Feb 27 17:00:38 crc systemd[1]: Starting Kubernetes Kubelet... Feb 27 17:00:38 crc restorecon[4695]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 27 17:00:38 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 17:00:39 crc restorecon[4695]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 17:00:39 crc restorecon[4695]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Feb 27 17:00:40 crc kubenswrapper[4700]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 27 17:00:40 crc kubenswrapper[4700]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Feb 27 17:00:40 crc kubenswrapper[4700]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 27 17:00:40 crc kubenswrapper[4700]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 27 17:00:40 crc kubenswrapper[4700]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Feb 27 17:00:40 crc kubenswrapper[4700]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.706069 4700 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711541 4700 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711574 4700 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711584 4700 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711593 4700 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711602 4700 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711611 4700 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711620 4700 feature_gate.go:330] unrecognized feature gate: Example Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711630 4700 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711638 4700 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711663 4700 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711671 4700 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711679 4700 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711687 4700 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711695 4700 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711703 4700 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711710 4700 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711718 4700 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711729 4700 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711739 4700 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711749 4700 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711758 4700 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711766 4700 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711775 4700 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711783 4700 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711790 4700 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711798 4700 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711805 4700 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711813 4700 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711820 4700 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711828 4700 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711835 4700 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711843 4700 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711850 4700 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711858 4700 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711866 4700 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711873 4700 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711883 4700 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711894 4700 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711905 4700 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711914 4700 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711923 4700 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711932 4700 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711941 4700 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711954 4700 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711964 4700 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711973 4700 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711982 4700 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711991 4700 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.711999 4700 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.712009 4700 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.712017 4700 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.712025 4700 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.712032 4700 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.712040 4700 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.712048 4700 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.712055 4700 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.712063 4700 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.712072 4700 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.712082 4700 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.712091 4700 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.712099 4700 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.712107 4700 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.712114 4700 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.712122 4700 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.712129 4700 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.712137 4700 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.712145 4700 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.712154 4700 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.712162 4700 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.712172 4700 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.712180 4700 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713215 4700 flags.go:64] FLAG: --address="0.0.0.0" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713241 4700 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713259 4700 flags.go:64] FLAG: --anonymous-auth="true" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713272 4700 flags.go:64] FLAG: --application-metrics-count-limit="100" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713284 4700 flags.go:64] FLAG: --authentication-token-webhook="false" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713293 4700 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713305 4700 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713317 4700 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713326 4700 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713336 4700 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713345 4700 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713356 4700 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713365 4700 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713374 4700 flags.go:64] FLAG: --cgroup-root="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713384 4700 flags.go:64] FLAG: --cgroups-per-qos="true" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713392 4700 flags.go:64] FLAG: --client-ca-file="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713402 4700 flags.go:64] FLAG: --cloud-config="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713410 4700 flags.go:64] FLAG: --cloud-provider="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713419 4700 flags.go:64] FLAG: --cluster-dns="[]" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713430 4700 flags.go:64] FLAG: --cluster-domain="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713438 4700 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713448 4700 flags.go:64] FLAG: --config-dir="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713465 4700 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713498 4700 flags.go:64] FLAG: --container-log-max-files="5" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713509 4700 flags.go:64] FLAG: --container-log-max-size="10Mi" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713518 4700 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713528 4700 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713537 4700 flags.go:64] FLAG: --containerd-namespace="k8s.io" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713545 4700 flags.go:64] FLAG: --contention-profiling="false" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713555 4700 flags.go:64] FLAG: --cpu-cfs-quota="true" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713564 4700 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713574 4700 flags.go:64] FLAG: --cpu-manager-policy="none" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713582 4700 flags.go:64] FLAG: --cpu-manager-policy-options="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713593 4700 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713602 4700 flags.go:64] FLAG: --enable-controller-attach-detach="true" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713612 4700 flags.go:64] FLAG: --enable-debugging-handlers="true" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713621 4700 flags.go:64] FLAG: --enable-load-reader="false" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713630 4700 flags.go:64] FLAG: --enable-server="true" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713640 4700 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713652 4700 flags.go:64] FLAG: --event-burst="100" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713662 4700 flags.go:64] FLAG: --event-qps="50" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713671 4700 flags.go:64] FLAG: --event-storage-age-limit="default=0" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713680 4700 flags.go:64] FLAG: --event-storage-event-limit="default=0" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713689 4700 flags.go:64] FLAG: --eviction-hard="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713700 4700 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713709 4700 flags.go:64] FLAG: --eviction-minimum-reclaim="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713719 4700 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713728 4700 flags.go:64] FLAG: --eviction-soft="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713737 4700 flags.go:64] FLAG: --eviction-soft-grace-period="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713745 4700 flags.go:64] FLAG: --exit-on-lock-contention="false" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713755 4700 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713764 4700 flags.go:64] FLAG: --experimental-mounter-path="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713773 4700 flags.go:64] FLAG: --fail-cgroupv1="false" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713782 4700 flags.go:64] FLAG: --fail-swap-on="true" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713791 4700 flags.go:64] FLAG: --feature-gates="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713802 4700 flags.go:64] FLAG: --file-check-frequency="20s" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713812 4700 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713821 4700 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713830 4700 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713840 4700 flags.go:64] FLAG: --healthz-port="10248" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713849 4700 flags.go:64] FLAG: --help="false" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713859 4700 flags.go:64] FLAG: --hostname-override="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713868 4700 flags.go:64] FLAG: --housekeeping-interval="10s" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713877 4700 flags.go:64] FLAG: --http-check-frequency="20s" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713886 4700 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713895 4700 flags.go:64] FLAG: --image-credential-provider-config="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713904 4700 flags.go:64] FLAG: --image-gc-high-threshold="85" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713913 4700 flags.go:64] FLAG: --image-gc-low-threshold="80" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713922 4700 flags.go:64] FLAG: --image-service-endpoint="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713931 4700 flags.go:64] FLAG: --kernel-memcg-notification="false" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713940 4700 flags.go:64] FLAG: --kube-api-burst="100" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713949 4700 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713959 4700 flags.go:64] FLAG: --kube-api-qps="50" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713970 4700 flags.go:64] FLAG: --kube-reserved="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713980 4700 flags.go:64] FLAG: --kube-reserved-cgroup="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713989 4700 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.713998 4700 flags.go:64] FLAG: --kubelet-cgroups="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714007 4700 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714016 4700 flags.go:64] FLAG: --lock-file="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714031 4700 flags.go:64] FLAG: --log-cadvisor-usage="false" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714039 4700 flags.go:64] FLAG: --log-flush-frequency="5s" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714048 4700 flags.go:64] FLAG: --log-json-info-buffer-size="0" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714062 4700 flags.go:64] FLAG: --log-json-split-stream="false" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714072 4700 flags.go:64] FLAG: --log-text-info-buffer-size="0" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714081 4700 flags.go:64] FLAG: --log-text-split-stream="false" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714090 4700 flags.go:64] FLAG: --logging-format="text" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714099 4700 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714109 4700 flags.go:64] FLAG: --make-iptables-util-chains="true" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714118 4700 flags.go:64] FLAG: --manifest-url="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714127 4700 flags.go:64] FLAG: --manifest-url-header="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714139 4700 flags.go:64] FLAG: --max-housekeeping-interval="15s" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714149 4700 flags.go:64] FLAG: --max-open-files="1000000" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714159 4700 flags.go:64] FLAG: --max-pods="110" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714171 4700 flags.go:64] FLAG: --maximum-dead-containers="-1" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714180 4700 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714189 4700 flags.go:64] FLAG: --memory-manager-policy="None" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714199 4700 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714209 4700 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714218 4700 flags.go:64] FLAG: --node-ip="192.168.126.11" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714228 4700 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714247 4700 flags.go:64] FLAG: --node-status-max-images="50" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714256 4700 flags.go:64] FLAG: --node-status-update-frequency="10s" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714265 4700 flags.go:64] FLAG: --oom-score-adj="-999" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714275 4700 flags.go:64] FLAG: --pod-cidr="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714284 4700 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714298 4700 flags.go:64] FLAG: --pod-manifest-path="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714307 4700 flags.go:64] FLAG: --pod-max-pids="-1" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714316 4700 flags.go:64] FLAG: --pods-per-core="0" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714325 4700 flags.go:64] FLAG: --port="10250" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714336 4700 flags.go:64] FLAG: --protect-kernel-defaults="false" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714345 4700 flags.go:64] FLAG: --provider-id="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714357 4700 flags.go:64] FLAG: --qos-reserved="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714366 4700 flags.go:64] FLAG: --read-only-port="10255" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714375 4700 flags.go:64] FLAG: --register-node="true" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714384 4700 flags.go:64] FLAG: --register-schedulable="true" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714392 4700 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714406 4700 flags.go:64] FLAG: --registry-burst="10" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714415 4700 flags.go:64] FLAG: --registry-qps="5" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714424 4700 flags.go:64] FLAG: --reserved-cpus="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714432 4700 flags.go:64] FLAG: --reserved-memory="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714443 4700 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714452 4700 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714466 4700 flags.go:64] FLAG: --rotate-certificates="false" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714508 4700 flags.go:64] FLAG: --rotate-server-certificates="false" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714518 4700 flags.go:64] FLAG: --runonce="false" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714527 4700 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714536 4700 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714546 4700 flags.go:64] FLAG: --seccomp-default="false" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714555 4700 flags.go:64] FLAG: --serialize-image-pulls="true" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714564 4700 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714575 4700 flags.go:64] FLAG: --storage-driver-db="cadvisor" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714585 4700 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714594 4700 flags.go:64] FLAG: --storage-driver-password="root" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714602 4700 flags.go:64] FLAG: --storage-driver-secure="false" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714612 4700 flags.go:64] FLAG: --storage-driver-table="stats" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714620 4700 flags.go:64] FLAG: --storage-driver-user="root" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714629 4700 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714668 4700 flags.go:64] FLAG: --sync-frequency="1m0s" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714678 4700 flags.go:64] FLAG: --system-cgroups="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714687 4700 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714701 4700 flags.go:64] FLAG: --system-reserved-cgroup="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714710 4700 flags.go:64] FLAG: --tls-cert-file="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714719 4700 flags.go:64] FLAG: --tls-cipher-suites="[]" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714733 4700 flags.go:64] FLAG: --tls-min-version="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714742 4700 flags.go:64] FLAG: --tls-private-key-file="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714753 4700 flags.go:64] FLAG: --topology-manager-policy="none" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714762 4700 flags.go:64] FLAG: --topology-manager-policy-options="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714770 4700 flags.go:64] FLAG: --topology-manager-scope="container" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714779 4700 flags.go:64] FLAG: --v="2" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714791 4700 flags.go:64] FLAG: --version="false" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714803 4700 flags.go:64] FLAG: --vmodule="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714814 4700 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.714823 4700 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715049 4700 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715059 4700 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715068 4700 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715079 4700 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715089 4700 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715099 4700 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715109 4700 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715118 4700 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715126 4700 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715134 4700 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715142 4700 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715149 4700 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715157 4700 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715165 4700 feature_gate.go:330] unrecognized feature gate: Example Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715173 4700 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715181 4700 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715189 4700 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715196 4700 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715204 4700 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715212 4700 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715220 4700 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715227 4700 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715238 4700 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715246 4700 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715254 4700 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715261 4700 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715269 4700 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715276 4700 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715296 4700 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715304 4700 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715314 4700 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715324 4700 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715333 4700 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715342 4700 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715350 4700 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715359 4700 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715367 4700 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715375 4700 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715382 4700 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715390 4700 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715398 4700 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715405 4700 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715413 4700 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715420 4700 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715428 4700 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715436 4700 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715444 4700 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715451 4700 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715465 4700 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715496 4700 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715504 4700 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715514 4700 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715524 4700 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715533 4700 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715543 4700 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715551 4700 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715559 4700 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715569 4700 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715578 4700 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715587 4700 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715596 4700 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715605 4700 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715613 4700 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715620 4700 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715629 4700 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715638 4700 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715647 4700 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715655 4700 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715663 4700 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715671 4700 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.715679 4700 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.715703 4700 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.728689 4700 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.728747 4700 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.728894 4700 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.728915 4700 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.728925 4700 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.728935 4700 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.728943 4700 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.728952 4700 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.728959 4700 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.728967 4700 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.728975 4700 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.728983 4700 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.728991 4700 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.728998 4700 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729008 4700 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729018 4700 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729027 4700 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729036 4700 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729067 4700 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729080 4700 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729089 4700 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729097 4700 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729104 4700 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729112 4700 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729119 4700 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729127 4700 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729135 4700 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729144 4700 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729152 4700 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729159 4700 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729167 4700 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729175 4700 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729182 4700 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729193 4700 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729202 4700 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729211 4700 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729234 4700 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729242 4700 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729251 4700 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729258 4700 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729269 4700 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729281 4700 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729290 4700 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729298 4700 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729307 4700 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729315 4700 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729325 4700 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729334 4700 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729342 4700 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729351 4700 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729358 4700 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729370 4700 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729378 4700 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729386 4700 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729393 4700 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729401 4700 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729409 4700 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729417 4700 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729425 4700 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729432 4700 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729440 4700 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729451 4700 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729469 4700 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729506 4700 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729515 4700 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729524 4700 feature_gate.go:330] unrecognized feature gate: Example Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729533 4700 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729541 4700 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729549 4700 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729557 4700 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729565 4700 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729573 4700 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729593 4700 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.729606 4700 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729904 4700 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729925 4700 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729935 4700 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729945 4700 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729953 4700 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729961 4700 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729969 4700 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729976 4700 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.729984 4700 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730009 4700 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730017 4700 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730026 4700 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730033 4700 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730041 4700 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730048 4700 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730056 4700 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730064 4700 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730071 4700 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730079 4700 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730087 4700 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730097 4700 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730107 4700 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730116 4700 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730126 4700 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730135 4700 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730142 4700 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730152 4700 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730162 4700 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730171 4700 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730179 4700 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730187 4700 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730195 4700 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730203 4700 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730211 4700 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730231 4700 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730238 4700 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730246 4700 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730254 4700 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730262 4700 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730270 4700 feature_gate.go:330] unrecognized feature gate: Example Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730278 4700 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730285 4700 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730293 4700 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730302 4700 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730309 4700 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730317 4700 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730324 4700 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730332 4700 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730340 4700 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730348 4700 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730356 4700 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730364 4700 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730371 4700 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730379 4700 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730386 4700 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730395 4700 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730402 4700 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730410 4700 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730419 4700 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730428 4700 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730436 4700 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730444 4700 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730452 4700 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730465 4700 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730473 4700 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730504 4700 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730512 4700 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730519 4700 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730527 4700 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730534 4700 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.730554 4700 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.730566 4700 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.731615 4700 server.go:940] "Client rotation is on, will bootstrap in background" Feb 27 17:00:40 crc kubenswrapper[4700]: E0227 17:00:40.737710 4700 bootstrap.go:266] "Unhandled Error" err="part of the existing bootstrap client certificate in /var/lib/kubelet/kubeconfig is expired: 2026-02-24 05:52:08 +0000 UTC" logger="UnhandledError" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.742234 4700 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.742376 4700 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.744614 4700 server.go:997] "Starting client certificate rotation" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.744666 4700 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.744954 4700 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.770238 4700 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 27 17:00:40 crc kubenswrapper[4700]: E0227 17:00:40.775952 4700 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.222:6443: connect: connection refused" logger="UnhandledError" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.776063 4700 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.789904 4700 log.go:25] "Validated CRI v1 runtime API" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.831831 4700 log.go:25] "Validated CRI v1 image API" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.834385 4700 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.840976 4700 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-02-27-16-55-45-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.841027 4700 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.874455 4700 manager.go:217] Machine: {Timestamp:2026-02-27 17:00:40.87072076 +0000 UTC m=+0.856033587 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654132736 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:c03ccb1e-6d2c-4062-a2f2-4418103b1627 BootID:aa2cf3e6-9e57-4694-81ae-b6e008a63c28 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730829824 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827068416 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:ca:d6:db Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:ca:d6:db Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:2f:aa:8f Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:4a:59:68 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:ac:a4:47 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:c4:17:5a Speed:-1 Mtu:1496} {Name:eth10 MacAddress:ba:77:1a:29:32:28 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:3a:5e:b2:2b:cd:5d Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654132736 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.875236 4700 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.875410 4700 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.877923 4700 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.878230 4700 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.878285 4700 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.878653 4700 topology_manager.go:138] "Creating topology manager with none policy" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.878673 4700 container_manager_linux.go:303] "Creating device plugin manager" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.879227 4700 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.879276 4700 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.879512 4700 state_mem.go:36] "Initialized new in-memory state store" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.879655 4700 server.go:1245] "Using root directory" path="/var/lib/kubelet" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.883465 4700 kubelet.go:418] "Attempting to sync node with API server" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.883520 4700 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.883562 4700 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.883587 4700 kubelet.go:324] "Adding apiserver pod source" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.883633 4700 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.891558 4700 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.892909 4700 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.222:6443: connect: connection refused Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.893024 4700 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.222:6443: connect: connection refused Feb 27 17:00:40 crc kubenswrapper[4700]: E0227 17:00:40.893054 4700 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.222:6443: connect: connection refused" logger="UnhandledError" Feb 27 17:00:40 crc kubenswrapper[4700]: E0227 17:00:40.893176 4700 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.222:6443: connect: connection refused" logger="UnhandledError" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.895613 4700 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.899051 4700 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.900892 4700 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.900937 4700 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.900952 4700 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.900968 4700 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.900991 4700 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.901006 4700 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.901023 4700 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.901046 4700 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.901062 4700 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.901078 4700 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.901113 4700 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.901128 4700 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.903117 4700 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.903766 4700 server.go:1280] "Started kubelet" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.904901 4700 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.904885 4700 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.905119 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.222:6443: connect: connection refused Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.905790 4700 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 27 17:00:40 crc systemd[1]: Started Kubernetes Kubelet. Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.906749 4700 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.906795 4700 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 27 17:00:40 crc kubenswrapper[4700]: E0227 17:00:40.907076 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.907145 4700 volume_manager.go:287] "The desired_state_of_world populator starts" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.907158 4700 volume_manager.go:289] "Starting Kubelet Volume Manager" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.907280 4700 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.909264 4700 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.222:6443: connect: connection refused Feb 27 17:00:40 crc kubenswrapper[4700]: E0227 17:00:40.909494 4700 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.222:6443: connect: connection refused" logger="UnhandledError" Feb 27 17:00:40 crc kubenswrapper[4700]: E0227 17:00:40.909284 4700 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.222:6443: connect: connection refused" interval="200ms" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.910273 4700 server.go:460] "Adding debug handlers to kubelet server" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.910299 4700 factory.go:153] Registering CRI-O factory Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.910514 4700 factory.go:221] Registration of the crio container factory successfully Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.910644 4700 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.910679 4700 factory.go:55] Registering systemd factory Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.910698 4700 factory.go:221] Registration of the systemd container factory successfully Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.910734 4700 factory.go:103] Registering Raw factory Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.910767 4700 manager.go:1196] Started watching for new ooms in manager Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.913131 4700 manager.go:319] Starting recovery of all containers Feb 27 17:00:40 crc kubenswrapper[4700]: E0227 17:00:40.916097 4700 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.222:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189829190ae0ef99 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:40.903724953 +0000 UTC m=+0.889037740,LastTimestamp:2026-02-27 17:00:40.903724953 +0000 UTC m=+0.889037740,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.932893 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.933035 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.933071 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.933098 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.933126 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.933155 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.933186 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.933270 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.933305 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.933337 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.933364 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.933391 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.933417 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.933449 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.933542 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.933574 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.933605 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.933632 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.933661 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.933689 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.933714 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.933765 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.933798 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.933825 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.933852 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.933877 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.933906 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.933934 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.934077 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.934109 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.934140 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.934169 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.934208 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.934235 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.934262 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.934289 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.934323 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.934439 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.934472 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.934581 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.934612 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.934639 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.934669 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.934696 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.934727 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.934755 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.934784 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.934811 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.934842 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.934867 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.934894 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.934923 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.935007 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.935043 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.935077 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.935106 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.935137 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.935207 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.935235 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.935262 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.935290 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.935320 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.935350 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.935384 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.935413 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.935442 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.935506 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.935543 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.935578 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.935608 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.935636 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.935662 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.935692 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.935720 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.935750 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.935777 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.935812 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.935843 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.935870 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.935899 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.935984 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.936011 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.936041 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.936067 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.936097 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.936128 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.936157 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.936183 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.936209 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.936236 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.936266 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.936294 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.936320 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.936346 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.936372 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.936400 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.936428 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.936451 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.936515 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.936547 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.936577 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.936604 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.936630 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.936657 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.936695 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.936726 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.936755 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.936798 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.936831 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.936862 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.936891 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.936919 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.936946 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.936989 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.937018 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.937047 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.937076 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.937104 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.939868 4700 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.939940 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.939973 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.940005 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.940035 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.940064 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.940091 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.940122 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.940152 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.940178 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.940201 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.940222 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.940416 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.940520 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.940693 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.940731 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.940762 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.940794 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.940821 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.940849 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.940878 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.940907 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.940936 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.940964 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.940991 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.941019 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.941049 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.941078 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.941107 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.941134 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.941166 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.941195 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.941227 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.941259 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.941289 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.941320 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.941351 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.941380 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.941413 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.941445 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.941542 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.941579 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.941608 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.941638 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.941669 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.941749 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.941779 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.941806 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.941834 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.941861 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.941889 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.941915 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.941940 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.941967 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.941994 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.942020 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.942044 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.942076 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.942104 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.942132 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.942158 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.942190 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.942220 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.942250 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.942280 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.942311 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.942338 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.942367 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.942395 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.942421 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.942451 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.942529 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.942559 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.942588 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.942614 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.942644 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.942719 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.942749 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.942782 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.942809 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.942845 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.942911 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.942939 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.942967 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.942993 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.943021 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.943050 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.943077 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.943107 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.943138 4700 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.943163 4700 reconstruct.go:97] "Volume reconstruction finished" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.943181 4700 reconciler.go:26] "Reconciler: start to sync state" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.949933 4700 manager.go:324] Recovery completed Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.966159 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.969196 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.969295 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.969321 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.970525 4700 cpu_manager.go:225] "Starting CPU manager" policy="none" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.970550 4700 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.970574 4700 state_mem.go:36] "Initialized new in-memory state store" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.975208 4700 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.979567 4700 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.979719 4700 status_manager.go:217] "Starting to sync pod status with apiserver" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.979815 4700 kubelet.go:2335] "Starting kubelet main sync loop" Feb 27 17:00:40 crc kubenswrapper[4700]: E0227 17:00:40.979963 4700 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 27 17:00:40 crc kubenswrapper[4700]: W0227 17:00:40.980774 4700 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.222:6443: connect: connection refused Feb 27 17:00:40 crc kubenswrapper[4700]: E0227 17:00:40.980877 4700 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.222:6443: connect: connection refused" logger="UnhandledError" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.988410 4700 policy_none.go:49] "None policy: Start" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.989828 4700 memory_manager.go:170] "Starting memorymanager" policy="None" Feb 27 17:00:40 crc kubenswrapper[4700]: I0227 17:00:40.989878 4700 state_mem.go:35] "Initializing new in-memory state store" Feb 27 17:00:41 crc kubenswrapper[4700]: E0227 17:00:41.007633 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.067259 4700 manager.go:334] "Starting Device Plugin manager" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.067366 4700 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.067389 4700 server.go:79] "Starting device plugin registration server" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.068078 4700 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.068229 4700 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.068538 4700 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.068712 4700 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.068728 4700 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.081012 4700 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.081144 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.083985 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.084085 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.084153 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.084760 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.084856 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.084949 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:41 crc kubenswrapper[4700]: E0227 17:00:41.085655 4700 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.087627 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.087687 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.087710 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.087771 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.087901 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.087972 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.088502 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.088790 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.088858 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.090798 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.090878 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.090900 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.091047 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.091133 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.091157 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.091159 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.092959 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.093080 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.093731 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.093797 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.093815 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.094052 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.094424 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.094589 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.095453 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.095521 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.095544 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.095555 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.095586 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.095602 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.096158 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.096328 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.096933 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.096983 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.097040 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.098212 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.098280 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.098311 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:41 crc kubenswrapper[4700]: E0227 17:00:41.111410 4700 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.222:6443: connect: connection refused" interval="400ms" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.146756 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.146903 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.149386 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.149451 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.149526 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.149571 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.149608 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.149660 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.149705 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.149785 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.150260 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.150613 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.150702 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.150798 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.150907 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.168778 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.170613 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.170663 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.170708 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.170755 4700 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 17:00:41 crc kubenswrapper[4700]: E0227 17:00:41.171605 4700 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.222:6443: connect: connection refused" node="crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.253700 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.253916 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.254386 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.254692 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.254402 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.254906 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.254986 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.255039 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.255085 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.255121 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.255152 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.255155 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.255185 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.255221 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.255251 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.255259 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.255289 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.255295 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.255227 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.255330 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.255342 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.255363 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.255395 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.255426 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.255434 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.255521 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.255571 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.255619 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.255646 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.255668 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.371902 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.373869 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.373947 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.374033 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.374078 4700 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 17:00:41 crc kubenswrapper[4700]: E0227 17:00:41.374815 4700 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.222:6443: connect: connection refused" node="crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.451630 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.468935 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.489509 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.499254 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.504318 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 17:00:41 crc kubenswrapper[4700]: E0227 17:00:41.513343 4700 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.222:6443: connect: connection refused" interval="800ms" Feb 27 17:00:41 crc kubenswrapper[4700]: W0227 17:00:41.520723 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-fb28cda052540f77161787a50367fd4a7476701a09ff21db462d95dc0fbe6ae6 WatchSource:0}: Error finding container fb28cda052540f77161787a50367fd4a7476701a09ff21db462d95dc0fbe6ae6: Status 404 returned error can't find the container with id fb28cda052540f77161787a50367fd4a7476701a09ff21db462d95dc0fbe6ae6 Feb 27 17:00:41 crc kubenswrapper[4700]: W0227 17:00:41.524975 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-2468460ce7e9398adcc09960b54a642333de73d167e2ac2a7db4c3cacf0e057a WatchSource:0}: Error finding container 2468460ce7e9398adcc09960b54a642333de73d167e2ac2a7db4c3cacf0e057a: Status 404 returned error can't find the container with id 2468460ce7e9398adcc09960b54a642333de73d167e2ac2a7db4c3cacf0e057a Feb 27 17:00:41 crc kubenswrapper[4700]: W0227 17:00:41.534795 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-a4fa55a71dd7ba69361941a20d43914e5e77cf91931bb3cb53fa07b5452616d0 WatchSource:0}: Error finding container a4fa55a71dd7ba69361941a20d43914e5e77cf91931bb3cb53fa07b5452616d0: Status 404 returned error can't find the container with id a4fa55a71dd7ba69361941a20d43914e5e77cf91931bb3cb53fa07b5452616d0 Feb 27 17:00:41 crc kubenswrapper[4700]: W0227 17:00:41.542643 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-68ba1b31fd1e521666860645dac8932d47069e5c1e012a595a2812944c2ff353 WatchSource:0}: Error finding container 68ba1b31fd1e521666860645dac8932d47069e5c1e012a595a2812944c2ff353: Status 404 returned error can't find the container with id 68ba1b31fd1e521666860645dac8932d47069e5c1e012a595a2812944c2ff353 Feb 27 17:00:41 crc kubenswrapper[4700]: W0227 17:00:41.547377 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-5ce9c7e49f19f4f6368b56d67d733773289c571101b02e2910facafbb8fba64c WatchSource:0}: Error finding container 5ce9c7e49f19f4f6368b56d67d733773289c571101b02e2910facafbb8fba64c: Status 404 returned error can't find the container with id 5ce9c7e49f19f4f6368b56d67d733773289c571101b02e2910facafbb8fba64c Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.775401 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.778554 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.778643 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.778667 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.778713 4700 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 17:00:41 crc kubenswrapper[4700]: E0227 17:00:41.779413 4700 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.222:6443: connect: connection refused" node="crc" Feb 27 17:00:41 crc kubenswrapper[4700]: W0227 17:00:41.820521 4700 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.222:6443: connect: connection refused Feb 27 17:00:41 crc kubenswrapper[4700]: E0227 17:00:41.820669 4700 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.222:6443: connect: connection refused" logger="UnhandledError" Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.906312 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.222:6443: connect: connection refused Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.993722 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5ce9c7e49f19f4f6368b56d67d733773289c571101b02e2910facafbb8fba64c"} Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.995156 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"68ba1b31fd1e521666860645dac8932d47069e5c1e012a595a2812944c2ff353"} Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.996839 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a4fa55a71dd7ba69361941a20d43914e5e77cf91931bb3cb53fa07b5452616d0"} Feb 27 17:00:41 crc kubenswrapper[4700]: I0227 17:00:41.998892 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2468460ce7e9398adcc09960b54a642333de73d167e2ac2a7db4c3cacf0e057a"} Feb 27 17:00:42 crc kubenswrapper[4700]: I0227 17:00:42.000281 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"fb28cda052540f77161787a50367fd4a7476701a09ff21db462d95dc0fbe6ae6"} Feb 27 17:00:42 crc kubenswrapper[4700]: W0227 17:00:42.186216 4700 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.222:6443: connect: connection refused Feb 27 17:00:42 crc kubenswrapper[4700]: E0227 17:00:42.186340 4700 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.222:6443: connect: connection refused" logger="UnhandledError" Feb 27 17:00:42 crc kubenswrapper[4700]: E0227 17:00:42.314863 4700 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.222:6443: connect: connection refused" interval="1.6s" Feb 27 17:00:42 crc kubenswrapper[4700]: W0227 17:00:42.345180 4700 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.222:6443: connect: connection refused Feb 27 17:00:42 crc kubenswrapper[4700]: E0227 17:00:42.345428 4700 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.222:6443: connect: connection refused" logger="UnhandledError" Feb 27 17:00:42 crc kubenswrapper[4700]: W0227 17:00:42.371954 4700 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.222:6443: connect: connection refused Feb 27 17:00:42 crc kubenswrapper[4700]: E0227 17:00:42.372114 4700 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.222:6443: connect: connection refused" logger="UnhandledError" Feb 27 17:00:42 crc kubenswrapper[4700]: I0227 17:00:42.579646 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:42 crc kubenswrapper[4700]: I0227 17:00:42.581812 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:42 crc kubenswrapper[4700]: I0227 17:00:42.581889 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:42 crc kubenswrapper[4700]: I0227 17:00:42.581911 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:42 crc kubenswrapper[4700]: I0227 17:00:42.581950 4700 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 17:00:42 crc kubenswrapper[4700]: E0227 17:00:42.582762 4700 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.222:6443: connect: connection refused" node="crc" Feb 27 17:00:42 crc kubenswrapper[4700]: I0227 17:00:42.827523 4700 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 27 17:00:42 crc kubenswrapper[4700]: E0227 17:00:42.828954 4700 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.222:6443: connect: connection refused" logger="UnhandledError" Feb 27 17:00:42 crc kubenswrapper[4700]: I0227 17:00:42.906540 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.222:6443: connect: connection refused Feb 27 17:00:43 crc kubenswrapper[4700]: I0227 17:00:43.005831 4700 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30" exitCode=0 Feb 27 17:00:43 crc kubenswrapper[4700]: I0227 17:00:43.005964 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30"} Feb 27 17:00:43 crc kubenswrapper[4700]: I0227 17:00:43.006078 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:43 crc kubenswrapper[4700]: I0227 17:00:43.007675 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:43 crc kubenswrapper[4700]: I0227 17:00:43.007727 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:43 crc kubenswrapper[4700]: I0227 17:00:43.007744 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:43 crc kubenswrapper[4700]: I0227 17:00:43.008712 4700 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0" exitCode=0 Feb 27 17:00:43 crc kubenswrapper[4700]: I0227 17:00:43.008827 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0"} Feb 27 17:00:43 crc kubenswrapper[4700]: I0227 17:00:43.008876 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:43 crc kubenswrapper[4700]: I0227 17:00:43.010647 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:43 crc kubenswrapper[4700]: I0227 17:00:43.010695 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:43 crc kubenswrapper[4700]: I0227 17:00:43.010717 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:43 crc kubenswrapper[4700]: I0227 17:00:43.012424 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"02f172e74f73a1db01fc2f510976b9d061a51b44eec618fbad2e2ac2e0f33f45"} Feb 27 17:00:43 crc kubenswrapper[4700]: I0227 17:00:43.012485 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"523f47c5e3ee759e5c1f7a782c10d08a2eb10fba3833c7fe40d6105867420b77"} Feb 27 17:00:43 crc kubenswrapper[4700]: I0227 17:00:43.012510 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"555793b646e823bff19401b489c4ab8de44ee3cd8f82ebcd66b5b08e67450be7"} Feb 27 17:00:43 crc kubenswrapper[4700]: I0227 17:00:43.015044 4700 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="3b6649f7710d4c1435bfd703120b70e4514f95b4750b247b8ad0a69ed6952a3b" exitCode=0 Feb 27 17:00:43 crc kubenswrapper[4700]: I0227 17:00:43.015113 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"3b6649f7710d4c1435bfd703120b70e4514f95b4750b247b8ad0a69ed6952a3b"} Feb 27 17:00:43 crc kubenswrapper[4700]: I0227 17:00:43.015200 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:43 crc kubenswrapper[4700]: I0227 17:00:43.016448 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:43 crc kubenswrapper[4700]: I0227 17:00:43.016547 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:43 crc kubenswrapper[4700]: I0227 17:00:43.016573 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:43 crc kubenswrapper[4700]: I0227 17:00:43.017209 4700 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c" exitCode=0 Feb 27 17:00:43 crc kubenswrapper[4700]: I0227 17:00:43.017251 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c"} Feb 27 17:00:43 crc kubenswrapper[4700]: I0227 17:00:43.017360 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:43 crc kubenswrapper[4700]: I0227 17:00:43.018659 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:43 crc kubenswrapper[4700]: I0227 17:00:43.018720 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:43 crc kubenswrapper[4700]: I0227 17:00:43.018746 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:43 crc kubenswrapper[4700]: I0227 17:00:43.020623 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:43 crc kubenswrapper[4700]: I0227 17:00:43.021673 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:43 crc kubenswrapper[4700]: I0227 17:00:43.021736 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:43 crc kubenswrapper[4700]: I0227 17:00:43.021762 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:43 crc kubenswrapper[4700]: I0227 17:00:43.906515 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.222:6443: connect: connection refused Feb 27 17:00:43 crc kubenswrapper[4700]: E0227 17:00:43.916047 4700 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.222:6443: connect: connection refused" interval="3.2s" Feb 27 17:00:44 crc kubenswrapper[4700]: I0227 17:00:44.023309 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"4222e543e1fff164630ef807cefe3110095d68d3b8c90fe8dffdcdb5708abceb"} Feb 27 17:00:44 crc kubenswrapper[4700]: I0227 17:00:44.023409 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:44 crc kubenswrapper[4700]: I0227 17:00:44.024809 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:44 crc kubenswrapper[4700]: I0227 17:00:44.024850 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:44 crc kubenswrapper[4700]: I0227 17:00:44.024863 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:44 crc kubenswrapper[4700]: I0227 17:00:44.027800 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"cf6d6f90f1a5acb192ed932732bab2c2a2bebe9f03fddc79c5dfe6423fab012c"} Feb 27 17:00:44 crc kubenswrapper[4700]: I0227 17:00:44.027826 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"f5a0dc977696434dce7529115ee1d6cce69cbd4c501bbfb77ed0d240d6340d2b"} Feb 27 17:00:44 crc kubenswrapper[4700]: I0227 17:00:44.027838 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"9da83072905af757ff9f9cecda0d94b2acfcd847aeeba7ca36ac291c4ff6a3e9"} Feb 27 17:00:44 crc kubenswrapper[4700]: I0227 17:00:44.027964 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:44 crc kubenswrapper[4700]: I0227 17:00:44.029278 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:44 crc kubenswrapper[4700]: I0227 17:00:44.029345 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:44 crc kubenswrapper[4700]: I0227 17:00:44.029369 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:44 crc kubenswrapper[4700]: I0227 17:00:44.030895 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951"} Feb 27 17:00:44 crc kubenswrapper[4700]: I0227 17:00:44.030966 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5"} Feb 27 17:00:44 crc kubenswrapper[4700]: I0227 17:00:44.030995 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb"} Feb 27 17:00:44 crc kubenswrapper[4700]: I0227 17:00:44.032456 4700 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b" exitCode=0 Feb 27 17:00:44 crc kubenswrapper[4700]: I0227 17:00:44.032818 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b"} Feb 27 17:00:44 crc kubenswrapper[4700]: I0227 17:00:44.032866 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:44 crc kubenswrapper[4700]: I0227 17:00:44.034255 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:44 crc kubenswrapper[4700]: I0227 17:00:44.034312 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:44 crc kubenswrapper[4700]: I0227 17:00:44.034336 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:44 crc kubenswrapper[4700]: I0227 17:00:44.035143 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"eabca65587272a0e2ce09d13f103fbbe5374fb6d5ed49bfa0e26de8336d9e683"} Feb 27 17:00:44 crc kubenswrapper[4700]: I0227 17:00:44.035295 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:44 crc kubenswrapper[4700]: I0227 17:00:44.036589 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:44 crc kubenswrapper[4700]: I0227 17:00:44.036627 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:44 crc kubenswrapper[4700]: I0227 17:00:44.036647 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:44 crc kubenswrapper[4700]: W0227 17:00:44.179221 4700 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.222:6443: connect: connection refused Feb 27 17:00:44 crc kubenswrapper[4700]: E0227 17:00:44.179324 4700 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.222:6443: connect: connection refused" logger="UnhandledError" Feb 27 17:00:44 crc kubenswrapper[4700]: I0227 17:00:44.183577 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:44 crc kubenswrapper[4700]: I0227 17:00:44.184791 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:44 crc kubenswrapper[4700]: I0227 17:00:44.184820 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:44 crc kubenswrapper[4700]: I0227 17:00:44.184832 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:44 crc kubenswrapper[4700]: I0227 17:00:44.184853 4700 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 17:00:44 crc kubenswrapper[4700]: E0227 17:00:44.185108 4700 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.222:6443: connect: connection refused" node="crc" Feb 27 17:00:44 crc kubenswrapper[4700]: W0227 17:00:44.288445 4700 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.222:6443: connect: connection refused Feb 27 17:00:44 crc kubenswrapper[4700]: E0227 17:00:44.288610 4700 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.222:6443: connect: connection refused" logger="UnhandledError" Feb 27 17:00:45 crc kubenswrapper[4700]: I0227 17:00:45.041666 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6b31fa0e2228210e4d48763af2d528541eadbcd37ee2b338c2af05c9db0f399a"} Feb 27 17:00:45 crc kubenswrapper[4700]: I0227 17:00:45.041715 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35"} Feb 27 17:00:45 crc kubenswrapper[4700]: I0227 17:00:45.041828 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:45 crc kubenswrapper[4700]: I0227 17:00:45.043255 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:45 crc kubenswrapper[4700]: I0227 17:00:45.043346 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:45 crc kubenswrapper[4700]: I0227 17:00:45.043368 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:45 crc kubenswrapper[4700]: I0227 17:00:45.044356 4700 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef" exitCode=0 Feb 27 17:00:45 crc kubenswrapper[4700]: I0227 17:00:45.044493 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:45 crc kubenswrapper[4700]: I0227 17:00:45.044504 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:45 crc kubenswrapper[4700]: I0227 17:00:45.044559 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef"} Feb 27 17:00:45 crc kubenswrapper[4700]: I0227 17:00:45.044509 4700 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 27 17:00:45 crc kubenswrapper[4700]: I0227 17:00:45.044696 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:45 crc kubenswrapper[4700]: I0227 17:00:45.044717 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:45 crc kubenswrapper[4700]: I0227 17:00:45.046505 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:45 crc kubenswrapper[4700]: I0227 17:00:45.046525 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:45 crc kubenswrapper[4700]: I0227 17:00:45.046637 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:45 crc kubenswrapper[4700]: I0227 17:00:45.046556 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:45 crc kubenswrapper[4700]: I0227 17:00:45.046665 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:45 crc kubenswrapper[4700]: I0227 17:00:45.046693 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:45 crc kubenswrapper[4700]: I0227 17:00:45.050362 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:45 crc kubenswrapper[4700]: I0227 17:00:45.050380 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:45 crc kubenswrapper[4700]: I0227 17:00:45.050406 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:45 crc kubenswrapper[4700]: I0227 17:00:45.050441 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:45 crc kubenswrapper[4700]: I0227 17:00:45.050522 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:45 crc kubenswrapper[4700]: I0227 17:00:45.050519 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:45 crc kubenswrapper[4700]: I0227 17:00:45.302896 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:00:46 crc kubenswrapper[4700]: I0227 17:00:46.054116 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8a14230f224fb3b7b9b0ced93634313b66ac3abc376dca228bc5802890b0c393"} Feb 27 17:00:46 crc kubenswrapper[4700]: I0227 17:00:46.054190 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9af0992b3383953e562bf540bf7659d41e60aeee5d6103105eaab164797486dd"} Feb 27 17:00:46 crc kubenswrapper[4700]: I0227 17:00:46.054218 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e64972d043accab95774784cc95dc1144e70c855eff777664c66cb121d519d4b"} Feb 27 17:00:46 crc kubenswrapper[4700]: I0227 17:00:46.054273 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:46 crc kubenswrapper[4700]: I0227 17:00:46.054443 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:00:46 crc kubenswrapper[4700]: I0227 17:00:46.055771 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:46 crc kubenswrapper[4700]: I0227 17:00:46.055821 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:46 crc kubenswrapper[4700]: I0227 17:00:46.055840 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:46 crc kubenswrapper[4700]: I0227 17:00:46.295678 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 17:00:46 crc kubenswrapper[4700]: I0227 17:00:46.296036 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:46 crc kubenswrapper[4700]: I0227 17:00:46.297924 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:46 crc kubenswrapper[4700]: I0227 17:00:46.298020 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:46 crc kubenswrapper[4700]: I0227 17:00:46.298048 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:46 crc kubenswrapper[4700]: I0227 17:00:46.306337 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 17:00:47 crc kubenswrapper[4700]: I0227 17:00:47.013308 4700 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 27 17:00:47 crc kubenswrapper[4700]: I0227 17:00:47.063051 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4fa2391deb934395fe4743249887832731bf7c6ea457d66484bcb6b556e4ab33"} Feb 27 17:00:47 crc kubenswrapper[4700]: I0227 17:00:47.063131 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"66f09ed7c87894bc0473c7015f0e0d55a28f893f5a2e8bf9085c86654c464371"} Feb 27 17:00:47 crc kubenswrapper[4700]: I0227 17:00:47.063270 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:47 crc kubenswrapper[4700]: I0227 17:00:47.063333 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:47 crc kubenswrapper[4700]: I0227 17:00:47.064382 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:47 crc kubenswrapper[4700]: I0227 17:00:47.064829 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:47 crc kubenswrapper[4700]: I0227 17:00:47.064885 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:47 crc kubenswrapper[4700]: I0227 17:00:47.064903 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:47 crc kubenswrapper[4700]: I0227 17:00:47.065216 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:47 crc kubenswrapper[4700]: I0227 17:00:47.065393 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:47 crc kubenswrapper[4700]: I0227 17:00:47.065419 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:47 crc kubenswrapper[4700]: I0227 17:00:47.065873 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:47 crc kubenswrapper[4700]: I0227 17:00:47.065915 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:47 crc kubenswrapper[4700]: I0227 17:00:47.065933 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:47 crc kubenswrapper[4700]: I0227 17:00:47.100347 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Feb 27 17:00:47 crc kubenswrapper[4700]: I0227 17:00:47.160161 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 17:00:47 crc kubenswrapper[4700]: I0227 17:00:47.385992 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:47 crc kubenswrapper[4700]: I0227 17:00:47.388205 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:47 crc kubenswrapper[4700]: I0227 17:00:47.388325 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:47 crc kubenswrapper[4700]: I0227 17:00:47.388355 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:47 crc kubenswrapper[4700]: I0227 17:00:47.388404 4700 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 17:00:47 crc kubenswrapper[4700]: I0227 17:00:47.646537 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:00:47 crc kubenswrapper[4700]: I0227 17:00:47.824687 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 17:00:47 crc kubenswrapper[4700]: I0227 17:00:47.824983 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:47 crc kubenswrapper[4700]: I0227 17:00:47.826699 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:47 crc kubenswrapper[4700]: I0227 17:00:47.826761 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:47 crc kubenswrapper[4700]: I0227 17:00:47.826771 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:48 crc kubenswrapper[4700]: I0227 17:00:48.066039 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:48 crc kubenswrapper[4700]: I0227 17:00:48.066115 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:48 crc kubenswrapper[4700]: I0227 17:00:48.066194 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:48 crc kubenswrapper[4700]: I0227 17:00:48.068419 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:48 crc kubenswrapper[4700]: I0227 17:00:48.068456 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:48 crc kubenswrapper[4700]: I0227 17:00:48.068453 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:48 crc kubenswrapper[4700]: I0227 17:00:48.068499 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:48 crc kubenswrapper[4700]: I0227 17:00:48.068519 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:48 crc kubenswrapper[4700]: I0227 17:00:48.068540 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:48 crc kubenswrapper[4700]: I0227 17:00:48.068570 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:48 crc kubenswrapper[4700]: I0227 17:00:48.068620 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:48 crc kubenswrapper[4700]: I0227 17:00:48.068639 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:48 crc kubenswrapper[4700]: I0227 17:00:48.575667 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 17:00:48 crc kubenswrapper[4700]: I0227 17:00:48.914584 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Feb 27 17:00:49 crc kubenswrapper[4700]: I0227 17:00:49.068752 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:49 crc kubenswrapper[4700]: I0227 17:00:49.068847 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:49 crc kubenswrapper[4700]: I0227 17:00:49.069989 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:49 crc kubenswrapper[4700]: I0227 17:00:49.070066 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:49 crc kubenswrapper[4700]: I0227 17:00:49.070085 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:49 crc kubenswrapper[4700]: I0227 17:00:49.070186 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:49 crc kubenswrapper[4700]: I0227 17:00:49.070221 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:49 crc kubenswrapper[4700]: I0227 17:00:49.070239 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:50 crc kubenswrapper[4700]: I0227 17:00:50.072425 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:50 crc kubenswrapper[4700]: I0227 17:00:50.073716 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:50 crc kubenswrapper[4700]: I0227 17:00:50.073749 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:50 crc kubenswrapper[4700]: I0227 17:00:50.073764 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:51 crc kubenswrapper[4700]: E0227 17:00:51.086674 4700 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 27 17:00:53 crc kubenswrapper[4700]: I0227 17:00:53.409326 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 17:00:53 crc kubenswrapper[4700]: I0227 17:00:53.409616 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:53 crc kubenswrapper[4700]: I0227 17:00:53.411504 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:53 crc kubenswrapper[4700]: I0227 17:00:53.411568 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:53 crc kubenswrapper[4700]: I0227 17:00:53.411588 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:53 crc kubenswrapper[4700]: I0227 17:00:53.417218 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 17:00:54 crc kubenswrapper[4700]: I0227 17:00:54.083838 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:54 crc kubenswrapper[4700]: I0227 17:00:54.084914 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:54 crc kubenswrapper[4700]: I0227 17:00:54.084975 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:54 crc kubenswrapper[4700]: I0227 17:00:54.084998 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:54 crc kubenswrapper[4700]: I0227 17:00:54.907025 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Feb 27 17:00:55 crc kubenswrapper[4700]: W0227 17:00:55.145993 4700 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Feb 27 17:00:55 crc kubenswrapper[4700]: I0227 17:00:55.146125 4700 trace.go:236] Trace[666023737]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (27-Feb-2026 17:00:45.144) (total time: 10001ms): Feb 27 17:00:55 crc kubenswrapper[4700]: Trace[666023737]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (17:00:55.145) Feb 27 17:00:55 crc kubenswrapper[4700]: Trace[666023737]: [10.001988376s] [10.001988376s] END Feb 27 17:00:55 crc kubenswrapper[4700]: E0227 17:00:55.146162 4700 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Feb 27 17:00:55 crc kubenswrapper[4700]: W0227 17:00:55.259425 4700 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Feb 27 17:00:55 crc kubenswrapper[4700]: I0227 17:00:55.259648 4700 trace.go:236] Trace[979762033]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (27-Feb-2026 17:00:45.257) (total time: 10001ms): Feb 27 17:00:55 crc kubenswrapper[4700]: Trace[979762033]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (17:00:55.259) Feb 27 17:00:55 crc kubenswrapper[4700]: Trace[979762033]: [10.001739348s] [10.001739348s] END Feb 27 17:00:55 crc kubenswrapper[4700]: E0227 17:00:55.259701 4700 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Feb 27 17:00:55 crc kubenswrapper[4700]: E0227 17:00:55.626502 4700 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": net/http: TLS handshake timeout" event="&Event{ObjectMeta:{crc.189829190ae0ef99 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:40.903724953 +0000 UTC m=+0.889037740,LastTimestamp:2026-02-27 17:00:40.903724953 +0000 UTC m=+0.889037740,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:00:55 crc kubenswrapper[4700]: E0227 17:00:55.735368 4700 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:00:55Z is after 2026-02-23T05:33:13Z" node="crc" Feb 27 17:00:55 crc kubenswrapper[4700]: W0227 17:00:55.739688 4700 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:00:55Z is after 2026-02-23T05:33:13Z Feb 27 17:00:55 crc kubenswrapper[4700]: E0227 17:00:55.739779 4700 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:00:55Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 27 17:00:55 crc kubenswrapper[4700]: I0227 17:00:55.740345 4700 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:53758->192.168.126.11:17697: read: connection reset by peer" start-of-body= Feb 27 17:00:55 crc kubenswrapper[4700]: I0227 17:00:55.740405 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:53758->192.168.126.11:17697: read: connection reset by peer" Feb 27 17:00:55 crc kubenswrapper[4700]: W0227 17:00:55.743526 4700 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:00:55Z is after 2026-02-23T05:33:13Z Feb 27 17:00:55 crc kubenswrapper[4700]: E0227 17:00:55.743599 4700 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:00:55Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 27 17:00:55 crc kubenswrapper[4700]: I0227 17:00:55.744819 4700 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 27 17:00:55 crc kubenswrapper[4700]: I0227 17:00:55.744909 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 27 17:00:55 crc kubenswrapper[4700]: E0227 17:00:55.747319 4700 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:00:55Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 27 17:00:55 crc kubenswrapper[4700]: E0227 17:00:55.752432 4700 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:00:55Z is after 2026-02-23T05:33:13Z" interval="6.4s" Feb 27 17:00:55 crc kubenswrapper[4700]: I0227 17:00:55.754406 4700 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 27 17:00:55 crc kubenswrapper[4700]: I0227 17:00:55.754460 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 27 17:00:55 crc kubenswrapper[4700]: I0227 17:00:55.913007 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:00:55Z is after 2026-02-23T05:33:13Z Feb 27 17:00:56 crc kubenswrapper[4700]: I0227 17:00:56.091046 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 27 17:00:56 crc kubenswrapper[4700]: I0227 17:00:56.093914 4700 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6b31fa0e2228210e4d48763af2d528541eadbcd37ee2b338c2af05c9db0f399a" exitCode=255 Feb 27 17:00:56 crc kubenswrapper[4700]: I0227 17:00:56.093969 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"6b31fa0e2228210e4d48763af2d528541eadbcd37ee2b338c2af05c9db0f399a"} Feb 27 17:00:56 crc kubenswrapper[4700]: I0227 17:00:56.094194 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:56 crc kubenswrapper[4700]: I0227 17:00:56.095359 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:56 crc kubenswrapper[4700]: I0227 17:00:56.095412 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:56 crc kubenswrapper[4700]: I0227 17:00:56.095431 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:56 crc kubenswrapper[4700]: I0227 17:00:56.096329 4700 scope.go:117] "RemoveContainer" containerID="6b31fa0e2228210e4d48763af2d528541eadbcd37ee2b338c2af05c9db0f399a" Feb 27 17:00:56 crc kubenswrapper[4700]: I0227 17:00:56.420603 4700 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 17:00:56 crc kubenswrapper[4700]: I0227 17:00:56.420727 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 17:00:56 crc kubenswrapper[4700]: I0227 17:00:56.909601 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:00:56Z is after 2026-02-23T05:33:13Z Feb 27 17:00:57 crc kubenswrapper[4700]: I0227 17:00:57.099384 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 27 17:00:57 crc kubenswrapper[4700]: I0227 17:00:57.102010 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d3dcb3b876b32e2e8dc9c1c45bffa575532863bacee7c89f103fa057a023c109"} Feb 27 17:00:57 crc kubenswrapper[4700]: I0227 17:00:57.102201 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:57 crc kubenswrapper[4700]: I0227 17:00:57.103405 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:57 crc kubenswrapper[4700]: I0227 17:00:57.103547 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:57 crc kubenswrapper[4700]: I0227 17:00:57.103577 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:57 crc kubenswrapper[4700]: I0227 17:00:57.654915 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:00:57 crc kubenswrapper[4700]: I0227 17:00:57.911167 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:00:57Z is after 2026-02-23T05:33:13Z Feb 27 17:00:58 crc kubenswrapper[4700]: I0227 17:00:58.108425 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 27 17:00:58 crc kubenswrapper[4700]: I0227 17:00:58.109299 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 27 17:00:58 crc kubenswrapper[4700]: I0227 17:00:58.112106 4700 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d3dcb3b876b32e2e8dc9c1c45bffa575532863bacee7c89f103fa057a023c109" exitCode=255 Feb 27 17:00:58 crc kubenswrapper[4700]: I0227 17:00:58.112171 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"d3dcb3b876b32e2e8dc9c1c45bffa575532863bacee7c89f103fa057a023c109"} Feb 27 17:00:58 crc kubenswrapper[4700]: I0227 17:00:58.112235 4700 scope.go:117] "RemoveContainer" containerID="6b31fa0e2228210e4d48763af2d528541eadbcd37ee2b338c2af05c9db0f399a" Feb 27 17:00:58 crc kubenswrapper[4700]: I0227 17:00:58.112277 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:58 crc kubenswrapper[4700]: I0227 17:00:58.113954 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:58 crc kubenswrapper[4700]: I0227 17:00:58.114016 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:58 crc kubenswrapper[4700]: I0227 17:00:58.114040 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:58 crc kubenswrapper[4700]: I0227 17:00:58.115123 4700 scope.go:117] "RemoveContainer" containerID="d3dcb3b876b32e2e8dc9c1c45bffa575532863bacee7c89f103fa057a023c109" Feb 27 17:00:58 crc kubenswrapper[4700]: E0227 17:00:58.115440 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 17:00:58 crc kubenswrapper[4700]: I0227 17:00:58.121010 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:00:58 crc kubenswrapper[4700]: I0227 17:00:58.913664 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:00:58Z is after 2026-02-23T05:33:13Z Feb 27 17:00:58 crc kubenswrapper[4700]: I0227 17:00:58.957436 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Feb 27 17:00:58 crc kubenswrapper[4700]: I0227 17:00:58.957728 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:58 crc kubenswrapper[4700]: I0227 17:00:58.959656 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:58 crc kubenswrapper[4700]: I0227 17:00:58.959696 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:58 crc kubenswrapper[4700]: I0227 17:00:58.959713 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:58 crc kubenswrapper[4700]: I0227 17:00:58.979057 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Feb 27 17:00:59 crc kubenswrapper[4700]: I0227 17:00:59.118319 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 27 17:00:59 crc kubenswrapper[4700]: I0227 17:00:59.121592 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:59 crc kubenswrapper[4700]: I0227 17:00:59.121707 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:00:59 crc kubenswrapper[4700]: I0227 17:00:59.125340 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:59 crc kubenswrapper[4700]: I0227 17:00:59.125482 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:59 crc kubenswrapper[4700]: I0227 17:00:59.125512 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:59 crc kubenswrapper[4700]: I0227 17:00:59.129931 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:00:59 crc kubenswrapper[4700]: I0227 17:00:59.130012 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:00:59 crc kubenswrapper[4700]: I0227 17:00:59.130042 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:00:59 crc kubenswrapper[4700]: I0227 17:00:59.130988 4700 scope.go:117] "RemoveContainer" containerID="d3dcb3b876b32e2e8dc9c1c45bffa575532863bacee7c89f103fa057a023c109" Feb 27 17:00:59 crc kubenswrapper[4700]: E0227 17:00:59.131299 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 17:00:59 crc kubenswrapper[4700]: W0227 17:00:59.692793 4700 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:00:59Z is after 2026-02-23T05:33:13Z Feb 27 17:00:59 crc kubenswrapper[4700]: E0227 17:00:59.692907 4700 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:00:59Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 27 17:00:59 crc kubenswrapper[4700]: I0227 17:00:59.911252 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:00:59Z is after 2026-02-23T05:33:13Z Feb 27 17:01:00 crc kubenswrapper[4700]: W0227 17:01:00.094130 4700 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:01:00Z is after 2026-02-23T05:33:13Z Feb 27 17:01:00 crc kubenswrapper[4700]: E0227 17:01:00.094243 4700 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:01:00Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 27 17:01:00 crc kubenswrapper[4700]: I0227 17:01:00.124633 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:01:00 crc kubenswrapper[4700]: I0227 17:01:00.125916 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:01:00 crc kubenswrapper[4700]: I0227 17:01:00.125996 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:01:00 crc kubenswrapper[4700]: I0227 17:01:00.126018 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:01:00 crc kubenswrapper[4700]: I0227 17:01:00.126899 4700 scope.go:117] "RemoveContainer" containerID="d3dcb3b876b32e2e8dc9c1c45bffa575532863bacee7c89f103fa057a023c109" Feb 27 17:01:00 crc kubenswrapper[4700]: E0227 17:01:00.127230 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 17:01:00 crc kubenswrapper[4700]: I0227 17:01:00.910731 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:01:00Z is after 2026-02-23T05:33:13Z Feb 27 17:01:01 crc kubenswrapper[4700]: E0227 17:01:01.087370 4700 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 27 17:01:01 crc kubenswrapper[4700]: I0227 17:01:01.520540 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:01:01 crc kubenswrapper[4700]: I0227 17:01:01.520790 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:01:01 crc kubenswrapper[4700]: I0227 17:01:01.522913 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:01:01 crc kubenswrapper[4700]: I0227 17:01:01.522974 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:01:01 crc kubenswrapper[4700]: I0227 17:01:01.522998 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:01:01 crc kubenswrapper[4700]: I0227 17:01:01.524025 4700 scope.go:117] "RemoveContainer" containerID="d3dcb3b876b32e2e8dc9c1c45bffa575532863bacee7c89f103fa057a023c109" Feb 27 17:01:01 crc kubenswrapper[4700]: E0227 17:01:01.524333 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 17:01:01 crc kubenswrapper[4700]: I0227 17:01:01.910532 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:01:01Z is after 2026-02-23T05:33:13Z Feb 27 17:01:02 crc kubenswrapper[4700]: I0227 17:01:02.136210 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:01:02 crc kubenswrapper[4700]: I0227 17:01:02.137883 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:01:02 crc kubenswrapper[4700]: I0227 17:01:02.137978 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:01:02 crc kubenswrapper[4700]: I0227 17:01:02.138009 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:01:02 crc kubenswrapper[4700]: I0227 17:01:02.138065 4700 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 17:01:02 crc kubenswrapper[4700]: E0227 17:01:02.143340 4700 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:01:02Z is after 2026-02-23T05:33:13Z" node="crc" Feb 27 17:01:02 crc kubenswrapper[4700]: E0227 17:01:02.159153 4700 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:01:02Z is after 2026-02-23T05:33:13Z" interval="7s" Feb 27 17:01:02 crc kubenswrapper[4700]: I0227 17:01:02.913274 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:03 crc kubenswrapper[4700]: W0227 17:01:03.221399 4700 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:03 crc kubenswrapper[4700]: E0227 17:01:03.221502 4700 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" Feb 27 17:01:03 crc kubenswrapper[4700]: I0227 17:01:03.913871 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:04 crc kubenswrapper[4700]: I0227 17:01:04.438258 4700 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 27 17:01:04 crc kubenswrapper[4700]: I0227 17:01:04.459586 4700 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 27 17:01:04 crc kubenswrapper[4700]: I0227 17:01:04.575758 4700 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:01:04 crc kubenswrapper[4700]: I0227 17:01:04.576000 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:01:04 crc kubenswrapper[4700]: I0227 17:01:04.577714 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:01:04 crc kubenswrapper[4700]: I0227 17:01:04.577764 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:01:04 crc kubenswrapper[4700]: I0227 17:01:04.577785 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:01:04 crc kubenswrapper[4700]: I0227 17:01:04.578677 4700 scope.go:117] "RemoveContainer" containerID="d3dcb3b876b32e2e8dc9c1c45bffa575532863bacee7c89f103fa057a023c109" Feb 27 17:01:04 crc kubenswrapper[4700]: E0227 17:01:04.579047 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 17:01:04 crc kubenswrapper[4700]: I0227 17:01:04.913598 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.634892 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189829190ae0ef99 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:40.903724953 +0000 UTC m=+0.889037740,LastTimestamp:2026-02-27 17:00:40.903724953 +0000 UTC m=+0.889037740,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.642995 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189829190ec8fec3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:40.969264835 +0000 UTC m=+0.954577622,LastTimestamp:2026-02-27 17:00:40.969264835 +0000 UTC m=+0.954577622,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.650337 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189829190ec9b04f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:40.969310287 +0000 UTC m=+0.954623074,LastTimestamp:2026-02-27 17:00:40.969310287 +0000 UTC m=+0.954623074,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.657278 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189829190eca0649 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:40.969332297 +0000 UTC m=+0.954645084,LastTimestamp:2026-02-27 17:00:40.969332297 +0000 UTC m=+0.954645084,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.664125 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.1898291914e83dba default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeAllocatableEnforced,Message:Updated Node Allocatable limit across pods,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:41.071975866 +0000 UTC m=+1.057288643,LastTimestamp:2026-02-27 17:00:41.071975866 +0000 UTC m=+1.057288643,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.671584 4700 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189829190ec8fec3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189829190ec8fec3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:40.969264835 +0000 UTC m=+0.954577622,LastTimestamp:2026-02-27 17:00:41.084056766 +0000 UTC m=+1.069369553,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.678159 4700 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189829190ec9b04f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189829190ec9b04f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:40.969310287 +0000 UTC m=+0.954623074,LastTimestamp:2026-02-27 17:00:41.084145348 +0000 UTC m=+1.069458125,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.685427 4700 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189829190eca0649\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189829190eca0649 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:40.969332297 +0000 UTC m=+0.954645084,LastTimestamp:2026-02-27 17:00:41.084164759 +0000 UTC m=+1.069477546,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.692829 4700 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189829190ec8fec3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189829190ec8fec3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:40.969264835 +0000 UTC m=+0.954577622,LastTimestamp:2026-02-27 17:00:41.087673103 +0000 UTC m=+1.072985880,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.700136 4700 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189829190ec9b04f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189829190ec9b04f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:40.969310287 +0000 UTC m=+0.954623074,LastTimestamp:2026-02-27 17:00:41.087702524 +0000 UTC m=+1.073015311,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.707374 4700 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189829190eca0649\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189829190eca0649 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:40.969332297 +0000 UTC m=+0.954645084,LastTimestamp:2026-02-27 17:00:41.087721675 +0000 UTC m=+1.073034452,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.714770 4700 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189829190ec8fec3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189829190ec8fec3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:40.969264835 +0000 UTC m=+0.954577622,LastTimestamp:2026-02-27 17:00:41.087793247 +0000 UTC m=+1.073106034,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.721858 4700 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189829190ec9b04f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189829190ec9b04f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:40.969310287 +0000 UTC m=+0.954623074,LastTimestamp:2026-02-27 17:00:41.087914171 +0000 UTC m=+1.073226958,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.729144 4700 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189829190eca0649\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189829190eca0649 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:40.969332297 +0000 UTC m=+0.954645084,LastTimestamp:2026-02-27 17:00:41.087984883 +0000 UTC m=+1.073297670,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.736656 4700 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189829190ec8fec3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189829190ec8fec3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:40.969264835 +0000 UTC m=+0.954577622,LastTimestamp:2026-02-27 17:00:41.090826237 +0000 UTC m=+1.076139014,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.744583 4700 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189829190ec9b04f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189829190ec9b04f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:40.969310287 +0000 UTC m=+0.954623074,LastTimestamp:2026-02-27 17:00:41.090891429 +0000 UTC m=+1.076204206,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.751672 4700 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189829190eca0649\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189829190eca0649 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:40.969332297 +0000 UTC m=+0.954645084,LastTimestamp:2026-02-27 17:00:41.09091033 +0000 UTC m=+1.076223107,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.755682 4700 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189829190ec8fec3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189829190ec8fec3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:40.969264835 +0000 UTC m=+0.954577622,LastTimestamp:2026-02-27 17:00:41.091100256 +0000 UTC m=+1.076413043,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.761285 4700 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189829190ec9b04f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189829190ec9b04f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:40.969310287 +0000 UTC m=+0.954623074,LastTimestamp:2026-02-27 17:00:41.091148367 +0000 UTC m=+1.076461154,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.766581 4700 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189829190eca0649\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189829190eca0649 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:40.969332297 +0000 UTC m=+0.954645084,LastTimestamp:2026-02-27 17:00:41.09124948 +0000 UTC m=+1.076562277,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.769688 4700 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189829190ec8fec3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189829190ec8fec3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:40.969264835 +0000 UTC m=+0.954577622,LastTimestamp:2026-02-27 17:00:41.093776355 +0000 UTC m=+1.079089132,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.772712 4700 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189829190ec9b04f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189829190ec9b04f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:40.969310287 +0000 UTC m=+0.954623074,LastTimestamp:2026-02-27 17:00:41.093809216 +0000 UTC m=+1.079122003,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.776956 4700 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189829190eca0649\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189829190eca0649 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:40.969332297 +0000 UTC m=+0.954645084,LastTimestamp:2026-02-27 17:00:41.093835607 +0000 UTC m=+1.079148384,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.781164 4700 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189829190ec8fec3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189829190ec8fec3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:40.969264835 +0000 UTC m=+0.954577622,LastTimestamp:2026-02-27 17:00:41.095510417 +0000 UTC m=+1.080823194,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.784925 4700 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189829190ec9b04f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189829190ec9b04f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:40.969310287 +0000 UTC m=+0.954623074,LastTimestamp:2026-02-27 17:00:41.095535488 +0000 UTC m=+1.080848265,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.789423 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189829193061e5e0 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:41.5329336 +0000 UTC m=+1.518246387,LastTimestamp:2026-02-27 17:00:41.5329336 +0000 UTC m=+1.518246387,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.794005 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.1898291930633ac9 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:41.533020873 +0000 UTC m=+1.518333660,LastTimestamp:2026-02-27 17:00:41.533020873 +0000 UTC m=+1.518333660,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.798239 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189829193105a43c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:41.5436647 +0000 UTC m=+1.528977477,LastTimestamp:2026-02-27 17:00:41.5436647 +0000 UTC m=+1.528977477,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.802334 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.1898291931b5ec4c openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:41.555217484 +0000 UTC m=+1.540530271,LastTimestamp:2026-02-27 17:00:41.555217484 +0000 UTC m=+1.540530271,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.806161 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.1898291931bab1aa openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:41.555530154 +0000 UTC m=+1.540842941,LastTimestamp:2026-02-27 17:00:41.555530154 +0000 UTC m=+1.540842941,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.810047 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189829195ab0925f openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Created,Message:Created container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:42.242732639 +0000 UTC m=+2.228045416,LastTimestamp:2026-02-27 17:00:42.242732639 +0000 UTC m=+2.228045416,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.813969 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189829195aca24f9 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Created,Message:Created container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:42.244408569 +0000 UTC m=+2.229721346,LastTimestamp:2026-02-27 17:00:42.244408569 +0000 UTC m=+2.229721346,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.818398 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189829195acecb3e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:42.244713278 +0000 UTC m=+2.230026056,LastTimestamp:2026-02-27 17:00:42.244713278 +0000 UTC m=+2.230026056,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.822845 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189829195ae0b3a3 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:42.245886883 +0000 UTC m=+2.231199670,LastTimestamp:2026-02-27 17:00:42.245886883 +0000 UTC m=+2.231199670,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.827661 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189829195ae20008 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:42.245971976 +0000 UTC m=+2.231284763,LastTimestamp:2026-02-27 17:00:42.245971976 +0000 UTC m=+2.231284763,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.831528 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189829195ba9d36c openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Started,Message:Started container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:42.259067756 +0000 UTC m=+2.244380533,LastTimestamp:2026-02-27 17:00:42.259067756 +0000 UTC m=+2.244380533,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.835410 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189829195bb7c983 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Started,Message:Started container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:42.259982723 +0000 UTC m=+2.245295500,LastTimestamp:2026-02-27 17:00:42.259982723 +0000 UTC m=+2.245295500,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.838884 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189829195bd8dbb4 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:42.262150068 +0000 UTC m=+2.247462845,LastTimestamp:2026-02-27 17:00:42.262150068 +0000 UTC m=+2.247462845,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.843420 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189829195bd9bce8 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:42.26220772 +0000 UTC m=+2.247520497,LastTimestamp:2026-02-27 17:00:42.26220772 +0000 UTC m=+2.247520497,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.848944 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189829195bea9acf openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:42.263313103 +0000 UTC m=+2.248625880,LastTimestamp:2026-02-27 17:00:42.263313103 +0000 UTC m=+2.248625880,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.855586 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189829195c4f83c4 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:42.26992634 +0000 UTC m=+2.255239127,LastTimestamp:2026-02-27 17:00:42.26992634 +0000 UTC m=+2.255239127,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.862005 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18982919717e9585 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:42.625332613 +0000 UTC m=+2.610645360,LastTimestamp:2026-02-27 17:00:42.625332613 +0000 UTC m=+2.610645360,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.868243 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18982919726e757e openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:42.641053054 +0000 UTC m=+2.626365841,LastTimestamp:2026-02-27 17:00:42.641053054 +0000 UTC m=+2.626365841,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.872350 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189829197285c3b8 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:42.642580408 +0000 UTC m=+2.627893155,LastTimestamp:2026-02-27 17:00:42.642580408 +0000 UTC m=+2.627893155,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.878214 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.1898291980757d66 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Created,Message:Created container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:42.876394854 +0000 UTC m=+2.861707631,LastTimestamp:2026-02-27 17:00:42.876394854 +0000 UTC m=+2.861707631,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.885668 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.1898291981438573 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Started,Message:Started container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:42.889897331 +0000 UTC m=+2.875210118,LastTimestamp:2026-02-27 17:00:42.889897331 +0000 UTC m=+2.875210118,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.891811 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189829198154e2da openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:42.891035354 +0000 UTC m=+2.876348131,LastTimestamp:2026-02-27 17:00:42.891035354 +0000 UTC m=+2.876348131,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.898101 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898291988629d0a openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:43.009375498 +0000 UTC m=+2.994688245,LastTimestamp:2026-02-27 17:00:43.009375498 +0000 UTC m=+2.994688245,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.902608 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.1898291988949f70 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:43.012652912 +0000 UTC m=+2.997965659,LastTimestamp:2026-02-27 17:00:43.012652912 +0000 UTC m=+2.997965659,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: I0227 17:01:05.911201 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.911882 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.1898291988eb65f1 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:43.018339825 +0000 UTC m=+3.003652582,LastTimestamp:2026-02-27 17:00:43.018339825 +0000 UTC m=+3.003652582,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.914736 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18982919890b82af openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:43.020444335 +0000 UTC m=+3.005757102,LastTimestamp:2026-02-27 17:00:43.020444335 +0000 UTC m=+3.005757102,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.919072 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189829198e6e8fcd openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Created,Message:Created container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:43.110821837 +0000 UTC m=+3.096134594,LastTimestamp:2026-02-27 17:00:43.110821837 +0000 UTC m=+3.096134594,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.922434 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189829198f844b43 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Started,Message:Started container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:43.129023299 +0000 UTC m=+3.114336046,LastTimestamp:2026-02-27 17:00:43.129023299 +0000 UTC m=+3.114336046,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.926546 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.1898291999319e53 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Created,Message:Created container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:43.291377235 +0000 UTC m=+3.276689992,LastTimestamp:2026-02-27 17:00:43.291377235 +0000 UTC m=+3.276689992,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.932696 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18982919994401c4 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Created,Message:Created container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:43.29258234 +0000 UTC m=+3.277895097,LastTimestamp:2026-02-27 17:00:43.29258234 +0000 UTC m=+3.277895097,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.938813 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.18982919995ce605 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:43.294213637 +0000 UTC m=+3.279526404,LastTimestamp:2026-02-27 17:00:43.294213637 +0000 UTC m=+3.279526404,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.944890 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898291999a28ac6 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Created,Message:Created container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:43.298777798 +0000 UTC m=+3.284090555,LastTimestamp:2026-02-27 17:00:43.298777798 +0000 UTC m=+3.284090555,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.951874 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.1898291999ce22d2 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Started,Message:Started container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:43.30163477 +0000 UTC m=+3.286947527,LastTimestamp:2026-02-27 17:00:43.30163477 +0000 UTC m=+3.286947527,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.959227 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.1898291999ff361c openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:43.304850972 +0000 UTC m=+3.290163719,LastTimestamp:2026-02-27 17:00:43.304850972 +0000 UTC m=+3.290163719,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.966286 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189829199a374aed openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Started,Message:Started container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:43.308526317 +0000 UTC m=+3.293839064,LastTimestamp:2026-02-27 17:00:43.308526317 +0000 UTC m=+3.293839064,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.975011 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189829199a44d38f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:43.309413263 +0000 UTC m=+3.294726010,LastTimestamp:2026-02-27 17:00:43.309413263 +0000 UTC m=+3.294726010,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.978236 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189829199b3fba44 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:43.325856324 +0000 UTC m=+3.311169071,LastTimestamp:2026-02-27 17:00:43.325856324 +0000 UTC m=+3.311169071,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.983836 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189829199bbf5e02 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Started,Message:Started container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:43.334221314 +0000 UTC m=+3.319534061,LastTimestamp:2026-02-27 17:00:43.334221314 +0000 UTC m=+3.319534061,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:05 crc kubenswrapper[4700]: E0227 17:01:05.998757 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18982919a659bfe4 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Created,Message:Created container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:43.512111076 +0000 UTC m=+3.497423823,LastTimestamp:2026-02-27 17:00:43.512111076 +0000 UTC m=+3.497423823,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.007566 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18982919a6960c09 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Created,Message:Created container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:43.516062729 +0000 UTC m=+3.501375476,LastTimestamp:2026-02-27 17:00:43.516062729 +0000 UTC m=+3.501375476,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.014966 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18982919a73eca81 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Started,Message:Started container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:43.527121537 +0000 UTC m=+3.512434284,LastTimestamp:2026-02-27 17:00:43.527121537 +0000 UTC m=+3.512434284,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.024164 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18982919a751abe2 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:43.528358882 +0000 UTC m=+3.513671629,LastTimestamp:2026-02-27 17:00:43.528358882 +0000 UTC m=+3.513671629,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.033252 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18982919a759ef74 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Started,Message:Started container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:43.528900468 +0000 UTC m=+3.514213215,LastTimestamp:2026-02-27 17:00:43.528900468 +0000 UTC m=+3.514213215,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.041279 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18982919a76b36a0 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:43.5300328 +0000 UTC m=+3.515345557,LastTimestamp:2026-02-27 17:00:43.5300328 +0000 UTC m=+3.515345557,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.049386 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18982919b575d8bf openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Created,Message:Created container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:43.765610687 +0000 UTC m=+3.750923444,LastTimestamp:2026-02-27 17:00:43.765610687 +0000 UTC m=+3.750923444,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.058017 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18982919b598c190 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Created,Message:Created container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:43.767898512 +0000 UTC m=+3.753211269,LastTimestamp:2026-02-27 17:00:43.767898512 +0000 UTC m=+3.753211269,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.065627 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18982919b6523a57 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Started,Message:Started container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:43.780053591 +0000 UTC m=+3.765366378,LastTimestamp:2026-02-27 17:00:43.780053591 +0000 UTC m=+3.765366378,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.073529 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18982919b66533dc openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:43.781297116 +0000 UTC m=+3.766609873,LastTimestamp:2026-02-27 17:00:43.781297116 +0000 UTC m=+3.766609873,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.080495 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18982919b6a1da34 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Started,Message:Started container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:43.78527186 +0000 UTC m=+3.770584617,LastTimestamp:2026-02-27 17:00:43.78527186 +0000 UTC m=+3.770584617,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.084286 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18982919c5a9ffcd openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:44.037464013 +0000 UTC m=+4.022776790,LastTimestamp:2026-02-27 17:00:44.037464013 +0000 UTC m=+4.022776790,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.089105 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18982919c5c235a3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Created,Message:Created container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:44.039050659 +0000 UTC m=+4.024363446,LastTimestamp:2026-02-27 17:00:44.039050659 +0000 UTC m=+4.024363446,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.093136 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18982919c6b0c3c9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Started,Message:Started container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:44.054684617 +0000 UTC m=+4.039997404,LastTimestamp:2026-02-27 17:00:44.054684617 +0000 UTC m=+4.039997404,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.099448 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18982919c6dff382 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:44.057777026 +0000 UTC m=+4.043089773,LastTimestamp:2026-02-27 17:00:44.057777026 +0000 UTC m=+4.043089773,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.106715 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18982919d3efb4a9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:44.276913321 +0000 UTC m=+4.262226068,LastTimestamp:2026-02-27 17:00:44.276913321 +0000 UTC m=+4.262226068,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.113506 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18982919d401faeb openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Created,Message:Created container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:44.278110955 +0000 UTC m=+4.263423702,LastTimestamp:2026-02-27 17:00:44.278110955 +0000 UTC m=+4.263423702,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.123326 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18982919d4c3f9ac openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:44.29082462 +0000 UTC m=+4.276137407,LastTimestamp:2026-02-27 17:00:44.29082462 +0000 UTC m=+4.276137407,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.130393 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18982919d4ceccd8 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Started,Message:Started container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:44.29153404 +0000 UTC m=+4.276846787,LastTimestamp:2026-02-27 17:00:44.29153404 +0000 UTC m=+4.276846787,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.140526 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898291a0214f00e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:45.051105294 +0000 UTC m=+5.036418071,LastTimestamp:2026-02-27 17:00:45.051105294 +0000 UTC m=+5.036418071,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.148993 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898291a1130298a openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Created,Message:Created container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:45.304547722 +0000 UTC m=+5.289860499,LastTimestamp:2026-02-27 17:00:45.304547722 +0000 UTC m=+5.289860499,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.159518 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898291a11de6ac0 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Started,Message:Started container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:45.31596768 +0000 UTC m=+5.301280467,LastTimestamp:2026-02-27 17:00:45.31596768 +0000 UTC m=+5.301280467,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.167736 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898291a11fe5b96 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:45.31806095 +0000 UTC m=+5.303373727,LastTimestamp:2026-02-27 17:00:45.31806095 +0000 UTC m=+5.303373727,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.176077 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898291a2250de50 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Created,Message:Created container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:45.591903824 +0000 UTC m=+5.577216581,LastTimestamp:2026-02-27 17:00:45.591903824 +0000 UTC m=+5.577216581,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.183766 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898291a2361606e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Started,Message:Started container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:45.609762926 +0000 UTC m=+5.595075713,LastTimestamp:2026-02-27 17:00:45.609762926 +0000 UTC m=+5.595075713,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.191016 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898291a237870f9 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:45.611274489 +0000 UTC m=+5.596587276,LastTimestamp:2026-02-27 17:00:45.611274489 +0000 UTC m=+5.596587276,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.198848 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898291a341e008b openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Created,Message:Created container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:45.890560139 +0000 UTC m=+5.875872916,LastTimestamp:2026-02-27 17:00:45.890560139 +0000 UTC m=+5.875872916,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.206616 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898291a35170b43 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Started,Message:Started container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:45.906881347 +0000 UTC m=+5.892194134,LastTimestamp:2026-02-27 17:00:45.906881347 +0000 UTC m=+5.892194134,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.215373 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898291a352dba80 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:45.908368 +0000 UTC m=+5.893680777,LastTimestamp:2026-02-27 17:00:45.908368 +0000 UTC m=+5.893680777,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.224523 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898291a45ebf9b4 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Created,Message:Created container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:46.189271476 +0000 UTC m=+6.174584253,LastTimestamp:2026-02-27 17:00:46.189271476 +0000 UTC m=+6.174584253,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.231577 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898291a46f6ae45 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Started,Message:Started container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:46.206750277 +0000 UTC m=+6.192063054,LastTimestamp:2026-02-27 17:00:46.206750277 +0000 UTC m=+6.192063054,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.238409 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898291a47128eb0 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:46.2085772 +0000 UTC m=+6.193889977,LastTimestamp:2026-02-27 17:00:46.2085772 +0000 UTC m=+6.193889977,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.247200 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898291a563699e5 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Created,Message:Created container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:46.462597605 +0000 UTC m=+6.447910392,LastTimestamp:2026-02-27 17:00:46.462597605 +0000 UTC m=+6.447910392,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.254301 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1898291a57f07c2a openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Started,Message:Started container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:46.491556906 +0000 UTC m=+6.476869643,LastTimestamp:2026-02-27 17:00:46.491556906 +0000 UTC m=+6.476869643,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.267315 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Feb 27 17:01:06 crc kubenswrapper[4700]: &Event{ObjectMeta:{kube-apiserver-crc.1898291c7f367296 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:ProbeError,Message:Liveness probe error: Get "https://192.168.126.11:17697/healthz": read tcp 192.168.126.11:53758->192.168.126.11:17697: read: connection reset by peer Feb 27 17:01:06 crc kubenswrapper[4700]: body: Feb 27 17:01:06 crc kubenswrapper[4700]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:55.74038799 +0000 UTC m=+15.725700777,LastTimestamp:2026-02-27 17:00:55.74038799 +0000 UTC m=+15.725700777,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 27 17:01:06 crc kubenswrapper[4700]: > Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.274107 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1898291c7f381491 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Unhealthy,Message:Liveness probe failed: Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:53758->192.168.126.11:17697: read: connection reset by peer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:55.740494993 +0000 UTC m=+15.725807770,LastTimestamp:2026-02-27 17:00:55.740494993 +0000 UTC m=+15.725807770,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.281445 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Feb 27 17:01:06 crc kubenswrapper[4700]: &Event{ObjectMeta:{kube-apiserver-crc.1898291c7f7b0059 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Feb 27 17:01:06 crc kubenswrapper[4700]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 27 17:01:06 crc kubenswrapper[4700]: Feb 27 17:01:06 crc kubenswrapper[4700]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:55.744880729 +0000 UTC m=+15.730193516,LastTimestamp:2026-02-27 17:00:55.744880729 +0000 UTC m=+15.730193516,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 27 17:01:06 crc kubenswrapper[4700]: > Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.288417 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1898291c7f7c005f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:55.744946271 +0000 UTC m=+15.730259058,LastTimestamp:2026-02-27 17:00:55.744946271 +0000 UTC m=+15.730259058,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.290644 4700 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.1898291c7f7b0059\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Feb 27 17:01:06 crc kubenswrapper[4700]: &Event{ObjectMeta:{kube-apiserver-crc.1898291c7f7b0059 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Feb 27 17:01:06 crc kubenswrapper[4700]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 27 17:01:06 crc kubenswrapper[4700]: Feb 27 17:01:06 crc kubenswrapper[4700]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:55.744880729 +0000 UTC m=+15.730193516,LastTimestamp:2026-02-27 17:00:55.754440893 +0000 UTC m=+15.739753650,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 27 17:01:06 crc kubenswrapper[4700]: > Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.295197 4700 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.1898291c7f7c005f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1898291c7f7c005f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:55.744946271 +0000 UTC m=+15.730259058,LastTimestamp:2026-02-27 17:00:55.754496395 +0000 UTC m=+15.739809152,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.302605 4700 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.18982919c6dff382\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18982919c6dff382 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:44.057777026 +0000 UTC m=+4.043089773,LastTimestamp:2026-02-27 17:00:56.098258084 +0000 UTC m=+16.083570871,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.309385 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Feb 27 17:01:06 crc kubenswrapper[4700]: &Event{ObjectMeta:{kube-controller-manager-crc.1898291ca7c33ade openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Feb 27 17:01:06 crc kubenswrapper[4700]: body: Feb 27 17:01:06 crc kubenswrapper[4700]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:56.420702942 +0000 UTC m=+16.406015729,LastTimestamp:2026-02-27 17:00:56.420702942 +0000 UTC m=+16.406015729,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 27 17:01:06 crc kubenswrapper[4700]: > Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.316552 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.1898291ca7c423b9 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:56.420762553 +0000 UTC m=+16.406075340,LastTimestamp:2026-02-27 17:00:56.420762553 +0000 UTC m=+16.406075340,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: I0227 17:01:06.408875 4700 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 17:01:06 crc kubenswrapper[4700]: I0227 17:01:06.409059 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.420501 4700 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.1898291ca7c33ade\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Feb 27 17:01:06 crc kubenswrapper[4700]: &Event{ObjectMeta:{kube-controller-manager-crc.1898291ca7c33ade openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Feb 27 17:01:06 crc kubenswrapper[4700]: body: Feb 27 17:01:06 crc kubenswrapper[4700]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:56.420702942 +0000 UTC m=+16.406015729,LastTimestamp:2026-02-27 17:01:06.409018244 +0000 UTC m=+26.394331031,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 27 17:01:06 crc kubenswrapper[4700]: > Feb 27 17:01:06 crc kubenswrapper[4700]: E0227 17:01:06.427754 4700 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.1898291ca7c423b9\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.1898291ca7c423b9 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:56.420762553 +0000 UTC m=+16.406075340,LastTimestamp:2026-02-27 17:01:06.409109636 +0000 UTC m=+26.394422453,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:06 crc kubenswrapper[4700]: I0227 17:01:06.913513 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:07 crc kubenswrapper[4700]: W0227 17:01:07.464642 4700 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Feb 27 17:01:07 crc kubenswrapper[4700]: E0227 17:01:07.464733 4700 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" Feb 27 17:01:07 crc kubenswrapper[4700]: I0227 17:01:07.916830 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:08 crc kubenswrapper[4700]: I0227 17:01:08.914354 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:09 crc kubenswrapper[4700]: I0227 17:01:09.144171 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:01:09 crc kubenswrapper[4700]: I0227 17:01:09.146284 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:01:09 crc kubenswrapper[4700]: I0227 17:01:09.146352 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:01:09 crc kubenswrapper[4700]: I0227 17:01:09.146375 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:01:09 crc kubenswrapper[4700]: I0227 17:01:09.146457 4700 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 17:01:09 crc kubenswrapper[4700]: E0227 17:01:09.154536 4700 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Feb 27 17:01:09 crc kubenswrapper[4700]: E0227 17:01:09.166003 4700 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Feb 27 17:01:09 crc kubenswrapper[4700]: I0227 17:01:09.913830 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:09 crc kubenswrapper[4700]: W0227 17:01:09.946349 4700 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Feb 27 17:01:09 crc kubenswrapper[4700]: E0227 17:01:09.946446 4700 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" Feb 27 17:01:10 crc kubenswrapper[4700]: W0227 17:01:10.454718 4700 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes "crc" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Feb 27 17:01:10 crc kubenswrapper[4700]: E0227 17:01:10.454852 4700 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" Feb 27 17:01:10 crc kubenswrapper[4700]: I0227 17:01:10.912788 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:11 crc kubenswrapper[4700]: E0227 17:01:11.088143 4700 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 27 17:01:11 crc kubenswrapper[4700]: I0227 17:01:11.912941 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:12 crc kubenswrapper[4700]: I0227 17:01:12.911940 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:13 crc kubenswrapper[4700]: I0227 17:01:13.568527 4700 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": read tcp 192.168.126.11:37950->192.168.126.11:10357: read: connection reset by peer" start-of-body= Feb 27 17:01:13 crc kubenswrapper[4700]: I0227 17:01:13.568619 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": read tcp 192.168.126.11:37950->192.168.126.11:10357: read: connection reset by peer" Feb 27 17:01:13 crc kubenswrapper[4700]: I0227 17:01:13.568700 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 17:01:13 crc kubenswrapper[4700]: I0227 17:01:13.568920 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:01:13 crc kubenswrapper[4700]: I0227 17:01:13.571426 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:01:13 crc kubenswrapper[4700]: I0227 17:01:13.571787 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:01:13 crc kubenswrapper[4700]: I0227 17:01:13.571815 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:01:13 crc kubenswrapper[4700]: I0227 17:01:13.573023 4700 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"523f47c5e3ee759e5c1f7a782c10d08a2eb10fba3833c7fe40d6105867420b77"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Feb 27 17:01:13 crc kubenswrapper[4700]: I0227 17:01:13.573301 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://523f47c5e3ee759e5c1f7a782c10d08a2eb10fba3833c7fe40d6105867420b77" gracePeriod=30 Feb 27 17:01:13 crc kubenswrapper[4700]: E0227 17:01:13.578598 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Feb 27 17:01:13 crc kubenswrapper[4700]: &Event{ObjectMeta:{kube-controller-manager-crc.18982920a5db46f6 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": read tcp 192.168.126.11:37950->192.168.126.11:10357: read: connection reset by peer Feb 27 17:01:13 crc kubenswrapper[4700]: body: Feb 27 17:01:13 crc kubenswrapper[4700]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:01:13.568593654 +0000 UTC m=+33.553906431,LastTimestamp:2026-02-27 17:01:13.568593654 +0000 UTC m=+33.553906431,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 27 17:01:13 crc kubenswrapper[4700]: > Feb 27 17:01:13 crc kubenswrapper[4700]: E0227 17:01:13.585900 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18982920a5dc429b openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": read tcp 192.168.126.11:37950->192.168.126.11:10357: read: connection reset by peer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:01:13.568658075 +0000 UTC m=+33.553970852,LastTimestamp:2026-02-27 17:01:13.568658075 +0000 UTC m=+33.553970852,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:13 crc kubenswrapper[4700]: E0227 17:01:13.594001 4700 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18982920a622a908 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Killing,Message:Container cluster-policy-controller failed startup probe, will be restarted,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:01:13.573271816 +0000 UTC m=+33.558584603,LastTimestamp:2026-02-27 17:01:13.573271816 +0000 UTC m=+33.558584603,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:13 crc kubenswrapper[4700]: E0227 17:01:13.608930 4700 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189829195bd8dbb4\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189829195bd8dbb4 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:42.262150068 +0000 UTC m=+2.247462845,LastTimestamp:2026-02-27 17:01:13.600856536 +0000 UTC m=+33.586169323,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:13 crc kubenswrapper[4700]: E0227 17:01:13.850519 4700 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.18982919717e9585\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18982919717e9585 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:42.625332613 +0000 UTC m=+2.610645360,LastTimestamp:2026-02-27 17:01:13.842005421 +0000 UTC m=+33.827318208,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:13 crc kubenswrapper[4700]: E0227 17:01:13.863504 4700 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.18982919726e757e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18982919726e757e openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:42.641053054 +0000 UTC m=+2.626365841,LastTimestamp:2026-02-27 17:01:13.856041448 +0000 UTC m=+33.841354235,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:13 crc kubenswrapper[4700]: I0227 17:01:13.914367 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:14 crc kubenswrapper[4700]: I0227 17:01:14.168957 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Feb 27 17:01:14 crc kubenswrapper[4700]: I0227 17:01:14.170659 4700 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="523f47c5e3ee759e5c1f7a782c10d08a2eb10fba3833c7fe40d6105867420b77" exitCode=255 Feb 27 17:01:14 crc kubenswrapper[4700]: I0227 17:01:14.170722 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"523f47c5e3ee759e5c1f7a782c10d08a2eb10fba3833c7fe40d6105867420b77"} Feb 27 17:01:14 crc kubenswrapper[4700]: I0227 17:01:14.170794 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"fb6c59b5f96fb98a2defc2a18bab3d015046d2352a51c1f408bb020c43d5f9f3"} Feb 27 17:01:14 crc kubenswrapper[4700]: I0227 17:01:14.170931 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:01:14 crc kubenswrapper[4700]: I0227 17:01:14.172189 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:01:14 crc kubenswrapper[4700]: I0227 17:01:14.172243 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:01:14 crc kubenswrapper[4700]: I0227 17:01:14.172265 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:01:14 crc kubenswrapper[4700]: I0227 17:01:14.912379 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:15 crc kubenswrapper[4700]: I0227 17:01:15.913302 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:16 crc kubenswrapper[4700]: I0227 17:01:16.154910 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:01:16 crc kubenswrapper[4700]: I0227 17:01:16.156397 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:01:16 crc kubenswrapper[4700]: I0227 17:01:16.156487 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:01:16 crc kubenswrapper[4700]: I0227 17:01:16.156509 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:01:16 crc kubenswrapper[4700]: I0227 17:01:16.156548 4700 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 17:01:16 crc kubenswrapper[4700]: E0227 17:01:16.163616 4700 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Feb 27 17:01:16 crc kubenswrapper[4700]: E0227 17:01:16.173394 4700 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Feb 27 17:01:16 crc kubenswrapper[4700]: I0227 17:01:16.912934 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:17 crc kubenswrapper[4700]: W0227 17:01:17.312802 4700 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:17 crc kubenswrapper[4700]: E0227 17:01:17.312891 4700 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" Feb 27 17:01:17 crc kubenswrapper[4700]: I0227 17:01:17.913013 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:18 crc kubenswrapper[4700]: I0227 17:01:18.576766 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 17:01:18 crc kubenswrapper[4700]: I0227 17:01:18.577004 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:01:18 crc kubenswrapper[4700]: I0227 17:01:18.578604 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:01:18 crc kubenswrapper[4700]: I0227 17:01:18.578649 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:01:18 crc kubenswrapper[4700]: I0227 17:01:18.578666 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:01:18 crc kubenswrapper[4700]: I0227 17:01:18.913681 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:18 crc kubenswrapper[4700]: I0227 17:01:18.980349 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:01:18 crc kubenswrapper[4700]: I0227 17:01:18.982145 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:01:18 crc kubenswrapper[4700]: I0227 17:01:18.982216 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:01:18 crc kubenswrapper[4700]: I0227 17:01:18.982236 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:01:18 crc kubenswrapper[4700]: I0227 17:01:18.983275 4700 scope.go:117] "RemoveContainer" containerID="d3dcb3b876b32e2e8dc9c1c45bffa575532863bacee7c89f103fa057a023c109" Feb 27 17:01:19 crc kubenswrapper[4700]: I0227 17:01:19.911976 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:20 crc kubenswrapper[4700]: I0227 17:01:20.192247 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Feb 27 17:01:20 crc kubenswrapper[4700]: I0227 17:01:20.193115 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 27 17:01:20 crc kubenswrapper[4700]: I0227 17:01:20.196191 4700 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a785c0a3f8cf91c386fb0e3fa63d0bd9e57e702e96ba3e20011010d691066c8f" exitCode=255 Feb 27 17:01:20 crc kubenswrapper[4700]: I0227 17:01:20.196252 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"a785c0a3f8cf91c386fb0e3fa63d0bd9e57e702e96ba3e20011010d691066c8f"} Feb 27 17:01:20 crc kubenswrapper[4700]: I0227 17:01:20.196313 4700 scope.go:117] "RemoveContainer" containerID="d3dcb3b876b32e2e8dc9c1c45bffa575532863bacee7c89f103fa057a023c109" Feb 27 17:01:20 crc kubenswrapper[4700]: I0227 17:01:20.196582 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:01:20 crc kubenswrapper[4700]: I0227 17:01:20.197866 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:01:20 crc kubenswrapper[4700]: I0227 17:01:20.197914 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:01:20 crc kubenswrapper[4700]: I0227 17:01:20.197930 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:01:20 crc kubenswrapper[4700]: I0227 17:01:20.198794 4700 scope.go:117] "RemoveContainer" containerID="a785c0a3f8cf91c386fb0e3fa63d0bd9e57e702e96ba3e20011010d691066c8f" Feb 27 17:01:20 crc kubenswrapper[4700]: E0227 17:01:20.199020 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 17:01:20 crc kubenswrapper[4700]: W0227 17:01:20.670046 4700 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Feb 27 17:01:20 crc kubenswrapper[4700]: E0227 17:01:20.670127 4700 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" Feb 27 17:01:20 crc kubenswrapper[4700]: I0227 17:01:20.912092 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:21 crc kubenswrapper[4700]: E0227 17:01:21.088943 4700 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 27 17:01:21 crc kubenswrapper[4700]: I0227 17:01:21.203379 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Feb 27 17:01:21 crc kubenswrapper[4700]: I0227 17:01:21.521011 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:01:21 crc kubenswrapper[4700]: I0227 17:01:21.521275 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:01:21 crc kubenswrapper[4700]: I0227 17:01:21.522968 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:01:21 crc kubenswrapper[4700]: I0227 17:01:21.523047 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:01:21 crc kubenswrapper[4700]: I0227 17:01:21.523068 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:01:21 crc kubenswrapper[4700]: I0227 17:01:21.524024 4700 scope.go:117] "RemoveContainer" containerID="a785c0a3f8cf91c386fb0e3fa63d0bd9e57e702e96ba3e20011010d691066c8f" Feb 27 17:01:21 crc kubenswrapper[4700]: E0227 17:01:21.524418 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 17:01:21 crc kubenswrapper[4700]: I0227 17:01:21.913804 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:22 crc kubenswrapper[4700]: I0227 17:01:22.914006 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:23 crc kubenswrapper[4700]: I0227 17:01:23.164358 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:01:23 crc kubenswrapper[4700]: I0227 17:01:23.166299 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:01:23 crc kubenswrapper[4700]: I0227 17:01:23.166368 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:01:23 crc kubenswrapper[4700]: I0227 17:01:23.166389 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:01:23 crc kubenswrapper[4700]: I0227 17:01:23.166433 4700 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 17:01:23 crc kubenswrapper[4700]: E0227 17:01:23.173399 4700 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Feb 27 17:01:23 crc kubenswrapper[4700]: E0227 17:01:23.181184 4700 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Feb 27 17:01:23 crc kubenswrapper[4700]: I0227 17:01:23.409258 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 17:01:23 crc kubenswrapper[4700]: I0227 17:01:23.409547 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:01:23 crc kubenswrapper[4700]: I0227 17:01:23.411215 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:01:23 crc kubenswrapper[4700]: I0227 17:01:23.411261 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:01:23 crc kubenswrapper[4700]: I0227 17:01:23.411279 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:01:23 crc kubenswrapper[4700]: I0227 17:01:23.914214 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:24 crc kubenswrapper[4700]: I0227 17:01:24.576545 4700 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:01:24 crc kubenswrapper[4700]: I0227 17:01:24.576774 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:01:24 crc kubenswrapper[4700]: I0227 17:01:24.577943 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:01:24 crc kubenswrapper[4700]: I0227 17:01:24.577972 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:01:24 crc kubenswrapper[4700]: I0227 17:01:24.577981 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:01:24 crc kubenswrapper[4700]: I0227 17:01:24.578454 4700 scope.go:117] "RemoveContainer" containerID="a785c0a3f8cf91c386fb0e3fa63d0bd9e57e702e96ba3e20011010d691066c8f" Feb 27 17:01:24 crc kubenswrapper[4700]: E0227 17:01:24.578627 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 17:01:24 crc kubenswrapper[4700]: I0227 17:01:24.912787 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:25 crc kubenswrapper[4700]: I0227 17:01:25.914030 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:26 crc kubenswrapper[4700]: I0227 17:01:26.409309 4700 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 17:01:26 crc kubenswrapper[4700]: I0227 17:01:26.409416 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 17:01:26 crc kubenswrapper[4700]: E0227 17:01:26.417072 4700 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.1898291ca7c33ade\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Feb 27 17:01:26 crc kubenswrapper[4700]: &Event{ObjectMeta:{kube-controller-manager-crc.1898291ca7c33ade openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Feb 27 17:01:26 crc kubenswrapper[4700]: body: Feb 27 17:01:26 crc kubenswrapper[4700]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:56.420702942 +0000 UTC m=+16.406015729,LastTimestamp:2026-02-27 17:01:26.409390974 +0000 UTC m=+46.394703761,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 27 17:01:26 crc kubenswrapper[4700]: > Feb 27 17:01:26 crc kubenswrapper[4700]: E0227 17:01:26.424314 4700 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.1898291ca7c423b9\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.1898291ca7c423b9 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:00:56.420762553 +0000 UTC m=+16.406075340,LastTimestamp:2026-02-27 17:01:26.409449355 +0000 UTC m=+46.394762132,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:01:26 crc kubenswrapper[4700]: I0227 17:01:26.914253 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:27 crc kubenswrapper[4700]: W0227 17:01:27.832491 4700 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes "crc" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Feb 27 17:01:27 crc kubenswrapper[4700]: E0227 17:01:27.832563 4700 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" Feb 27 17:01:27 crc kubenswrapper[4700]: I0227 17:01:27.912168 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:28 crc kubenswrapper[4700]: I0227 17:01:28.911448 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:29 crc kubenswrapper[4700]: I0227 17:01:29.913996 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:30 crc kubenswrapper[4700]: I0227 17:01:30.173916 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:01:30 crc kubenswrapper[4700]: I0227 17:01:30.175140 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:01:30 crc kubenswrapper[4700]: I0227 17:01:30.175173 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:01:30 crc kubenswrapper[4700]: I0227 17:01:30.175183 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:01:30 crc kubenswrapper[4700]: I0227 17:01:30.175204 4700 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 17:01:30 crc kubenswrapper[4700]: E0227 17:01:30.181518 4700 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Feb 27 17:01:30 crc kubenswrapper[4700]: E0227 17:01:30.187811 4700 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Feb 27 17:01:30 crc kubenswrapper[4700]: I0227 17:01:30.914725 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:31 crc kubenswrapper[4700]: E0227 17:01:31.089070 4700 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 27 17:01:31 crc kubenswrapper[4700]: I0227 17:01:31.914646 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:32 crc kubenswrapper[4700]: I0227 17:01:32.914210 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:33 crc kubenswrapper[4700]: I0227 17:01:33.415666 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 17:01:33 crc kubenswrapper[4700]: I0227 17:01:33.415908 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:01:33 crc kubenswrapper[4700]: I0227 17:01:33.417651 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:01:33 crc kubenswrapper[4700]: I0227 17:01:33.417728 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:01:33 crc kubenswrapper[4700]: I0227 17:01:33.417749 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:01:33 crc kubenswrapper[4700]: I0227 17:01:33.422017 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 17:01:33 crc kubenswrapper[4700]: I0227 17:01:33.913060 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:34 crc kubenswrapper[4700]: I0227 17:01:34.251806 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:01:34 crc kubenswrapper[4700]: I0227 17:01:34.253392 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:01:34 crc kubenswrapper[4700]: I0227 17:01:34.253449 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:01:34 crc kubenswrapper[4700]: I0227 17:01:34.253501 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:01:34 crc kubenswrapper[4700]: I0227 17:01:34.910258 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:35 crc kubenswrapper[4700]: W0227 17:01:35.490509 4700 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Feb 27 17:01:35 crc kubenswrapper[4700]: E0227 17:01:35.490583 4700 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" Feb 27 17:01:35 crc kubenswrapper[4700]: I0227 17:01:35.911930 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:35 crc kubenswrapper[4700]: I0227 17:01:35.981150 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:01:35 crc kubenswrapper[4700]: I0227 17:01:35.982910 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:01:35 crc kubenswrapper[4700]: I0227 17:01:35.982964 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:01:35 crc kubenswrapper[4700]: I0227 17:01:35.982978 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:01:35 crc kubenswrapper[4700]: I0227 17:01:35.983655 4700 scope.go:117] "RemoveContainer" containerID="a785c0a3f8cf91c386fb0e3fa63d0bd9e57e702e96ba3e20011010d691066c8f" Feb 27 17:01:35 crc kubenswrapper[4700]: E0227 17:01:35.983862 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 17:01:36 crc kubenswrapper[4700]: I0227 17:01:36.913223 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:37 crc kubenswrapper[4700]: I0227 17:01:37.182630 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:01:37 crc kubenswrapper[4700]: I0227 17:01:37.184318 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:01:37 crc kubenswrapper[4700]: I0227 17:01:37.184368 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:01:37 crc kubenswrapper[4700]: I0227 17:01:37.184386 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:01:37 crc kubenswrapper[4700]: I0227 17:01:37.184427 4700 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 17:01:37 crc kubenswrapper[4700]: E0227 17:01:37.189667 4700 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Feb 27 17:01:37 crc kubenswrapper[4700]: E0227 17:01:37.191902 4700 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Feb 27 17:01:37 crc kubenswrapper[4700]: I0227 17:01:37.829918 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 17:01:37 crc kubenswrapper[4700]: I0227 17:01:37.830086 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:01:37 crc kubenswrapper[4700]: I0227 17:01:37.831256 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:01:37 crc kubenswrapper[4700]: I0227 17:01:37.831287 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:01:37 crc kubenswrapper[4700]: I0227 17:01:37.831298 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:01:37 crc kubenswrapper[4700]: I0227 17:01:37.912663 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:38 crc kubenswrapper[4700]: I0227 17:01:38.911076 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:39 crc kubenswrapper[4700]: I0227 17:01:39.912701 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:40 crc kubenswrapper[4700]: I0227 17:01:40.912184 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:41 crc kubenswrapper[4700]: E0227 17:01:41.089238 4700 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 27 17:01:41 crc kubenswrapper[4700]: I0227 17:01:41.912644 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:42 crc kubenswrapper[4700]: I0227 17:01:42.913725 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:43 crc kubenswrapper[4700]: I0227 17:01:43.912362 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:44 crc kubenswrapper[4700]: I0227 17:01:44.190121 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:01:44 crc kubenswrapper[4700]: I0227 17:01:44.192162 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:01:44 crc kubenswrapper[4700]: I0227 17:01:44.192764 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:01:44 crc kubenswrapper[4700]: I0227 17:01:44.192783 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:01:44 crc kubenswrapper[4700]: I0227 17:01:44.192828 4700 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 17:01:44 crc kubenswrapper[4700]: E0227 17:01:44.198162 4700 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Feb 27 17:01:44 crc kubenswrapper[4700]: E0227 17:01:44.198167 4700 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Feb 27 17:01:44 crc kubenswrapper[4700]: I0227 17:01:44.911625 4700 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 17:01:45 crc kubenswrapper[4700]: I0227 17:01:45.626590 4700 csr.go:261] certificate signing request csr-wp64s is approved, waiting to be issued Feb 27 17:01:45 crc kubenswrapper[4700]: I0227 17:01:45.643211 4700 csr.go:257] certificate signing request csr-wp64s is issued Feb 27 17:01:45 crc kubenswrapper[4700]: I0227 17:01:45.682649 4700 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Feb 27 17:01:45 crc kubenswrapper[4700]: I0227 17:01:45.744066 4700 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Feb 27 17:01:46 crc kubenswrapper[4700]: I0227 17:01:46.645053 4700 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-09 13:55:10.772937835 +0000 UTC Feb 27 17:01:46 crc kubenswrapper[4700]: I0227 17:01:46.645143 4700 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6836h53m24.127802661s for next certificate rotation Feb 27 17:01:50 crc kubenswrapper[4700]: I0227 17:01:50.980706 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:01:50 crc kubenswrapper[4700]: I0227 17:01:50.981912 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:01:50 crc kubenswrapper[4700]: I0227 17:01:50.981966 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:01:50 crc kubenswrapper[4700]: I0227 17:01:50.981978 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:01:50 crc kubenswrapper[4700]: I0227 17:01:50.983192 4700 scope.go:117] "RemoveContainer" containerID="a785c0a3f8cf91c386fb0e3fa63d0bd9e57e702e96ba3e20011010d691066c8f" Feb 27 17:01:51 crc kubenswrapper[4700]: E0227 17:01:51.089442 4700 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 27 17:01:51 crc kubenswrapper[4700]: I0227 17:01:51.199276 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:01:51 crc kubenswrapper[4700]: I0227 17:01:51.200589 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:01:51 crc kubenswrapper[4700]: I0227 17:01:51.200643 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:01:51 crc kubenswrapper[4700]: I0227 17:01:51.200661 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:01:51 crc kubenswrapper[4700]: I0227 17:01:51.200814 4700 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 17:01:51 crc kubenswrapper[4700]: I0227 17:01:51.213761 4700 kubelet_node_status.go:115] "Node was previously registered" node="crc" Feb 27 17:01:51 crc kubenswrapper[4700]: I0227 17:01:51.214189 4700 kubelet_node_status.go:79] "Successfully registered node" node="crc" Feb 27 17:01:51 crc kubenswrapper[4700]: E0227 17:01:51.214228 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Feb 27 17:01:51 crc kubenswrapper[4700]: I0227 17:01:51.218582 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:01:51 crc kubenswrapper[4700]: I0227 17:01:51.218627 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:01:51 crc kubenswrapper[4700]: I0227 17:01:51.218636 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:01:51 crc kubenswrapper[4700]: I0227 17:01:51.218657 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:01:51 crc kubenswrapper[4700]: I0227 17:01:51.218669 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:01:51Z","lastTransitionTime":"2026-02-27T17:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:01:51 crc kubenswrapper[4700]: E0227 17:01:51.242202 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 17:01:51 crc kubenswrapper[4700]: I0227 17:01:51.251764 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:01:51 crc kubenswrapper[4700]: I0227 17:01:51.251853 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:01:51 crc kubenswrapper[4700]: I0227 17:01:51.251869 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:01:51 crc kubenswrapper[4700]: I0227 17:01:51.251902 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:01:51 crc kubenswrapper[4700]: I0227 17:01:51.251922 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:01:51Z","lastTransitionTime":"2026-02-27T17:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:01:51 crc kubenswrapper[4700]: E0227 17:01:51.269198 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 17:01:51 crc kubenswrapper[4700]: I0227 17:01:51.279398 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:01:51 crc kubenswrapper[4700]: I0227 17:01:51.279490 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:01:51 crc kubenswrapper[4700]: I0227 17:01:51.279516 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:01:51 crc kubenswrapper[4700]: I0227 17:01:51.279548 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:01:51 crc kubenswrapper[4700]: I0227 17:01:51.279677 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:01:51Z","lastTransitionTime":"2026-02-27T17:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:01:51 crc kubenswrapper[4700]: I0227 17:01:51.306426 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Feb 27 17:01:51 crc kubenswrapper[4700]: I0227 17:01:51.309545 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d"} Feb 27 17:01:51 crc kubenswrapper[4700]: I0227 17:01:51.309715 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:01:51 crc kubenswrapper[4700]: I0227 17:01:51.310646 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:01:51 crc kubenswrapper[4700]: I0227 17:01:51.310705 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:01:51 crc kubenswrapper[4700]: I0227 17:01:51.310725 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:01:51 crc kubenswrapper[4700]: E0227 17:01:51.312804 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 17:01:51 crc kubenswrapper[4700]: I0227 17:01:51.322139 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:01:51 crc kubenswrapper[4700]: I0227 17:01:51.322187 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:01:51 crc kubenswrapper[4700]: I0227 17:01:51.322202 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:01:51 crc kubenswrapper[4700]: I0227 17:01:51.322222 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:01:51 crc kubenswrapper[4700]: I0227 17:01:51.322236 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:01:51Z","lastTransitionTime":"2026-02-27T17:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:01:51 crc kubenswrapper[4700]: E0227 17:01:51.332651 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 17:01:51 crc kubenswrapper[4700]: E0227 17:01:51.332763 4700 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 27 17:01:51 crc kubenswrapper[4700]: E0227 17:01:51.332788 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:51 crc kubenswrapper[4700]: E0227 17:01:51.433402 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:51 crc kubenswrapper[4700]: I0227 17:01:51.521234 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:01:51 crc kubenswrapper[4700]: E0227 17:01:51.534300 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:51 crc kubenswrapper[4700]: E0227 17:01:51.635126 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:51 crc kubenswrapper[4700]: E0227 17:01:51.736212 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:51 crc kubenswrapper[4700]: E0227 17:01:51.837253 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:51 crc kubenswrapper[4700]: E0227 17:01:51.937692 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:52 crc kubenswrapper[4700]: E0227 17:01:52.038021 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:52 crc kubenswrapper[4700]: E0227 17:01:52.139161 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:52 crc kubenswrapper[4700]: E0227 17:01:52.239857 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:52 crc kubenswrapper[4700]: I0227 17:01:52.314675 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Feb 27 17:01:52 crc kubenswrapper[4700]: I0227 17:01:52.315433 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Feb 27 17:01:52 crc kubenswrapper[4700]: I0227 17:01:52.317564 4700 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d" exitCode=255 Feb 27 17:01:52 crc kubenswrapper[4700]: I0227 17:01:52.317614 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d"} Feb 27 17:01:52 crc kubenswrapper[4700]: I0227 17:01:52.317663 4700 scope.go:117] "RemoveContainer" containerID="a785c0a3f8cf91c386fb0e3fa63d0bd9e57e702e96ba3e20011010d691066c8f" Feb 27 17:01:52 crc kubenswrapper[4700]: I0227 17:01:52.317878 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:01:52 crc kubenswrapper[4700]: I0227 17:01:52.319190 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:01:52 crc kubenswrapper[4700]: I0227 17:01:52.319241 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:01:52 crc kubenswrapper[4700]: I0227 17:01:52.319259 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:01:52 crc kubenswrapper[4700]: I0227 17:01:52.320193 4700 scope.go:117] "RemoveContainer" containerID="26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d" Feb 27 17:01:52 crc kubenswrapper[4700]: E0227 17:01:52.320495 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 17:01:52 crc kubenswrapper[4700]: E0227 17:01:52.340341 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:52 crc kubenswrapper[4700]: E0227 17:01:52.440971 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:52 crc kubenswrapper[4700]: E0227 17:01:52.541766 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:52 crc kubenswrapper[4700]: E0227 17:01:52.642517 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:52 crc kubenswrapper[4700]: E0227 17:01:52.742806 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:52 crc kubenswrapper[4700]: E0227 17:01:52.843674 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:52 crc kubenswrapper[4700]: E0227 17:01:52.944548 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:53 crc kubenswrapper[4700]: E0227 17:01:53.044979 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:53 crc kubenswrapper[4700]: E0227 17:01:53.146119 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:53 crc kubenswrapper[4700]: E0227 17:01:53.246892 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:53 crc kubenswrapper[4700]: I0227 17:01:53.322799 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Feb 27 17:01:53 crc kubenswrapper[4700]: I0227 17:01:53.326385 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:01:53 crc kubenswrapper[4700]: I0227 17:01:53.327427 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:01:53 crc kubenswrapper[4700]: I0227 17:01:53.327524 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:01:53 crc kubenswrapper[4700]: I0227 17:01:53.327544 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:01:53 crc kubenswrapper[4700]: I0227 17:01:53.328573 4700 scope.go:117] "RemoveContainer" containerID="26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d" Feb 27 17:01:53 crc kubenswrapper[4700]: E0227 17:01:53.328884 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 17:01:53 crc kubenswrapper[4700]: E0227 17:01:53.347456 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:53 crc kubenswrapper[4700]: E0227 17:01:53.447823 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:53 crc kubenswrapper[4700]: E0227 17:01:53.548927 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:53 crc kubenswrapper[4700]: E0227 17:01:53.649554 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:53 crc kubenswrapper[4700]: E0227 17:01:53.750800 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:53 crc kubenswrapper[4700]: E0227 17:01:53.851263 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:53 crc kubenswrapper[4700]: E0227 17:01:53.952187 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:54 crc kubenswrapper[4700]: E0227 17:01:54.052408 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:54 crc kubenswrapper[4700]: E0227 17:01:54.154133 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:54 crc kubenswrapper[4700]: E0227 17:01:54.255307 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:54 crc kubenswrapper[4700]: E0227 17:01:54.356642 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:54 crc kubenswrapper[4700]: E0227 17:01:54.457544 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:54 crc kubenswrapper[4700]: I0227 17:01:54.494445 4700 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 27 17:01:54 crc kubenswrapper[4700]: E0227 17:01:54.557765 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:54 crc kubenswrapper[4700]: I0227 17:01:54.576404 4700 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:01:54 crc kubenswrapper[4700]: I0227 17:01:54.576763 4700 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 17:01:54 crc kubenswrapper[4700]: I0227 17:01:54.587980 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:01:54 crc kubenswrapper[4700]: I0227 17:01:54.588059 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:01:54 crc kubenswrapper[4700]: I0227 17:01:54.588080 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:01:54 crc kubenswrapper[4700]: I0227 17:01:54.589328 4700 scope.go:117] "RemoveContainer" containerID="26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d" Feb 27 17:01:54 crc kubenswrapper[4700]: E0227 17:01:54.589676 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 17:01:54 crc kubenswrapper[4700]: E0227 17:01:54.658645 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:54 crc kubenswrapper[4700]: E0227 17:01:54.759720 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:54 crc kubenswrapper[4700]: E0227 17:01:54.860984 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:54 crc kubenswrapper[4700]: E0227 17:01:54.961629 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:55 crc kubenswrapper[4700]: E0227 17:01:55.062265 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:55 crc kubenswrapper[4700]: E0227 17:01:55.163583 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:55 crc kubenswrapper[4700]: E0227 17:01:55.264676 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:55 crc kubenswrapper[4700]: E0227 17:01:55.365067 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:55 crc kubenswrapper[4700]: E0227 17:01:55.465496 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:55 crc kubenswrapper[4700]: E0227 17:01:55.566151 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:55 crc kubenswrapper[4700]: E0227 17:01:55.667016 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:55 crc kubenswrapper[4700]: E0227 17:01:55.767597 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:55 crc kubenswrapper[4700]: E0227 17:01:55.868041 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:55 crc kubenswrapper[4700]: E0227 17:01:55.968315 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:56 crc kubenswrapper[4700]: E0227 17:01:56.069424 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:56 crc kubenswrapper[4700]: E0227 17:01:56.169586 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:56 crc kubenswrapper[4700]: E0227 17:01:56.270630 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:56 crc kubenswrapper[4700]: E0227 17:01:56.371340 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:56 crc kubenswrapper[4700]: E0227 17:01:56.472238 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:56 crc kubenswrapper[4700]: E0227 17:01:56.573014 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:56 crc kubenswrapper[4700]: E0227 17:01:56.674122 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:56 crc kubenswrapper[4700]: E0227 17:01:56.774875 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:56 crc kubenswrapper[4700]: I0227 17:01:56.786172 4700 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 27 17:01:56 crc kubenswrapper[4700]: E0227 17:01:56.875391 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:56 crc kubenswrapper[4700]: E0227 17:01:56.976502 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:57 crc kubenswrapper[4700]: E0227 17:01:57.076629 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:57 crc kubenswrapper[4700]: E0227 17:01:57.177679 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:57 crc kubenswrapper[4700]: E0227 17:01:57.278194 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:57 crc kubenswrapper[4700]: E0227 17:01:57.378653 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:57 crc kubenswrapper[4700]: E0227 17:01:57.479253 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:57 crc kubenswrapper[4700]: E0227 17:01:57.580015 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:57 crc kubenswrapper[4700]: E0227 17:01:57.680326 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:57 crc kubenswrapper[4700]: E0227 17:01:57.781003 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:57 crc kubenswrapper[4700]: E0227 17:01:57.881512 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:57 crc kubenswrapper[4700]: E0227 17:01:57.981779 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:58 crc kubenswrapper[4700]: E0227 17:01:58.081936 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:58 crc kubenswrapper[4700]: E0227 17:01:58.182545 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:58 crc kubenswrapper[4700]: E0227 17:01:58.283260 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:58 crc kubenswrapper[4700]: E0227 17:01:58.384337 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:58 crc kubenswrapper[4700]: E0227 17:01:58.485095 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:58 crc kubenswrapper[4700]: E0227 17:01:58.585656 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:58 crc kubenswrapper[4700]: E0227 17:01:58.686562 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:58 crc kubenswrapper[4700]: E0227 17:01:58.786836 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:58 crc kubenswrapper[4700]: E0227 17:01:58.887859 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:58 crc kubenswrapper[4700]: E0227 17:01:58.988223 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:59 crc kubenswrapper[4700]: E0227 17:01:59.089239 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:59 crc kubenswrapper[4700]: E0227 17:01:59.190355 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:59 crc kubenswrapper[4700]: E0227 17:01:59.290567 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:59 crc kubenswrapper[4700]: E0227 17:01:59.391288 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:59 crc kubenswrapper[4700]: E0227 17:01:59.492524 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:59 crc kubenswrapper[4700]: E0227 17:01:59.593682 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:59 crc kubenswrapper[4700]: E0227 17:01:59.694092 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:59 crc kubenswrapper[4700]: E0227 17:01:59.794516 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:59 crc kubenswrapper[4700]: E0227 17:01:59.894949 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:01:59 crc kubenswrapper[4700]: E0227 17:01:59.996155 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:00 crc kubenswrapper[4700]: E0227 17:02:00.096268 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:00 crc kubenswrapper[4700]: E0227 17:02:00.197112 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:00 crc kubenswrapper[4700]: E0227 17:02:00.297819 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:00 crc kubenswrapper[4700]: E0227 17:02:00.398399 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:00 crc kubenswrapper[4700]: E0227 17:02:00.499545 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:00 crc kubenswrapper[4700]: E0227 17:02:00.600349 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:00 crc kubenswrapper[4700]: E0227 17:02:00.701416 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:00 crc kubenswrapper[4700]: E0227 17:02:00.802690 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:00 crc kubenswrapper[4700]: E0227 17:02:00.903601 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:01 crc kubenswrapper[4700]: E0227 17:02:01.004779 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:01 crc kubenswrapper[4700]: E0227 17:02:01.090074 4700 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 27 17:02:01 crc kubenswrapper[4700]: E0227 17:02:01.105604 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:01 crc kubenswrapper[4700]: E0227 17:02:01.206235 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:01 crc kubenswrapper[4700]: E0227 17:02:01.306986 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:01 crc kubenswrapper[4700]: E0227 17:02:01.407095 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:01 crc kubenswrapper[4700]: E0227 17:02:01.478342 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Feb 27 17:02:01 crc kubenswrapper[4700]: I0227 17:02:01.484771 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:01 crc kubenswrapper[4700]: I0227 17:02:01.485025 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:01 crc kubenswrapper[4700]: I0227 17:02:01.485227 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:01 crc kubenswrapper[4700]: I0227 17:02:01.485388 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:01 crc kubenswrapper[4700]: I0227 17:02:01.485571 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:01Z","lastTransitionTime":"2026-02-27T17:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:01 crc kubenswrapper[4700]: E0227 17:02:01.497069 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 17:02:01 crc kubenswrapper[4700]: I0227 17:02:01.502813 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:01 crc kubenswrapper[4700]: I0227 17:02:01.502890 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:01 crc kubenswrapper[4700]: I0227 17:02:01.502912 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:01 crc kubenswrapper[4700]: I0227 17:02:01.502938 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:01 crc kubenswrapper[4700]: I0227 17:02:01.502960 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:01Z","lastTransitionTime":"2026-02-27T17:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:01 crc kubenswrapper[4700]: E0227 17:02:01.514548 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 17:02:01 crc kubenswrapper[4700]: I0227 17:02:01.520652 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:01 crc kubenswrapper[4700]: I0227 17:02:01.520728 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:01 crc kubenswrapper[4700]: I0227 17:02:01.520747 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:01 crc kubenswrapper[4700]: I0227 17:02:01.520774 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:01 crc kubenswrapper[4700]: I0227 17:02:01.520793 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:01Z","lastTransitionTime":"2026-02-27T17:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:01 crc kubenswrapper[4700]: E0227 17:02:01.536903 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 17:02:01 crc kubenswrapper[4700]: I0227 17:02:01.542203 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:01 crc kubenswrapper[4700]: I0227 17:02:01.542264 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:01 crc kubenswrapper[4700]: I0227 17:02:01.542283 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:01 crc kubenswrapper[4700]: I0227 17:02:01.542306 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:01 crc kubenswrapper[4700]: I0227 17:02:01.542324 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:01Z","lastTransitionTime":"2026-02-27T17:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:01 crc kubenswrapper[4700]: E0227 17:02:01.558331 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 17:02:01 crc kubenswrapper[4700]: E0227 17:02:01.558618 4700 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 27 17:02:01 crc kubenswrapper[4700]: E0227 17:02:01.558663 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:01 crc kubenswrapper[4700]: E0227 17:02:01.658774 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:01 crc kubenswrapper[4700]: E0227 17:02:01.759954 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:01 crc kubenswrapper[4700]: E0227 17:02:01.860499 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:01 crc kubenswrapper[4700]: E0227 17:02:01.960916 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:02 crc kubenswrapper[4700]: E0227 17:02:02.061410 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:02 crc kubenswrapper[4700]: E0227 17:02:02.161740 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:02 crc kubenswrapper[4700]: E0227 17:02:02.262139 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:02 crc kubenswrapper[4700]: E0227 17:02:02.362757 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:02 crc kubenswrapper[4700]: E0227 17:02:02.463952 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:02 crc kubenswrapper[4700]: E0227 17:02:02.565079 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:02 crc kubenswrapper[4700]: E0227 17:02:02.665492 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:02 crc kubenswrapper[4700]: I0227 17:02:02.746208 4700 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 27 17:02:02 crc kubenswrapper[4700]: E0227 17:02:02.766732 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:02 crc kubenswrapper[4700]: E0227 17:02:02.867368 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:02 crc kubenswrapper[4700]: E0227 17:02:02.968093 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:03 crc kubenswrapper[4700]: E0227 17:02:03.068541 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:03 crc kubenswrapper[4700]: E0227 17:02:03.169653 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:03 crc kubenswrapper[4700]: E0227 17:02:03.270686 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:03 crc kubenswrapper[4700]: E0227 17:02:03.371535 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:03 crc kubenswrapper[4700]: E0227 17:02:03.472137 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:03 crc kubenswrapper[4700]: E0227 17:02:03.573266 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:03 crc kubenswrapper[4700]: E0227 17:02:03.674421 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:03 crc kubenswrapper[4700]: E0227 17:02:03.774532 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:03 crc kubenswrapper[4700]: E0227 17:02:03.874738 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:03 crc kubenswrapper[4700]: E0227 17:02:03.975128 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:04 crc kubenswrapper[4700]: E0227 17:02:04.075766 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:04 crc kubenswrapper[4700]: E0227 17:02:04.176033 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:04 crc kubenswrapper[4700]: E0227 17:02:04.276099 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:04 crc kubenswrapper[4700]: E0227 17:02:04.376543 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:04 crc kubenswrapper[4700]: E0227 17:02:04.477226 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:04 crc kubenswrapper[4700]: E0227 17:02:04.578425 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:04 crc kubenswrapper[4700]: E0227 17:02:04.679341 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:04 crc kubenswrapper[4700]: E0227 17:02:04.779983 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:04 crc kubenswrapper[4700]: E0227 17:02:04.880571 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:04 crc kubenswrapper[4700]: E0227 17:02:04.980946 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:05 crc kubenswrapper[4700]: E0227 17:02:05.081415 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:05 crc kubenswrapper[4700]: E0227 17:02:05.182224 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:05 crc kubenswrapper[4700]: E0227 17:02:05.282900 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:05 crc kubenswrapper[4700]: E0227 17:02:05.383540 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:05 crc kubenswrapper[4700]: E0227 17:02:05.484314 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:05 crc kubenswrapper[4700]: E0227 17:02:05.584973 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:05 crc kubenswrapper[4700]: E0227 17:02:05.685350 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:05 crc kubenswrapper[4700]: E0227 17:02:05.786668 4700 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.822808 4700 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.890363 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.890444 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.890515 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.890555 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.890581 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:05Z","lastTransitionTime":"2026-02-27T17:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.934625 4700 apiserver.go:52] "Watching apiserver" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.942825 4700 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.943627 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-rbptz","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-operator/iptables-alerter-4ln5h","openshift-ovn-kubernetes/ovnkube-node-jtbqn","openshift-dns/node-resolver-dnkln","openshift-machine-config-operator/machine-config-daemon-f9dq7","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6","openshift-image-registry/node-ca-g66q8","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-node-identity/network-node-identity-vrzqb","openshift-multus/multus-additional-cni-plugins-8n2wb","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-multus/network-metrics-daemon-s8k74"] Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.944115 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.944343 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.944448 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.945013 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 27 17:02:05 crc kubenswrapper[4700]: E0227 17:02:05.945094 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.945044 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.945427 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.945491 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" Feb 27 17:02:05 crc kubenswrapper[4700]: E0227 17:02:05.945560 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.945747 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-g66q8" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.945896 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-dnkln" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.945918 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-rbptz" Feb 27 17:02:05 crc kubenswrapper[4700]: E0227 17:02:05.946423 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.946574 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.946849 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.947973 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.948180 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.948301 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:05 crc kubenswrapper[4700]: E0227 17:02:05.948550 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.948734 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.949311 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.949561 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.949497 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.949551 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.950812 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.950970 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.951390 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.952351 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.952401 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.952410 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.952740 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.952824 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.952903 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.952917 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.953188 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.953664 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.953675 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.953776 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.955448 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.955911 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.956115 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.956140 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.956660 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.956687 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.956660 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.956681 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.956878 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.957348 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.957441 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.957771 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.958748 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.958937 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.958576 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.959282 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.959324 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.977178 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.994064 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.994133 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.994156 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.994185 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.994204 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:05Z","lastTransitionTime":"2026-02-27T17:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:05 crc kubenswrapper[4700]: I0227 17:02:05.996508 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.009276 4700 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.010595 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b45511-b94c-479f-98db-fd2c4eceec46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f9dq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.011448 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.011523 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.011553 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.011578 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.011604 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.011632 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.011658 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.011686 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.011714 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.011740 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.011765 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.011791 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.011816 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.011841 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.011865 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.011891 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.011920 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.011952 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.011980 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012005 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012032 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012058 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012092 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012124 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012149 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012176 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012201 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012226 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012251 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012279 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012307 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012333 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012358 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012388 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012422 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012446 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012491 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012522 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012548 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012574 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012599 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012625 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012653 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012682 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012707 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012768 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012796 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012822 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012847 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012871 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012899 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012926 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012963 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.013005 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.013039 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.013072 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.013099 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.013135 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.013171 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.013200 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.013234 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.013275 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.013304 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.013336 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.013373 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.013410 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.013445 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.013513 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.013557 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.013597 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.013636 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.013675 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.013709 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.013748 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.013790 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.013827 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.013885 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.013922 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.013957 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.013994 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.014028 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.014066 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.014102 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.014139 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.014173 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.014204 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.014239 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.014271 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.014305 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.014340 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.014374 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.014410 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.014441 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.014498 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.014536 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.014568 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.014599 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.014947 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.015498 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012247 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.016623 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.016652 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.016700 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.017007 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.017327 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.017676 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.017752 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.017998 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.018289 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.018393 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.018433 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.018449 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.016949 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.018767 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.018800 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.018810 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.018830 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.018866 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.018896 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.018939 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.018979 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.019124 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.019169 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.019273 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.019682 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.019884 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.019998 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.020060 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.020123 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.020174 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.020215 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.020244 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.020271 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.020283 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.020304 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.020439 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.020507 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.020736 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.020782 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.020840 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.020897 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.020999 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.014877 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012560 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012552 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012808 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.021202 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.012995 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.013212 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.013323 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.013366 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.013680 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.013707 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.013927 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.014030 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.014009 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.014898 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.015728 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.015710 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.015896 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.016220 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.016421 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.016490 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.018451 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.021979 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.022019 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.022079 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.022099 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.021922 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.022152 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.022740 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.022753 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.022974 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.023032 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.023071 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.023156 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.023186 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.023266 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.023315 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.024397 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.024632 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.024987 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.024994 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.025356 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.025437 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.025681 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.025815 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.026043 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.026164 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.026225 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.026968 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.027151 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.027022 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.027575 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.027257 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.027321 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.027387 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-s8k74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5288317-4810-4321-baa0-31a118a4dc36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-s8k74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.027813 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.026966 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.028010 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.028048 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.028134 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.028564 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.028633 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.028661 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.028697 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.028740 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.028824 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.028867 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.028904 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.029049 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.029091 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.029127 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.029097 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.029159 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.029198 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.029286 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.029642 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.029685 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.029717 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.029749 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.029788 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.029822 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.029903 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.029936 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.030078 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.030115 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.030148 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.030182 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.030212 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.030252 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.030288 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.030321 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.030362 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.030401 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.030438 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.030507 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.030542 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.030587 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.030642 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.030681 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.030720 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.030755 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.030787 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.030820 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.030886 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.030928 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.030962 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.030999 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.031031 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.031069 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.031101 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.031137 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.031172 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.031211 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.031248 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.031282 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.031319 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.031646 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.031689 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.031726 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.032084 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.032132 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.032177 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.032217 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.032259 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.032300 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.032340 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.032377 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.032412 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.032444 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.032539 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.032587 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.032636 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.032671 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.032708 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.032741 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.032777 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.032810 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.032847 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.032880 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.032986 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.033028 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/66eb22fb-e593-40ff-9ff3-5bb32912972d-cni-binary-copy\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.033063 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-host-var-lib-cni-multus\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.033096 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-multus-conf-dir\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.033130 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-run-systemd\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.033267 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.029160 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.033303 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/47cf6a8e-df81-4ccb-992b-abb030862331-os-release\") pod \"multus-additional-cni-plugins-8n2wb\" (UID: \"47cf6a8e-df81-4ccb-992b-abb030862331\") " pod="openshift-multus/multus-additional-cni-plugins-8n2wb" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.029804 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.033345 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-etc-openvswitch\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.029920 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.030348 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.033382 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0f425b79-78da-4f9e-a1ad-205a09466052-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-cxnv6\" (UID: \"0f425b79-78da-4f9e-a1ad-205a09466052\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.033611 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-multus-cni-dir\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.033742 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-run-netns\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.033894 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.033898 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.033940 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f5288317-4810-4321-baa0-31a118a4dc36-metrics-certs\") pod \"network-metrics-daemon-s8k74\" (UID: \"f5288317-4810-4321-baa0-31a118a4dc36\") " pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.033981 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-multus-socket-dir-parent\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.034069 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-hostroot\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.030713 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.034119 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.030815 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.030798 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.031054 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.031702 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.031968 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.032033 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.034210 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/47cf6a8e-df81-4ccb-992b-abb030862331-cnibin\") pod \"multus-additional-cni-plugins-8n2wb\" (UID: \"47cf6a8e-df81-4ccb-992b-abb030862331\") " pod="openshift-multus/multus-additional-cni-plugins-8n2wb" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.032326 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.032608 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.032013 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.034298 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/47cf6a8e-df81-4ccb-992b-abb030862331-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-8n2wb\" (UID: \"47cf6a8e-df81-4ccb-992b-abb030862331\") " pod="openshift-multus/multus-additional-cni-plugins-8n2wb" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.034390 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.034719 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/84b45511-b94c-479f-98db-fd2c4eceec46-proxy-tls\") pod \"machine-config-daemon-f9dq7\" (UID: \"84b45511-b94c-479f-98db-fd2c4eceec46\") " pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.034799 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-slash\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.035564 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-cni-netd\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.035633 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3a34c0e3-2513-4e81-b6b1-80d1230475fd-env-overrides\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.035711 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.035799 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v92s4\" (UniqueName: \"kubernetes.io/projected/e772eecd-127f-4b26-8021-010fb507ce0c-kube-api-access-v92s4\") pod \"node-resolver-dnkln\" (UID: \"e772eecd-127f-4b26-8021-010fb507ce0c\") " pod="openshift-dns/node-resolver-dnkln" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.035856 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-systemd-units\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.035940 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-kubelet\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.036003 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5wt7\" (UniqueName: \"kubernetes.io/projected/f5288317-4810-4321-baa0-31a118a4dc36-kube-api-access-j5wt7\") pod \"network-metrics-daemon-s8k74\" (UID: \"f5288317-4810-4321-baa0-31a118a4dc36\") " pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.036057 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-host-run-multus-certs\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.036119 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/84b45511-b94c-479f-98db-fd2c4eceec46-rootfs\") pod \"machine-config-daemon-f9dq7\" (UID: \"84b45511-b94c-479f-98db-fd2c4eceec46\") " pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.036348 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2472dab1-91b8-4be5-bd0c-dbe6522f5b94-host\") pod \"node-ca-g66q8\" (UID: \"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\") " pod="openshift-image-registry/node-ca-g66q8" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.036417 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.036650 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-etc-kubernetes\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.036709 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2dtm\" (UniqueName: \"kubernetes.io/projected/66eb22fb-e593-40ff-9ff3-5bb32912972d-kube-api-access-v2dtm\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.036776 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.036839 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3a34c0e3-2513-4e81-b6b1-80d1230475fd-ovn-node-metrics-cert\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.036901 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.036961 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-system-cni-dir\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.037015 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-cnibin\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.037066 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-node-log\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.037118 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-log-socket\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.037201 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.037263 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/47cf6a8e-df81-4ccb-992b-abb030862331-tuning-conf-dir\") pod \"multus-additional-cni-plugins-8n2wb\" (UID: \"47cf6a8e-df81-4ccb-992b-abb030862331\") " pod="openshift-multus/multus-additional-cni-plugins-8n2wb" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.037325 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dt85\" (UniqueName: \"kubernetes.io/projected/0f425b79-78da-4f9e-a1ad-205a09466052-kube-api-access-8dt85\") pod \"ovnkube-control-plane-749d76644c-cxnv6\" (UID: \"0f425b79-78da-4f9e-a1ad-205a09466052\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.037788 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/47cf6a8e-df81-4ccb-992b-abb030862331-cni-binary-copy\") pod \"multus-additional-cni-plugins-8n2wb\" (UID: \"47cf6a8e-df81-4ccb-992b-abb030862331\") " pod="openshift-multus/multus-additional-cni-plugins-8n2wb" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.032943 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.034149 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.034252 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.034331 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.034544 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.034633 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.034677 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.034688 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.034724 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.034816 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.035214 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.035242 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.034975 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.035648 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.038937 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.035210 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: E0227 17:02:06.037603 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:02:06.537573564 +0000 UTC m=+86.522886341 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.038156 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.038367 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.038397 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.038627 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.039006 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: E0227 17:02:06.039047 4700 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.040178 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.040202 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.040545 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.040580 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.040701 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.040853 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: E0227 17:02:06.038752 4700 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.039263 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.039643 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.039590 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.039724 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.039995 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.039902 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.041355 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.041793 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.042035 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.042131 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.042206 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.042435 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.043312 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.043493 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.043553 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0f425b79-78da-4f9e-a1ad-205a09466052-env-overrides\") pod \"ovnkube-control-plane-749d76644c-cxnv6\" (UID: \"0f425b79-78da-4f9e-a1ad-205a09466052\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.043590 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-748j7\" (UniqueName: \"kubernetes.io/projected/3a34c0e3-2513-4e81-b6b1-80d1230475fd-kube-api-access-748j7\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.043699 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.043751 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/84b45511-b94c-479f-98db-fd2c4eceec46-mcd-auth-proxy-config\") pod \"machine-config-daemon-f9dq7\" (UID: \"84b45511-b94c-479f-98db-fd2c4eceec46\") " pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.043782 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/2472dab1-91b8-4be5-bd0c-dbe6522f5b94-serviceca\") pod \"node-ca-g66q8\" (UID: \"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\") " pod="openshift-image-registry/node-ca-g66q8" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.043869 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.044164 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.044746 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.045021 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5b6cx\" (UniqueName: \"kubernetes.io/projected/2472dab1-91b8-4be5-bd0c-dbe6522f5b94-kube-api-access-5b6cx\") pod \"node-ca-g66q8\" (UID: \"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\") " pod="openshift-image-registry/node-ca-g66q8" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.045203 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.045355 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.045407 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.045634 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: E0227 17:02:06.045840 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 17:02:06.545712985 +0000 UTC m=+86.531025762 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 17:02:06 crc kubenswrapper[4700]: E0227 17:02:06.045997 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 17:02:06.545980252 +0000 UTC m=+86.531293029 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.046641 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.046565 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.046525 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.046735 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.046804 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.046780 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.047164 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.047257 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.047284 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.047643 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.047668 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.047716 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.048033 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.048123 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f425b79-78da-4f9e-a1ad-205a09466052\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cxnv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.048274 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.048528 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.048597 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.045851 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-run-openvswitch\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.046225 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.048844 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.048891 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.048911 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.048932 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqdf7\" (UniqueName: \"kubernetes.io/projected/84b45511-b94c-479f-98db-fd2c4eceec46-kube-api-access-xqdf7\") pod \"machine-config-daemon-f9dq7\" (UID: \"84b45511-b94c-479f-98db-fd2c4eceec46\") " pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.048968 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.049032 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8d8w\" (UniqueName: \"kubernetes.io/projected/47cf6a8e-df81-4ccb-992b-abb030862331-kube-api-access-m8d8w\") pod \"multus-additional-cni-plugins-8n2wb\" (UID: \"47cf6a8e-df81-4ccb-992b-abb030862331\") " pod="openshift-multus/multus-additional-cni-plugins-8n2wb" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.049108 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.049166 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-var-lib-openvswitch\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.049218 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-run-ovn-kubernetes\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.049266 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-cni-bin\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.049328 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.049383 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3a34c0e3-2513-4e81-b6b1-80d1230475fd-ovnkube-config\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.049543 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-host-var-lib-kubelet\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.049600 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e772eecd-127f-4b26-8021-010fb507ce0c-hosts-file\") pod \"node-resolver-dnkln\" (UID: \"e772eecd-127f-4b26-8021-010fb507ce0c\") " pod="openshift-dns/node-resolver-dnkln" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.049653 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-run-ovn\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.049705 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-host-run-netns\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.049823 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/66eb22fb-e593-40ff-9ff3-5bb32912972d-multus-daemon-config\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.049978 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3a34c0e3-2513-4e81-b6b1-80d1230475fd-ovnkube-script-lib\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.050027 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.049823 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.050088 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0f425b79-78da-4f9e-a1ad-205a09466052-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-cxnv6\" (UID: \"0f425b79-78da-4f9e-a1ad-205a09466052\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.050139 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.050173 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-host-run-k8s-cni-cncf-io\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.050269 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-host-var-lib-cni-bin\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.050311 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/47cf6a8e-df81-4ccb-992b-abb030862331-system-cni-dir\") pod \"multus-additional-cni-plugins-8n2wb\" (UID: \"47cf6a8e-df81-4ccb-992b-abb030862331\") " pod="openshift-multus/multus-additional-cni-plugins-8n2wb" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.050342 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-os-release\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.050521 4700 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.050766 4700 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.050784 4700 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.050799 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.050815 4700 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.050829 4700 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.050839 4700 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.050852 4700 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.050863 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.050874 4700 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.050884 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.050994 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.052927 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.053090 4700 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.053105 4700 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.053117 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.053126 4700 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.053135 4700 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.053179 4700 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.053191 4700 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.053215 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.055383 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.055435 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.055447 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.055532 4700 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.055547 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.055557 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.056042 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.056604 4700 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.056810 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.056854 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.056849 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.056877 4700 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057005 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057029 4700 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057048 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057069 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057086 4700 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057100 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057115 4700 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057131 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057146 4700 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057164 4700 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057180 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057198 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057213 4700 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057228 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057246 4700 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057262 4700 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057283 4700 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057301 4700 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057317 4700 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057334 4700 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057350 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057365 4700 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057381 4700 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057395 4700 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057410 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057424 4700 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057441 4700 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057478 4700 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057497 4700 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057516 4700 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057534 4700 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057553 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057569 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057588 4700 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057604 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057620 4700 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057636 4700 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057653 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057668 4700 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057684 4700 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057700 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057717 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057733 4700 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057749 4700 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057806 4700 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057825 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057840 4700 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057857 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057873 4700 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057888 4700 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057904 4700 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057921 4700 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057937 4700 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057953 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057969 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057983 4700 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.057998 4700 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.058013 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.058029 4700 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.058047 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.058065 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.058081 4700 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.058096 4700 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.058111 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.058126 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.058143 4700 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.058159 4700 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.058176 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.058192 4700 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.058209 4700 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.058225 4700 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.058240 4700 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.058256 4700 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.058272 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.058288 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.058303 4700 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.058321 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.058337 4700 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.058354 4700 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.058371 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.058387 4700 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.058403 4700 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.058419 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.058437 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.058453 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.058852 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.058870 4700 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.058886 4700 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.058902 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.058918 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.058935 4700 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.058952 4700 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.058967 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.058983 4700 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059000 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059016 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059034 4700 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059050 4700 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059069 4700 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059088 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059106 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059124 4700 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059158 4700 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059175 4700 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059191 4700 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059209 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059225 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059240 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059256 4700 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059271 4700 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059286 4700 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059301 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059319 4700 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059334 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059349 4700 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059368 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059384 4700 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059398 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059415 4700 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059430 4700 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059447 4700 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059488 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059504 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059521 4700 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059537 4700 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059553 4700 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059569 4700 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059593 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059609 4700 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059625 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059641 4700 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059657 4700 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059676 4700 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059694 4700 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059711 4700 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059728 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059744 4700 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059762 4700 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.059778 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.061089 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.061145 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.061216 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.062115 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.062482 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.062772 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.062853 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.064230 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.064546 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.064838 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.065415 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: E0227 17:02:06.065851 4700 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 17:02:06 crc kubenswrapper[4700]: E0227 17:02:06.065887 4700 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 17:02:06 crc kubenswrapper[4700]: E0227 17:02:06.065909 4700 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 17:02:06 crc kubenswrapper[4700]: E0227 17:02:06.066007 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-27 17:02:06.56598264 +0000 UTC m=+86.551295427 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.066326 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.066120 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.068618 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.069531 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.069771 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.069658 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.069804 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: E0227 17:02:06.070103 4700 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 17:02:06 crc kubenswrapper[4700]: E0227 17:02:06.070143 4700 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 17:02:06 crc kubenswrapper[4700]: E0227 17:02:06.070167 4700 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 17:02:06 crc kubenswrapper[4700]: E0227 17:02:06.070264 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-27 17:02:06.57022818 +0000 UTC m=+86.555541157 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.071832 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.074515 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.074885 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.075550 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.075933 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.077270 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.080783 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.081081 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.084794 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-g66q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b6cx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-g66q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.086789 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.089869 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.096838 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rbptz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66eb22fb-e593-40ff-9ff3-5bb32912972d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2dtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rbptz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.098064 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.098101 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.098117 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.098143 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.098161 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:06Z","lastTransitionTime":"2026-02-27T17:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.107996 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.109943 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.117315 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.122163 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.133568 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.145061 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dnkln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e772eecd-127f-4b26-8021-010fb507ce0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v92s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dnkln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.161095 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/47cf6a8e-df81-4ccb-992b-abb030862331-cni-binary-copy\") pod \"multus-additional-cni-plugins-8n2wb\" (UID: \"47cf6a8e-df81-4ccb-992b-abb030862331\") " pod="openshift-multus/multus-additional-cni-plugins-8n2wb" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.161176 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0f425b79-78da-4f9e-a1ad-205a09466052-env-overrides\") pod \"ovnkube-control-plane-749d76644c-cxnv6\" (UID: \"0f425b79-78da-4f9e-a1ad-205a09466052\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.161222 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/84b45511-b94c-479f-98db-fd2c4eceec46-mcd-auth-proxy-config\") pod \"machine-config-daemon-f9dq7\" (UID: \"84b45511-b94c-479f-98db-fd2c4eceec46\") " pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.161256 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/2472dab1-91b8-4be5-bd0c-dbe6522f5b94-serviceca\") pod \"node-ca-g66q8\" (UID: \"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\") " pod="openshift-image-registry/node-ca-g66q8" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.161297 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-748j7\" (UniqueName: \"kubernetes.io/projected/3a34c0e3-2513-4e81-b6b1-80d1230475fd-kube-api-access-748j7\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.161334 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqdf7\" (UniqueName: \"kubernetes.io/projected/84b45511-b94c-479f-98db-fd2c4eceec46-kube-api-access-xqdf7\") pod \"machine-config-daemon-f9dq7\" (UID: \"84b45511-b94c-479f-98db-fd2c4eceec46\") " pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.161369 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5b6cx\" (UniqueName: \"kubernetes.io/projected/2472dab1-91b8-4be5-bd0c-dbe6522f5b94-kube-api-access-5b6cx\") pod \"node-ca-g66q8\" (UID: \"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\") " pod="openshift-image-registry/node-ca-g66q8" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.161402 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-run-openvswitch\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.161437 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-run-ovn-kubernetes\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.161512 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-cni-bin\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.161593 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8d8w\" (UniqueName: \"kubernetes.io/projected/47cf6a8e-df81-4ccb-992b-abb030862331-kube-api-access-m8d8w\") pod \"multus-additional-cni-plugins-8n2wb\" (UID: \"47cf6a8e-df81-4ccb-992b-abb030862331\") " pod="openshift-multus/multus-additional-cni-plugins-8n2wb" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.161632 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-var-lib-openvswitch\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.161668 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3a34c0e3-2513-4e81-b6b1-80d1230475fd-ovnkube-config\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.161705 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-host-var-lib-kubelet\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.161739 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e772eecd-127f-4b26-8021-010fb507ce0c-hosts-file\") pod \"node-resolver-dnkln\" (UID: \"e772eecd-127f-4b26-8021-010fb507ce0c\") " pod="openshift-dns/node-resolver-dnkln" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.161773 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-run-ovn\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.161818 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0f425b79-78da-4f9e-a1ad-205a09466052-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-cxnv6\" (UID: \"0f425b79-78da-4f9e-a1ad-205a09466052\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.161852 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-host-run-netns\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.161887 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/66eb22fb-e593-40ff-9ff3-5bb32912972d-multus-daemon-config\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.161954 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3a34c0e3-2513-4e81-b6b1-80d1230475fd-ovnkube-script-lib\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.162018 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/47cf6a8e-df81-4ccb-992b-abb030862331-system-cni-dir\") pod \"multus-additional-cni-plugins-8n2wb\" (UID: \"47cf6a8e-df81-4ccb-992b-abb030862331\") " pod="openshift-multus/multus-additional-cni-plugins-8n2wb" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.162067 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-os-release\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.162086 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-cni-bin\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.162120 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-host-run-k8s-cni-cncf-io\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.162171 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-run-ovn-kubernetes\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.162180 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-host-var-lib-cni-bin\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.162189 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0f425b79-78da-4f9e-a1ad-205a09466052-env-overrides\") pod \"ovnkube-control-plane-749d76644c-cxnv6\" (UID: \"0f425b79-78da-4f9e-a1ad-205a09466052\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.162227 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-multus-conf-dir\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.162240 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/47cf6a8e-df81-4ccb-992b-abb030862331-cni-binary-copy\") pod \"multus-additional-cni-plugins-8n2wb\" (UID: \"47cf6a8e-df81-4ccb-992b-abb030862331\") " pod="openshift-multus/multus-additional-cni-plugins-8n2wb" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.162202 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-var-lib-openvswitch\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.162278 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-run-systemd\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.162330 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.162367 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/47cf6a8e-df81-4ccb-992b-abb030862331-os-release\") pod \"multus-additional-cni-plugins-8n2wb\" (UID: \"47cf6a8e-df81-4ccb-992b-abb030862331\") " pod="openshift-multus/multus-additional-cni-plugins-8n2wb" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.162401 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.162434 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/66eb22fb-e593-40ff-9ff3-5bb32912972d-cni-binary-copy\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.162499 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-host-var-lib-cni-multus\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.162051 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-run-openvswitch\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.162576 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-etc-openvswitch\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.162585 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-host-var-lib-cni-bin\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.162590 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-host-run-k8s-cni-cncf-io\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.162531 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-etc-openvswitch\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.162628 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/47cf6a8e-df81-4ccb-992b-abb030862331-system-cni-dir\") pod \"multus-additional-cni-plugins-8n2wb\" (UID: \"47cf6a8e-df81-4ccb-992b-abb030862331\") " pod="openshift-multus/multus-additional-cni-plugins-8n2wb" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.162693 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0f425b79-78da-4f9e-a1ad-205a09466052-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-cxnv6\" (UID: \"0f425b79-78da-4f9e-a1ad-205a09466052\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.162675 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-os-release\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.162713 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/47cf6a8e-df81-4ccb-992b-abb030862331-os-release\") pod \"multus-additional-cni-plugins-8n2wb\" (UID: \"47cf6a8e-df81-4ccb-992b-abb030862331\") " pod="openshift-multus/multus-additional-cni-plugins-8n2wb" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.162528 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-multus-conf-dir\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.162627 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-run-systemd\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.162665 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.162778 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-multus-cni-dir\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.162800 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-run-ovn\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.162797 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-host-var-lib-cni-multus\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.162907 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3a34c0e3-2513-4e81-b6b1-80d1230475fd-ovnkube-script-lib\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.162919 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-run-netns\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.162954 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-hostroot\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.162985 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.163002 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/47cf6a8e-df81-4ccb-992b-abb030862331-cnibin\") pod \"multus-additional-cni-plugins-8n2wb\" (UID: \"47cf6a8e-df81-4ccb-992b-abb030862331\") " pod="openshift-multus/multus-additional-cni-plugins-8n2wb" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.162982 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3a34c0e3-2513-4e81-b6b1-80d1230475fd-ovnkube-config\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.163036 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-hostroot\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.163059 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/47cf6a8e-df81-4ccb-992b-abb030862331-cnibin\") pod \"multus-additional-cni-plugins-8n2wb\" (UID: \"47cf6a8e-df81-4ccb-992b-abb030862331\") " pod="openshift-multus/multus-additional-cni-plugins-8n2wb" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.163072 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-multus-cni-dir\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.163082 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-run-netns\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.163093 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-host-var-lib-kubelet\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.163128 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-host-run-netns\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.163186 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/47cf6a8e-df81-4ccb-992b-abb030862331-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-8n2wb\" (UID: \"47cf6a8e-df81-4ccb-992b-abb030862331\") " pod="openshift-multus/multus-additional-cni-plugins-8n2wb" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.163402 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e772eecd-127f-4b26-8021-010fb507ce0c-hosts-file\") pod \"node-resolver-dnkln\" (UID: \"e772eecd-127f-4b26-8021-010fb507ce0c\") " pod="openshift-dns/node-resolver-dnkln" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.163854 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/66eb22fb-e593-40ff-9ff3-5bb32912972d-cni-binary-copy\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.163936 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f5288317-4810-4321-baa0-31a118a4dc36-metrics-certs\") pod \"network-metrics-daemon-s8k74\" (UID: \"f5288317-4810-4321-baa0-31a118a4dc36\") " pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.163970 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-multus-socket-dir-parent\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.164017 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/84b45511-b94c-479f-98db-fd2c4eceec46-proxy-tls\") pod \"machine-config-daemon-f9dq7\" (UID: \"84b45511-b94c-479f-98db-fd2c4eceec46\") " pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.164049 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v92s4\" (UniqueName: \"kubernetes.io/projected/e772eecd-127f-4b26-8021-010fb507ce0c-kube-api-access-v92s4\") pod \"node-resolver-dnkln\" (UID: \"e772eecd-127f-4b26-8021-010fb507ce0c\") " pod="openshift-dns/node-resolver-dnkln" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.164061 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-multus-socket-dir-parent\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.164080 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-systemd-units\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.164108 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-slash\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.164139 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-cni-netd\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: E0227 17:02:06.164158 4700 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.164172 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3a34c0e3-2513-4e81-b6b1-80d1230475fd-env-overrides\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.164190 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0f425b79-78da-4f9e-a1ad-205a09466052-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-cxnv6\" (UID: \"0f425b79-78da-4f9e-a1ad-205a09466052\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" Feb 27 17:02:06 crc kubenswrapper[4700]: E0227 17:02:06.164224 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f5288317-4810-4321-baa0-31a118a4dc36-metrics-certs podName:f5288317-4810-4321-baa0-31a118a4dc36 nodeName:}" failed. No retries permitted until 2026-02-27 17:02:06.664203094 +0000 UTC m=+86.649515861 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f5288317-4810-4321-baa0-31a118a4dc36-metrics-certs") pod "network-metrics-daemon-s8k74" (UID: "f5288317-4810-4321-baa0-31a118a4dc36") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.164205 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5wt7\" (UniqueName: \"kubernetes.io/projected/f5288317-4810-4321-baa0-31a118a4dc36-kube-api-access-j5wt7\") pod \"network-metrics-daemon-s8k74\" (UID: \"f5288317-4810-4321-baa0-31a118a4dc36\") " pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.164270 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-systemd-units\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.164290 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-host-run-multus-certs\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.164325 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-kubelet\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.164358 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-etc-kubernetes\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.164388 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2dtm\" (UniqueName: \"kubernetes.io/projected/66eb22fb-e593-40ff-9ff3-5bb32912972d-kube-api-access-v2dtm\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.164420 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/47cf6a8e-df81-4ccb-992b-abb030862331-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-8n2wb\" (UID: \"47cf6a8e-df81-4ccb-992b-abb030862331\") " pod="openshift-multus/multus-additional-cni-plugins-8n2wb" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.164435 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/84b45511-b94c-479f-98db-fd2c4eceec46-rootfs\") pod \"machine-config-daemon-f9dq7\" (UID: \"84b45511-b94c-479f-98db-fd2c4eceec46\") " pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.164493 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2472dab1-91b8-4be5-bd0c-dbe6522f5b94-host\") pod \"node-ca-g66q8\" (UID: \"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\") " pod="openshift-image-registry/node-ca-g66q8" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.164520 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-kubelet\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.164530 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.164523 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/2472dab1-91b8-4be5-bd0c-dbe6522f5b94-serviceca\") pod \"node-ca-g66q8\" (UID: \"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\") " pod="openshift-image-registry/node-ca-g66q8" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.164570 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3a34c0e3-2513-4e81-b6b1-80d1230475fd-ovn-node-metrics-cert\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.164599 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-host-run-multus-certs\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.164607 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-slash\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.164620 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-log-socket\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.164662 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2472dab1-91b8-4be5-bd0c-dbe6522f5b94-host\") pod \"node-ca-g66q8\" (UID: \"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\") " pod="openshift-image-registry/node-ca-g66q8" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.164667 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-log-socket\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.164573 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-cni-netd\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.165033 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/47cf6a8e-df81-4ccb-992b-abb030862331-tuning-conf-dir\") pod \"multus-additional-cni-plugins-8n2wb\" (UID: \"47cf6a8e-df81-4ccb-992b-abb030862331\") " pod="openshift-multus/multus-additional-cni-plugins-8n2wb" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.165077 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dt85\" (UniqueName: \"kubernetes.io/projected/0f425b79-78da-4f9e-a1ad-205a09466052-kube-api-access-8dt85\") pod \"ovnkube-control-plane-749d76644c-cxnv6\" (UID: \"0f425b79-78da-4f9e-a1ad-205a09466052\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.165054 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/84b45511-b94c-479f-98db-fd2c4eceec46-mcd-auth-proxy-config\") pod \"machine-config-daemon-f9dq7\" (UID: \"84b45511-b94c-479f-98db-fd2c4eceec46\") " pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.165203 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-system-cni-dir\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.165249 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3a34c0e3-2513-4e81-b6b1-80d1230475fd-env-overrides\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.165113 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-system-cni-dir\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.165310 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.165625 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-cnibin\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.165976 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-node-log\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.166006 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-cnibin\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.166030 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/66eb22fb-e593-40ff-9ff3-5bb32912972d-etc-kubernetes\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.166289 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/47cf6a8e-df81-4ccb-992b-abb030862331-tuning-conf-dir\") pod \"multus-additional-cni-plugins-8n2wb\" (UID: \"47cf6a8e-df81-4ccb-992b-abb030862331\") " pod="openshift-multus/multus-additional-cni-plugins-8n2wb" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.164920 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/84b45511-b94c-479f-98db-fd2c4eceec46-rootfs\") pod \"machine-config-daemon-f9dq7\" (UID: \"84b45511-b94c-479f-98db-fd2c4eceec46\") " pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.166500 4700 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.166529 4700 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.166548 4700 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.166563 4700 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.166579 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.166594 4700 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.166610 4700 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.166626 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.166640 4700 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.166656 4700 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.166650 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-node-log\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.166680 4700 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.166698 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.166715 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.166734 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.166752 4700 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.166769 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.166785 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.166800 4700 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.166816 4700 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.166831 4700 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.166846 4700 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.166860 4700 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.166876 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.166894 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.166911 4700 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.166926 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.170503 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/66eb22fb-e593-40ff-9ff3-5bb32912972d-multus-daemon-config\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.174168 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0f425b79-78da-4f9e-a1ad-205a09466052-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-cxnv6\" (UID: \"0f425b79-78da-4f9e-a1ad-205a09466052\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.174067 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47cf6a8e-df81-4ccb-992b-abb030862331\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8n2wb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.174738 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/84b45511-b94c-479f-98db-fd2c4eceec46-proxy-tls\") pod \"machine-config-daemon-f9dq7\" (UID: \"84b45511-b94c-479f-98db-fd2c4eceec46\") " pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.174837 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3a34c0e3-2513-4e81-b6b1-80d1230475fd-ovn-node-metrics-cert\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.183960 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5b6cx\" (UniqueName: \"kubernetes.io/projected/2472dab1-91b8-4be5-bd0c-dbe6522f5b94-kube-api-access-5b6cx\") pod \"node-ca-g66q8\" (UID: \"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\") " pod="openshift-image-registry/node-ca-g66q8" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.186126 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v92s4\" (UniqueName: \"kubernetes.io/projected/e772eecd-127f-4b26-8021-010fb507ce0c-kube-api-access-v92s4\") pod \"node-resolver-dnkln\" (UID: \"e772eecd-127f-4b26-8021-010fb507ce0c\") " pod="openshift-dns/node-resolver-dnkln" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.188853 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqdf7\" (UniqueName: \"kubernetes.io/projected/84b45511-b94c-479f-98db-fd2c4eceec46-kube-api-access-xqdf7\") pod \"machine-config-daemon-f9dq7\" (UID: \"84b45511-b94c-479f-98db-fd2c4eceec46\") " pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.189526 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.190379 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8d8w\" (UniqueName: \"kubernetes.io/projected/47cf6a8e-df81-4ccb-992b-abb030862331-kube-api-access-m8d8w\") pod \"multus-additional-cni-plugins-8n2wb\" (UID: \"47cf6a8e-df81-4ccb-992b-abb030862331\") " pod="openshift-multus/multus-additional-cni-plugins-8n2wb" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.192961 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5wt7\" (UniqueName: \"kubernetes.io/projected/f5288317-4810-4321-baa0-31a118a4dc36-kube-api-access-j5wt7\") pod \"network-metrics-daemon-s8k74\" (UID: \"f5288317-4810-4321-baa0-31a118a4dc36\") " pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.193536 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-748j7\" (UniqueName: \"kubernetes.io/projected/3a34c0e3-2513-4e81-b6b1-80d1230475fd-kube-api-access-748j7\") pod \"ovnkube-node-jtbqn\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.196557 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2dtm\" (UniqueName: \"kubernetes.io/projected/66eb22fb-e593-40ff-9ff3-5bb32912972d-kube-api-access-v2dtm\") pod \"multus-rbptz\" (UID: \"66eb22fb-e593-40ff-9ff3-5bb32912972d\") " pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.198686 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dt85\" (UniqueName: \"kubernetes.io/projected/0f425b79-78da-4f9e-a1ad-205a09466052-kube-api-access-8dt85\") pod \"ovnkube-control-plane-749d76644c-cxnv6\" (UID: \"0f425b79-78da-4f9e-a1ad-205a09466052\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.201556 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.201636 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.201664 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.201696 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.201725 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:06Z","lastTransitionTime":"2026-02-27T17:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.211239 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a34c0e3-2513-4e81-b6b1-80d1230475fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jtbqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.266430 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.286828 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.305549 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.305635 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.305692 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.305723 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.305776 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:06Z","lastTransitionTime":"2026-02-27T17:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.307313 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.331520 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-g66q8" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.356985 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-dnkln" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.367701 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"197643c2c6bb6413d2c6ca4dde021b9b029d3eb48f5556706d061a9f0f64e089"} Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.369125 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"5f1cb9329c8d879680feedffad9cdba7acfd3e9013bebd2eb8858553df908b45"} Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.370914 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"0a7a29dc5383f044fa496b0cfcd2380e45d3c7d68fd3ee34bbdb99a90766e891"} Feb 27 17:02:06 crc kubenswrapper[4700]: W0227 17:02:06.371276 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2472dab1_91b8_4be5_bd0c_dbe6522f5b94.slice/crio-b2cff83fa8e9db1f95f458802b4ac349d23502836735279f3f510c2151ad694e WatchSource:0}: Error finding container b2cff83fa8e9db1f95f458802b4ac349d23502836735279f3f510c2151ad694e: Status 404 returned error can't find the container with id b2cff83fa8e9db1f95f458802b4ac349d23502836735279f3f510c2151ad694e Feb 27 17:02:06 crc kubenswrapper[4700]: W0227 17:02:06.372218 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode772eecd_127f_4b26_8021_010fb507ce0c.slice/crio-c8a6de94bb9e45b32b18554554a13e3ecae1d9658fd7aadbf53c432ee4d47161 WatchSource:0}: Error finding container c8a6de94bb9e45b32b18554554a13e3ecae1d9658fd7aadbf53c432ee4d47161: Status 404 returned error can't find the container with id c8a6de94bb9e45b32b18554554a13e3ecae1d9658fd7aadbf53c432ee4d47161 Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.396554 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-rbptz" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.409098 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.409131 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.409140 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.409154 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.409165 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:06Z","lastTransitionTime":"2026-02-27T17:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.409568 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" Feb 27 17:02:06 crc kubenswrapper[4700]: W0227 17:02:06.415902 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66eb22fb_e593_40ff_9ff3_5bb32912972d.slice/crio-58884a088c8e4fbb777d12e5594eec752c86296f8c202a1756da0a2baf7a26ab WatchSource:0}: Error finding container 58884a088c8e4fbb777d12e5594eec752c86296f8c202a1756da0a2baf7a26ab: Status 404 returned error can't find the container with id 58884a088c8e4fbb777d12e5594eec752c86296f8c202a1756da0a2baf7a26ab Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.418877 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" Feb 27 17:02:06 crc kubenswrapper[4700]: W0227 17:02:06.432164 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod84b45511_b94c_479f_98db_fd2c4eceec46.slice/crio-1008e2edd091c3e6ec5e205877a02172515d40347ae908e1b9905eec11f54402 WatchSource:0}: Error finding container 1008e2edd091c3e6ec5e205877a02172515d40347ae908e1b9905eec11f54402: Status 404 returned error can't find the container with id 1008e2edd091c3e6ec5e205877a02172515d40347ae908e1b9905eec11f54402 Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.433408 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" Feb 27 17:02:06 crc kubenswrapper[4700]: W0227 17:02:06.442640 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod47cf6a8e_df81_4ccb_992b_abb030862331.slice/crio-9815e2b2cf4388a3895c1f1765897e4b486962a988da48e5803a8db3941f73a6 WatchSource:0}: Error finding container 9815e2b2cf4388a3895c1f1765897e4b486962a988da48e5803a8db3941f73a6: Status 404 returned error can't find the container with id 9815e2b2cf4388a3895c1f1765897e4b486962a988da48e5803a8db3941f73a6 Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.445325 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:06 crc kubenswrapper[4700]: W0227 17:02:06.475877 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0f425b79_78da_4f9e_a1ad_205a09466052.slice/crio-97a341239cdf658000b513b2819581c90baa8d8c114961521cdf462a7709d7d5 WatchSource:0}: Error finding container 97a341239cdf658000b513b2819581c90baa8d8c114961521cdf462a7709d7d5: Status 404 returned error can't find the container with id 97a341239cdf658000b513b2819581c90baa8d8c114961521cdf462a7709d7d5 Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.513046 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.513084 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.513095 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.513116 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.513137 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:06Z","lastTransitionTime":"2026-02-27T17:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.573410 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.573573 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.573600 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:02:06 crc kubenswrapper[4700]: E0227 17:02:06.573676 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:02:07.573635348 +0000 UTC m=+87.558948095 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:02:06 crc kubenswrapper[4700]: E0227 17:02:06.573736 4700 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 17:02:06 crc kubenswrapper[4700]: E0227 17:02:06.573765 4700 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 17:02:06 crc kubenswrapper[4700]: E0227 17:02:06.573777 4700 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 17:02:06 crc kubenswrapper[4700]: E0227 17:02:06.573788 4700 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 17:02:06 crc kubenswrapper[4700]: E0227 17:02:06.573805 4700 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.573804 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:06 crc kubenswrapper[4700]: E0227 17:02:06.573816 4700 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 17:02:06 crc kubenswrapper[4700]: E0227 17:02:06.573893 4700 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 17:02:06 crc kubenswrapper[4700]: E0227 17:02:06.573839 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-27 17:02:07.573817673 +0000 UTC m=+87.559130410 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 17:02:06 crc kubenswrapper[4700]: E0227 17:02:06.573924 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 17:02:07.573916486 +0000 UTC m=+87.559229233 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.573954 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:06 crc kubenswrapper[4700]: E0227 17:02:06.574018 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-27 17:02:07.574010248 +0000 UTC m=+87.559322995 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 17:02:06 crc kubenswrapper[4700]: E0227 17:02:06.574101 4700 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 17:02:06 crc kubenswrapper[4700]: E0227 17:02:06.574185 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 17:02:07.574159662 +0000 UTC m=+87.559472419 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.615702 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.615735 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.615745 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.615760 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.615770 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:06Z","lastTransitionTime":"2026-02-27T17:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.675036 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f5288317-4810-4321-baa0-31a118a4dc36-metrics-certs\") pod \"network-metrics-daemon-s8k74\" (UID: \"f5288317-4810-4321-baa0-31a118a4dc36\") " pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:06 crc kubenswrapper[4700]: E0227 17:02:06.675234 4700 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 17:02:06 crc kubenswrapper[4700]: E0227 17:02:06.675316 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f5288317-4810-4321-baa0-31a118a4dc36-metrics-certs podName:f5288317-4810-4321-baa0-31a118a4dc36 nodeName:}" failed. No retries permitted until 2026-02-27 17:02:07.675291081 +0000 UTC m=+87.660603838 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f5288317-4810-4321-baa0-31a118a4dc36-metrics-certs") pod "network-metrics-daemon-s8k74" (UID: "f5288317-4810-4321-baa0-31a118a4dc36") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.718866 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.719262 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.719275 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.719293 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.719308 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:06Z","lastTransitionTime":"2026-02-27T17:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.821498 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.821555 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.821565 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.821585 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.821600 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:06Z","lastTransitionTime":"2026-02-27T17:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.924726 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.924766 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.924774 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.924791 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.924800 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:06Z","lastTransitionTime":"2026-02-27T17:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.986250 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.987438 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.989183 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.990087 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.991799 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.992650 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.993629 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.994913 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.995852 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.997219 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Feb 27 17:02:06 crc kubenswrapper[4700]: I0227 17:02:06.998216 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.000203 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.001116 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.001891 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.004223 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.005297 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.006995 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.007857 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.008844 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.010158 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.010854 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.012336 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.012849 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.013917 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.014359 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.015414 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.016107 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.016653 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.017637 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.018142 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.018987 4700 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.019090 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.021104 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.022162 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.022792 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.024721 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.025537 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.026554 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.027286 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.027581 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.027642 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.027655 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.027679 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.027760 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:07Z","lastTransitionTime":"2026-02-27T17:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.028358 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.028839 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.029841 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.031298 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.032606 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.034043 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.035056 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.036242 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.037097 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.038136 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.039445 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.041284 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.043507 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.044756 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.048009 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.131361 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.131421 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.131439 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.131528 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.131552 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:07Z","lastTransitionTime":"2026-02-27T17:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.233479 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.233530 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.233544 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.233559 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.233569 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:07Z","lastTransitionTime":"2026-02-27T17:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.336373 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.336420 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.336429 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.336449 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.336480 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:07Z","lastTransitionTime":"2026-02-27T17:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.375731 4700 generic.go:334] "Generic (PLEG): container finished" podID="47cf6a8e-df81-4ccb-992b-abb030862331" containerID="d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919" exitCode=0 Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.375816 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" event={"ID":"47cf6a8e-df81-4ccb-992b-abb030862331","Type":"ContainerDied","Data":"d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919"} Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.375857 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" event={"ID":"47cf6a8e-df81-4ccb-992b-abb030862331","Type":"ContainerStarted","Data":"9815e2b2cf4388a3895c1f1765897e4b486962a988da48e5803a8db3941f73a6"} Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.380540 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerStarted","Data":"bddf1a7b7264d36174b90f1d0d4e49f029781b1acfce88582aa5b2575dcdd09a"} Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.380603 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerStarted","Data":"2ef7c37b61de7a039756248f222bd1769b774703c406de5ad11abd9fb2f46dd2"} Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.380617 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerStarted","Data":"1008e2edd091c3e6ec5e205877a02172515d40347ae908e1b9905eec11f54402"} Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.383207 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"c501ee5970cba15e45e1f1c8f0e33768019511f94dbbe509a7bbfea2c5b4a064"} Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.383370 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"0f4cd0745c61d33ac717a4a655f00713e2bd935c3cbba23c4ca0913859f6016f"} Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.386316 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" event={"ID":"0f425b79-78da-4f9e-a1ad-205a09466052","Type":"ContainerStarted","Data":"bd29e6d703945e0f366f2d109b7c35fdd6360c77091b165616e9832dddac1c05"} Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.386353 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" event={"ID":"0f425b79-78da-4f9e-a1ad-205a09466052","Type":"ContainerStarted","Data":"3baf8927284b91f3edeec1dc906c2928955badb19062087790d720936b6f7624"} Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.386369 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" event={"ID":"0f425b79-78da-4f9e-a1ad-205a09466052","Type":"ContainerStarted","Data":"97a341239cdf658000b513b2819581c90baa8d8c114961521cdf462a7709d7d5"} Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.388230 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-g66q8" event={"ID":"2472dab1-91b8-4be5-bd0c-dbe6522f5b94","Type":"ContainerStarted","Data":"55d4f1eeba71eeb576c1df386f5acaf2bc354e7ca08aa9aac0971f61db72f244"} Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.388493 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-g66q8" event={"ID":"2472dab1-91b8-4be5-bd0c-dbe6522f5b94","Type":"ContainerStarted","Data":"b2cff83fa8e9db1f95f458802b4ac349d23502836735279f3f510c2151ad694e"} Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.390133 4700 generic.go:334] "Generic (PLEG): container finished" podID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerID="d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1" exitCode=0 Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.390205 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" event={"ID":"3a34c0e3-2513-4e81-b6b1-80d1230475fd","Type":"ContainerDied","Data":"d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1"} Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.390312 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" event={"ID":"3a34c0e3-2513-4e81-b6b1-80d1230475fd","Type":"ContainerStarted","Data":"59b3e89a8e1c655b3a026b9f6a993dcc53a00351a6765f6762f67e5f72758623"} Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.391316 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-dnkln" event={"ID":"e772eecd-127f-4b26-8021-010fb507ce0c","Type":"ContainerStarted","Data":"4e6df6de510d7e55695db4c89cbe9db84f91aca0e4bfbf2ed852407047c32822"} Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.391356 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-dnkln" event={"ID":"e772eecd-127f-4b26-8021-010fb507ce0c","Type":"ContainerStarted","Data":"c8a6de94bb9e45b32b18554554a13e3ecae1d9658fd7aadbf53c432ee4d47161"} Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.394103 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-rbptz" event={"ID":"66eb22fb-e593-40ff-9ff3-5bb32912972d","Type":"ContainerStarted","Data":"f58f4814976a1e25bef53555d7f5d14fa1cc800c8c5220801960758d5262b552"} Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.394139 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-rbptz" event={"ID":"66eb22fb-e593-40ff-9ff3-5bb32912972d","Type":"ContainerStarted","Data":"58884a088c8e4fbb777d12e5594eec752c86296f8c202a1756da0a2baf7a26ab"} Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.396343 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"a336c2f16e12b411f111f6291366aa1b466f6b405896c152d1667b832a589b7c"} Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.410234 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:07Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.424779 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-g66q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b6cx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-g66q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:07Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.440795 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.440842 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.440859 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.440887 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.440904 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:07Z","lastTransitionTime":"2026-02-27T17:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.450507 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rbptz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66eb22fb-e593-40ff-9ff3-5bb32912972d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2dtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rbptz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:07Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.474110 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:07Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.494620 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:07Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.513389 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:07Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.527020 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dnkln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e772eecd-127f-4b26-8021-010fb507ce0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v92s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dnkln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:07Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.544149 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.544185 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.544197 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.544219 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.544231 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:07Z","lastTransitionTime":"2026-02-27T17:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.551074 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47cf6a8e-df81-4ccb-992b-abb030862331\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8n2wb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:07Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.571772 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:07Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.585534 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.585672 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.585702 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:07 crc kubenswrapper[4700]: E0227 17:02:07.585726 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:02:09.585697449 +0000 UTC m=+89.571010216 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.585775 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:02:07 crc kubenswrapper[4700]: E0227 17:02:07.585818 4700 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.585827 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:02:07 crc kubenswrapper[4700]: E0227 17:02:07.585877 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 17:02:09.585860734 +0000 UTC m=+89.571173531 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 17:02:07 crc kubenswrapper[4700]: E0227 17:02:07.586020 4700 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 17:02:07 crc kubenswrapper[4700]: E0227 17:02:07.586035 4700 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 17:02:07 crc kubenswrapper[4700]: E0227 17:02:07.586048 4700 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 17:02:07 crc kubenswrapper[4700]: E0227 17:02:07.586088 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-27 17:02:09.586078869 +0000 UTC m=+89.571391636 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 17:02:07 crc kubenswrapper[4700]: E0227 17:02:07.586144 4700 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 17:02:07 crc kubenswrapper[4700]: E0227 17:02:07.586156 4700 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 17:02:07 crc kubenswrapper[4700]: E0227 17:02:07.586167 4700 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 17:02:07 crc kubenswrapper[4700]: E0227 17:02:07.586194 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-27 17:02:09.586185712 +0000 UTC m=+89.571498469 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 17:02:07 crc kubenswrapper[4700]: E0227 17:02:07.586253 4700 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 17:02:07 crc kubenswrapper[4700]: E0227 17:02:07.586279 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 17:02:09.586271354 +0000 UTC m=+89.571584201 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.600774 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a34c0e3-2513-4e81-b6b1-80d1230475fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jtbqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:07Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.619573 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:07Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.636533 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b45511-b94c-479f-98db-fd2c4eceec46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f9dq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:07Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.646542 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.646592 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.646606 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.646625 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.646641 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:07Z","lastTransitionTime":"2026-02-27T17:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.663915 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-s8k74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5288317-4810-4321-baa0-31a118a4dc36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-s8k74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:07Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.683061 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f425b79-78da-4f9e-a1ad-205a09466052\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cxnv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:07Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.686795 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f5288317-4810-4321-baa0-31a118a4dc36-metrics-certs\") pod \"network-metrics-daemon-s8k74\" (UID: \"f5288317-4810-4321-baa0-31a118a4dc36\") " pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:07 crc kubenswrapper[4700]: E0227 17:02:07.686979 4700 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 17:02:07 crc kubenswrapper[4700]: E0227 17:02:07.687037 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f5288317-4810-4321-baa0-31a118a4dc36-metrics-certs podName:f5288317-4810-4321-baa0-31a118a4dc36 nodeName:}" failed. No retries permitted until 2026-02-27 17:02:09.687021724 +0000 UTC m=+89.672334471 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f5288317-4810-4321-baa0-31a118a4dc36-metrics-certs") pod "network-metrics-daemon-s8k74" (UID: "f5288317-4810-4321-baa0-31a118a4dc36") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.697730 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:07Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.712274 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-g66q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d4f1eeba71eeb576c1df386f5acaf2bc354e7ca08aa9aac0971f61db72f244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b6cx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-g66q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:07Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.726214 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rbptz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66eb22fb-e593-40ff-9ff3-5bb32912972d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f58f4814976a1e25bef53555d7f5d14fa1cc800c8c5220801960758d5262b552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2dtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rbptz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:07Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.739095 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dnkln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e772eecd-127f-4b26-8021-010fb507ce0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6df6de510d7e55695db4c89cbe9db84f91aca0e4bfbf2ed852407047c32822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v92s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dnkln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:07Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.749404 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.749783 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.749793 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.749811 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.749820 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:07Z","lastTransitionTime":"2026-02-27T17:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.757294 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47cf6a8e-df81-4ccb-992b-abb030862331\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8n2wb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:07Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.770803 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:07Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.784063 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c501ee5970cba15e45e1f1c8f0e33768019511f94dbbe509a7bbfea2c5b4a064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f4cd0745c61d33ac717a4a655f00713e2bd935c3cbba23c4ca0913859f6016f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:07Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.796841 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:07Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.814487 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a336c2f16e12b411f111f6291366aa1b466f6b405896c152d1667b832a589b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:07Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.844329 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a34c0e3-2513-4e81-b6b1-80d1230475fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jtbqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:07Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.852396 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.852435 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.852447 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.852493 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.852504 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:07Z","lastTransitionTime":"2026-02-27T17:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.859493 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-s8k74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5288317-4810-4321-baa0-31a118a4dc36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-s8k74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:07Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.871943 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f425b79-78da-4f9e-a1ad-205a09466052\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3baf8927284b91f3edeec1dc906c2928955badb19062087790d720936b6f7624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd29e6d703945e0f366f2d109b7c35fdd6360c77091b165616e9832dddac1c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cxnv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:07Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.890897 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:07Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.903689 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b45511-b94c-479f-98db-fd2c4eceec46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bddf1a7b7264d36174b90f1d0d4e49f029781b1acfce88582aa5b2575dcdd09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ef7c37b61de7a039756248f222bd1769b774703c406de5ad11abd9fb2f46dd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f9dq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:07Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.957682 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.957749 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.957774 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.957802 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.957821 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:07Z","lastTransitionTime":"2026-02-27T17:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.980426 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.980543 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.980445 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:07 crc kubenswrapper[4700]: I0227 17:02:07.980673 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:02:07 crc kubenswrapper[4700]: E0227 17:02:07.980680 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:02:07 crc kubenswrapper[4700]: E0227 17:02:07.980818 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:02:07 crc kubenswrapper[4700]: E0227 17:02:07.980864 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:02:07 crc kubenswrapper[4700]: E0227 17:02:07.980945 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.060715 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.060777 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.060796 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.060823 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.060841 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:08Z","lastTransitionTime":"2026-02-27T17:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.069432 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.069519 4700 scope.go:117] "RemoveContainer" containerID="26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d" Feb 27 17:02:08 crc kubenswrapper[4700]: E0227 17:02:08.069873 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.164987 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.165042 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.165055 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.165082 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.165094 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:08Z","lastTransitionTime":"2026-02-27T17:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.268231 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.268724 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.268740 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.268759 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.268771 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:08Z","lastTransitionTime":"2026-02-27T17:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.371890 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.371937 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.371947 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.371966 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.371983 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:08Z","lastTransitionTime":"2026-02-27T17:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.402034 4700 generic.go:334] "Generic (PLEG): container finished" podID="47cf6a8e-df81-4ccb-992b-abb030862331" containerID="45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1" exitCode=0 Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.402178 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" event={"ID":"47cf6a8e-df81-4ccb-992b-abb030862331","Type":"ContainerDied","Data":"45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1"} Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.408222 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" event={"ID":"3a34c0e3-2513-4e81-b6b1-80d1230475fd","Type":"ContainerStarted","Data":"6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64"} Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.408290 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" event={"ID":"3a34c0e3-2513-4e81-b6b1-80d1230475fd","Type":"ContainerStarted","Data":"a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a"} Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.408308 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" event={"ID":"3a34c0e3-2513-4e81-b6b1-80d1230475fd","Type":"ContainerStarted","Data":"ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8"} Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.408328 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" event={"ID":"3a34c0e3-2513-4e81-b6b1-80d1230475fd","Type":"ContainerStarted","Data":"b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0"} Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.408635 4700 scope.go:117] "RemoveContainer" containerID="26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d" Feb 27 17:02:08 crc kubenswrapper[4700]: E0227 17:02:08.408859 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.420821 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:08Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.442499 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c501ee5970cba15e45e1f1c8f0e33768019511f94dbbe509a7bbfea2c5b4a064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f4cd0745c61d33ac717a4a655f00713e2bd935c3cbba23c4ca0913859f6016f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:08Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.457296 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:08Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.469152 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dnkln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e772eecd-127f-4b26-8021-010fb507ce0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6df6de510d7e55695db4c89cbe9db84f91aca0e4bfbf2ed852407047c32822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v92s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dnkln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:08Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.478779 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.478827 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.478842 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.478864 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.478876 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:08Z","lastTransitionTime":"2026-02-27T17:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.487646 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47cf6a8e-df81-4ccb-992b-abb030862331\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8n2wb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:08Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.502297 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a336c2f16e12b411f111f6291366aa1b466f6b405896c152d1667b832a589b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:08Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.524184 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a34c0e3-2513-4e81-b6b1-80d1230475fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jtbqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:08Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.546067 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a2d1b57-7fc1-4614-aac5-7144cdd812d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"27 17:01:51.956663 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 17:01:51.956671 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 17:01:51.956678 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 17:01:51.956715 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0227 17:01:51.960396 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1772211711\\\\\\\\\\\\\\\" (2026-02-27 17:01:50 +0000 UTC to 2026-03-29 17:01:51 +0000 UTC (now=2026-02-27 17:01:51.960134581 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960704 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1772211711\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1772211711\\\\\\\\\\\\\\\" (2026-02-27 16:01:51 +0000 UTC to 2027-02-27 16:01:51 +0000 UTC (now=2026-02-27 17:01:51.960667784 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960745 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0227 17:01:51.960790 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0227 17:01:51.960837 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960878 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960925 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\"\\\\nI0227 17:01:51.961071 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0227 17:01:51.958239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:01:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:08Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.560510 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:08Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.572557 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b45511-b94c-479f-98db-fd2c4eceec46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bddf1a7b7264d36174b90f1d0d4e49f029781b1acfce88582aa5b2575dcdd09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ef7c37b61de7a039756248f222bd1769b774703c406de5ad11abd9fb2f46dd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f9dq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:08Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.581119 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.581152 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.581165 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.581185 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.581198 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:08Z","lastTransitionTime":"2026-02-27T17:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.584045 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-s8k74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5288317-4810-4321-baa0-31a118a4dc36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-s8k74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:08Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.598188 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f425b79-78da-4f9e-a1ad-205a09466052\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3baf8927284b91f3edeec1dc906c2928955badb19062087790d720936b6f7624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd29e6d703945e0f366f2d109b7c35fdd6360c77091b165616e9832dddac1c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cxnv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:08Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.615131 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:08Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.627229 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-g66q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d4f1eeba71eeb576c1df386f5acaf2bc354e7ca08aa9aac0971f61db72f244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b6cx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-g66q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:08Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.639379 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rbptz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66eb22fb-e593-40ff-9ff3-5bb32912972d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f58f4814976a1e25bef53555d7f5d14fa1cc800c8c5220801960758d5262b552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2dtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rbptz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:08Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.683208 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.683251 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.683265 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.683286 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.683296 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:08Z","lastTransitionTime":"2026-02-27T17:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.786447 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.786495 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.786504 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.786519 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.786529 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:08Z","lastTransitionTime":"2026-02-27T17:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.890094 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.890167 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.890186 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.890213 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.890230 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:08Z","lastTransitionTime":"2026-02-27T17:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.995409 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.995469 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.995483 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.995500 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:08 crc kubenswrapper[4700]: I0227 17:02:08.995513 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:08Z","lastTransitionTime":"2026-02-27T17:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.098547 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.098628 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.098648 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.098674 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.098692 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:09Z","lastTransitionTime":"2026-02-27T17:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.201811 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.201872 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.201942 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.201966 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.202032 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:09Z","lastTransitionTime":"2026-02-27T17:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.305021 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.305102 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.305172 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.305205 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.305240 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:09Z","lastTransitionTime":"2026-02-27T17:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.409086 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.409153 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.409177 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.409208 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.409233 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:09Z","lastTransitionTime":"2026-02-27T17:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.422604 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" event={"ID":"3a34c0e3-2513-4e81-b6b1-80d1230475fd","Type":"ContainerStarted","Data":"c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf"} Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.422661 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" event={"ID":"3a34c0e3-2513-4e81-b6b1-80d1230475fd","Type":"ContainerStarted","Data":"611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265"} Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.428101 4700 generic.go:334] "Generic (PLEG): container finished" podID="47cf6a8e-df81-4ccb-992b-abb030862331" containerID="edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9" exitCode=0 Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.428149 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" event={"ID":"47cf6a8e-df81-4ccb-992b-abb030862331","Type":"ContainerDied","Data":"edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9"} Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.448651 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dnkln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e772eecd-127f-4b26-8021-010fb507ce0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6df6de510d7e55695db4c89cbe9db84f91aca0e4bfbf2ed852407047c32822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v92s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dnkln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:09Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.469904 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47cf6a8e-df81-4ccb-992b-abb030862331\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8n2wb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:09Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.490298 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:09Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.511143 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c501ee5970cba15e45e1f1c8f0e33768019511f94dbbe509a7bbfea2c5b4a064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f4cd0745c61d33ac717a4a655f00713e2bd935c3cbba23c4ca0913859f6016f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:09Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.514393 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.514438 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.514453 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.514498 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.514514 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:09Z","lastTransitionTime":"2026-02-27T17:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.532255 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:09Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.553069 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a336c2f16e12b411f111f6291366aa1b466f6b405896c152d1667b832a589b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:09Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.583408 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a34c0e3-2513-4e81-b6b1-80d1230475fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jtbqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:09Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.600392 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-s8k74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5288317-4810-4321-baa0-31a118a4dc36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-s8k74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:09Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.610191 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.610301 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.610333 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.610374 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.610419 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:09 crc kubenswrapper[4700]: E0227 17:02:09.610537 4700 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 17:02:09 crc kubenswrapper[4700]: E0227 17:02:09.610599 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 17:02:13.610572373 +0000 UTC m=+93.595885130 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 17:02:09 crc kubenswrapper[4700]: E0227 17:02:09.610867 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:02:13.6108561 +0000 UTC m=+93.596168857 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:02:09 crc kubenswrapper[4700]: E0227 17:02:09.610934 4700 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 17:02:09 crc kubenswrapper[4700]: E0227 17:02:09.610947 4700 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 17:02:09 crc kubenswrapper[4700]: E0227 17:02:09.610959 4700 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 17:02:09 crc kubenswrapper[4700]: E0227 17:02:09.610987 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-27 17:02:13.610978863 +0000 UTC m=+93.596291620 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 17:02:09 crc kubenswrapper[4700]: E0227 17:02:09.611034 4700 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 17:02:09 crc kubenswrapper[4700]: E0227 17:02:09.611046 4700 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 17:02:09 crc kubenswrapper[4700]: E0227 17:02:09.611054 4700 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 17:02:09 crc kubenswrapper[4700]: E0227 17:02:09.611094 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-27 17:02:13.611087156 +0000 UTC m=+93.596399913 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 17:02:09 crc kubenswrapper[4700]: E0227 17:02:09.611127 4700 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 17:02:09 crc kubenswrapper[4700]: E0227 17:02:09.611151 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 17:02:13.611144478 +0000 UTC m=+93.596457235 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.618172 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f425b79-78da-4f9e-a1ad-205a09466052\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3baf8927284b91f3edeec1dc906c2928955badb19062087790d720936b6f7624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd29e6d703945e0f366f2d109b7c35fdd6360c77091b165616e9832dddac1c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cxnv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:09Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.618708 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.618727 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.618737 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.618752 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.618763 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:09Z","lastTransitionTime":"2026-02-27T17:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.636339 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a2d1b57-7fc1-4614-aac5-7144cdd812d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"27 17:01:51.956663 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 17:01:51.956671 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 17:01:51.956678 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 17:01:51.956715 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0227 17:01:51.960396 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1772211711\\\\\\\\\\\\\\\" (2026-02-27 17:01:50 +0000 UTC to 2026-03-29 17:01:51 +0000 UTC (now=2026-02-27 17:01:51.960134581 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960704 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1772211711\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1772211711\\\\\\\\\\\\\\\" (2026-02-27 16:01:51 +0000 UTC to 2027-02-27 16:01:51 +0000 UTC (now=2026-02-27 17:01:51.960667784 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960745 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0227 17:01:51.960790 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0227 17:01:51.960837 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960878 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960925 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\"\\\\nI0227 17:01:51.961071 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0227 17:01:51.958239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:01:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:09Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.656219 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:09Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.668951 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b45511-b94c-479f-98db-fd2c4eceec46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bddf1a7b7264d36174b90f1d0d4e49f029781b1acfce88582aa5b2575dcdd09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ef7c37b61de7a039756248f222bd1769b774703c406de5ad11abd9fb2f46dd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f9dq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:09Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.683845 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:09Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.695035 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-g66q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d4f1eeba71eeb576c1df386f5acaf2bc354e7ca08aa9aac0971f61db72f244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b6cx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-g66q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:09Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.709338 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rbptz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66eb22fb-e593-40ff-9ff3-5bb32912972d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f58f4814976a1e25bef53555d7f5d14fa1cc800c8c5220801960758d5262b552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2dtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rbptz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:09Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.711192 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f5288317-4810-4321-baa0-31a118a4dc36-metrics-certs\") pod \"network-metrics-daemon-s8k74\" (UID: \"f5288317-4810-4321-baa0-31a118a4dc36\") " pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:09 crc kubenswrapper[4700]: E0227 17:02:09.711453 4700 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 17:02:09 crc kubenswrapper[4700]: E0227 17:02:09.711580 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f5288317-4810-4321-baa0-31a118a4dc36-metrics-certs podName:f5288317-4810-4321-baa0-31a118a4dc36 nodeName:}" failed. No retries permitted until 2026-02-27 17:02:13.711552478 +0000 UTC m=+93.696865265 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f5288317-4810-4321-baa0-31a118a4dc36-metrics-certs") pod "network-metrics-daemon-s8k74" (UID: "f5288317-4810-4321-baa0-31a118a4dc36") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.726565 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.726609 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.726618 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.726637 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.726648 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:09Z","lastTransitionTime":"2026-02-27T17:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.830059 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.830104 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.830116 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.830134 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.830145 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:09Z","lastTransitionTime":"2026-02-27T17:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.932507 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.932603 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.932624 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.932656 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.932673 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:09Z","lastTransitionTime":"2026-02-27T17:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.980442 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.980506 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.980504 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:02:09 crc kubenswrapper[4700]: I0227 17:02:09.980639 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:09 crc kubenswrapper[4700]: E0227 17:02:09.980678 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:02:09 crc kubenswrapper[4700]: E0227 17:02:09.981074 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:02:09 crc kubenswrapper[4700]: E0227 17:02:09.981222 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:02:09 crc kubenswrapper[4700]: E0227 17:02:09.981356 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.036120 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.036198 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.036227 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.036263 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.036281 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:10Z","lastTransitionTime":"2026-02-27T17:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.139632 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.139726 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.139768 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.139810 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.139830 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:10Z","lastTransitionTime":"2026-02-27T17:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.243496 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.243565 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.243582 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.243610 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.243628 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:10Z","lastTransitionTime":"2026-02-27T17:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.347218 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.347285 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.347309 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.347340 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.347359 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:10Z","lastTransitionTime":"2026-02-27T17:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.437597 4700 generic.go:334] "Generic (PLEG): container finished" podID="47cf6a8e-df81-4ccb-992b-abb030862331" containerID="c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af" exitCode=0 Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.437679 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" event={"ID":"47cf6a8e-df81-4ccb-992b-abb030862331","Type":"ContainerDied","Data":"c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af"} Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.441699 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"6fcde75bb1f9dd52a2ac77ffaa7d60d8c785c399bbe45a7cbff39d1b926d5541"} Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.449584 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.449638 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.449657 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.449682 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.449701 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:10Z","lastTransitionTime":"2026-02-27T17:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.457413 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a2d1b57-7fc1-4614-aac5-7144cdd812d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"27 17:01:51.956663 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 17:01:51.956671 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 17:01:51.956678 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 17:01:51.956715 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0227 17:01:51.960396 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1772211711\\\\\\\\\\\\\\\" (2026-02-27 17:01:50 +0000 UTC to 2026-03-29 17:01:51 +0000 UTC (now=2026-02-27 17:01:51.960134581 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960704 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1772211711\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1772211711\\\\\\\\\\\\\\\" (2026-02-27 16:01:51 +0000 UTC to 2027-02-27 16:01:51 +0000 UTC (now=2026-02-27 17:01:51.960667784 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960745 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0227 17:01:51.960790 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0227 17:01:51.960837 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960878 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960925 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\"\\\\nI0227 17:01:51.961071 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0227 17:01:51.958239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:01:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:10Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.469363 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:10Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.482086 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b45511-b94c-479f-98db-fd2c4eceec46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bddf1a7b7264d36174b90f1d0d4e49f029781b1acfce88582aa5b2575dcdd09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ef7c37b61de7a039756248f222bd1769b774703c406de5ad11abd9fb2f46dd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f9dq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:10Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.493307 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-s8k74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5288317-4810-4321-baa0-31a118a4dc36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-s8k74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:10Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.508257 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f425b79-78da-4f9e-a1ad-205a09466052\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3baf8927284b91f3edeec1dc906c2928955badb19062087790d720936b6f7624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd29e6d703945e0f366f2d109b7c35fdd6360c77091b165616e9832dddac1c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cxnv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:10Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.521834 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:10Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.534306 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-g66q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d4f1eeba71eeb576c1df386f5acaf2bc354e7ca08aa9aac0971f61db72f244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b6cx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-g66q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:10Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.550295 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rbptz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66eb22fb-e593-40ff-9ff3-5bb32912972d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f58f4814976a1e25bef53555d7f5d14fa1cc800c8c5220801960758d5262b552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2dtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rbptz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:10Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.554584 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.554640 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.554656 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.554687 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.554701 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:10Z","lastTransitionTime":"2026-02-27T17:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.564622 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:10Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.579354 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c501ee5970cba15e45e1f1c8f0e33768019511f94dbbe509a7bbfea2c5b4a064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f4cd0745c61d33ac717a4a655f00713e2bd935c3cbba23c4ca0913859f6016f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:10Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.593852 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:10Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.606836 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dnkln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e772eecd-127f-4b26-8021-010fb507ce0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6df6de510d7e55695db4c89cbe9db84f91aca0e4bfbf2ed852407047c32822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v92s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dnkln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:10Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.628197 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47cf6a8e-df81-4ccb-992b-abb030862331\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8n2wb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:10Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.645114 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a336c2f16e12b411f111f6291366aa1b466f6b405896c152d1667b832a589b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:10Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.658329 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.658375 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.658389 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.658406 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.658417 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:10Z","lastTransitionTime":"2026-02-27T17:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.672823 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a34c0e3-2513-4e81-b6b1-80d1230475fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jtbqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:10Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.686650 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dnkln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e772eecd-127f-4b26-8021-010fb507ce0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6df6de510d7e55695db4c89cbe9db84f91aca0e4bfbf2ed852407047c32822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v92s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dnkln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:10Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.706216 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47cf6a8e-df81-4ccb-992b-abb030862331\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8n2wb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:10Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.720338 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:10Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.734109 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c501ee5970cba15e45e1f1c8f0e33768019511f94dbbe509a7bbfea2c5b4a064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f4cd0745c61d33ac717a4a655f00713e2bd935c3cbba23c4ca0913859f6016f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:10Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.761437 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.761511 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.761526 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.761547 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.761559 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:10Z","lastTransitionTime":"2026-02-27T17:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.768315 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:10Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.785488 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a336c2f16e12b411f111f6291366aa1b466f6b405896c152d1667b832a589b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:10Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.808299 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a34c0e3-2513-4e81-b6b1-80d1230475fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jtbqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:10Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.823952 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-s8k74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5288317-4810-4321-baa0-31a118a4dc36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-s8k74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:10Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.837749 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f425b79-78da-4f9e-a1ad-205a09466052\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3baf8927284b91f3edeec1dc906c2928955badb19062087790d720936b6f7624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd29e6d703945e0f366f2d109b7c35fdd6360c77091b165616e9832dddac1c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cxnv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:10Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.851498 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a2d1b57-7fc1-4614-aac5-7144cdd812d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"27 17:01:51.956663 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 17:01:51.956671 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 17:01:51.956678 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 17:01:51.956715 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0227 17:01:51.960396 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1772211711\\\\\\\\\\\\\\\" (2026-02-27 17:01:50 +0000 UTC to 2026-03-29 17:01:51 +0000 UTC (now=2026-02-27 17:01:51.960134581 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960704 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1772211711\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1772211711\\\\\\\\\\\\\\\" (2026-02-27 16:01:51 +0000 UTC to 2027-02-27 16:01:51 +0000 UTC (now=2026-02-27 17:01:51.960667784 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960745 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0227 17:01:51.960790 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0227 17:01:51.960837 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960878 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960925 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\"\\\\nI0227 17:01:51.961071 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0227 17:01:51.958239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:01:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:10Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.864075 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.864119 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.864129 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.864145 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.864153 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:10Z","lastTransitionTime":"2026-02-27T17:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.865755 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:10Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.878101 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b45511-b94c-479f-98db-fd2c4eceec46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bddf1a7b7264d36174b90f1d0d4e49f029781b1acfce88582aa5b2575dcdd09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ef7c37b61de7a039756248f222bd1769b774703c406de5ad11abd9fb2f46dd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f9dq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:10Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.889729 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fcde75bb1f9dd52a2ac77ffaa7d60d8c785c399bbe45a7cbff39d1b926d5541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:10Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.901764 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-g66q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d4f1eeba71eeb576c1df386f5acaf2bc354e7ca08aa9aac0971f61db72f244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b6cx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-g66q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:10Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.915198 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rbptz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66eb22fb-e593-40ff-9ff3-5bb32912972d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f58f4814976a1e25bef53555d7f5d14fa1cc800c8c5220801960758d5262b552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2dtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rbptz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:10Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.966675 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.966724 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.966736 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.966755 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.966768 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:10Z","lastTransitionTime":"2026-02-27T17:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.992146 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:10Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:10 crc kubenswrapper[4700]: I0227 17:02:10.996364 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.009125 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c501ee5970cba15e45e1f1c8f0e33768019511f94dbbe509a7bbfea2c5b4a064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f4cd0745c61d33ac717a4a655f00713e2bd935c3cbba23c4ca0913859f6016f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.030048 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.040916 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dnkln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e772eecd-127f-4b26-8021-010fb507ce0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6df6de510d7e55695db4c89cbe9db84f91aca0e4bfbf2ed852407047c32822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v92s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dnkln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.054189 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47cf6a8e-df81-4ccb-992b-abb030862331\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8n2wb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.066424 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a336c2f16e12b411f111f6291366aa1b466f6b405896c152d1667b832a589b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.069268 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.069338 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.069355 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.069373 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.069386 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:11Z","lastTransitionTime":"2026-02-27T17:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.090003 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a34c0e3-2513-4e81-b6b1-80d1230475fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jtbqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.110226 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a2d1b57-7fc1-4614-aac5-7144cdd812d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"27 17:01:51.956663 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 17:01:51.956671 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 17:01:51.956678 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 17:01:51.956715 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0227 17:01:51.960396 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1772211711\\\\\\\\\\\\\\\" (2026-02-27 17:01:50 +0000 UTC to 2026-03-29 17:01:51 +0000 UTC (now=2026-02-27 17:01:51.960134581 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960704 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1772211711\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1772211711\\\\\\\\\\\\\\\" (2026-02-27 16:01:51 +0000 UTC to 2027-02-27 16:01:51 +0000 UTC (now=2026-02-27 17:01:51.960667784 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960745 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0227 17:01:51.960790 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0227 17:01:51.960837 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960878 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960925 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\"\\\\nI0227 17:01:51.961071 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0227 17:01:51.958239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:01:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.131998 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.150078 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b45511-b94c-479f-98db-fd2c4eceec46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bddf1a7b7264d36174b90f1d0d4e49f029781b1acfce88582aa5b2575dcdd09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ef7c37b61de7a039756248f222bd1769b774703c406de5ad11abd9fb2f46dd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f9dq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.163816 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-s8k74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5288317-4810-4321-baa0-31a118a4dc36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-s8k74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.172476 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.172507 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.172517 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.172533 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.172543 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:11Z","lastTransitionTime":"2026-02-27T17:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.174161 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f425b79-78da-4f9e-a1ad-205a09466052\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3baf8927284b91f3edeec1dc906c2928955badb19062087790d720936b6f7624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd29e6d703945e0f366f2d109b7c35fdd6360c77091b165616e9832dddac1c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cxnv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.185920 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fcde75bb1f9dd52a2ac77ffaa7d60d8c785c399bbe45a7cbff39d1b926d5541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.197910 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-g66q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d4f1eeba71eeb576c1df386f5acaf2bc354e7ca08aa9aac0971f61db72f244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b6cx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-g66q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.212601 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rbptz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66eb22fb-e593-40ff-9ff3-5bb32912972d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f58f4814976a1e25bef53555d7f5d14fa1cc800c8c5220801960758d5262b552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2dtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rbptz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.275981 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.276047 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.276063 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.276089 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.276105 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:11Z","lastTransitionTime":"2026-02-27T17:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.379813 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.379880 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.379895 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.379918 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.379934 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:11Z","lastTransitionTime":"2026-02-27T17:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.451533 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" event={"ID":"3a34c0e3-2513-4e81-b6b1-80d1230475fd","Type":"ContainerStarted","Data":"3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191"} Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.456018 4700 generic.go:334] "Generic (PLEG): container finished" podID="47cf6a8e-df81-4ccb-992b-abb030862331" containerID="fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942" exitCode=0 Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.456079 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" event={"ID":"47cf6a8e-df81-4ccb-992b-abb030862331","Type":"ContainerDied","Data":"fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942"} Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.479469 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fcde75bb1f9dd52a2ac77ffaa7d60d8c785c399bbe45a7cbff39d1b926d5541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.482632 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.482666 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.482679 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.482696 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.482707 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:11Z","lastTransitionTime":"2026-02-27T17:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.494215 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-g66q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d4f1eeba71eeb576c1df386f5acaf2bc354e7ca08aa9aac0971f61db72f244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b6cx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-g66q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.513253 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rbptz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66eb22fb-e593-40ff-9ff3-5bb32912972d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f58f4814976a1e25bef53555d7f5d14fa1cc800c8c5220801960758d5262b552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2dtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rbptz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.529432 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.549703 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c501ee5970cba15e45e1f1c8f0e33768019511f94dbbe509a7bbfea2c5b4a064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f4cd0745c61d33ac717a4a655f00713e2bd935c3cbba23c4ca0913859f6016f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.567290 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.581333 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dnkln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e772eecd-127f-4b26-8021-010fb507ce0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6df6de510d7e55695db4c89cbe9db84f91aca0e4bfbf2ed852407047c32822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v92s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dnkln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.585420 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.585449 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.585472 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.585488 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.585499 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:11Z","lastTransitionTime":"2026-02-27T17:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.602264 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47cf6a8e-df81-4ccb-992b-abb030862331\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8n2wb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.624062 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a336c2f16e12b411f111f6291366aa1b466f6b405896c152d1667b832a589b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.655242 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a34c0e3-2513-4e81-b6b1-80d1230475fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jtbqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.687338 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ee6013b-8a2d-4651-ae5f-98aaac016d22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af0992b3383953e562bf540bf7659d41e60aeee5d6103105eaab164797486dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a14230f224fb3b7b9b0ced93634313b66ac3abc376dca228bc5802890b0c393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f09ed7c87894bc0473c7015f0e0d55a28f893f5a2e8bf9085c86654c464371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa2391deb934395fe4743249887832731bf7c6ea457d66484bcb6b556e4ab33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64972d043accab95774784cc95dc1144e70c855eff777664c66cb121d519d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.689498 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.689719 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.689869 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.690035 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.690172 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:11Z","lastTransitionTime":"2026-02-27T17:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.710375 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a2d1b57-7fc1-4614-aac5-7144cdd812d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"27 17:01:51.956663 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 17:01:51.956671 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 17:01:51.956678 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 17:01:51.956715 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0227 17:01:51.960396 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1772211711\\\\\\\\\\\\\\\" (2026-02-27 17:01:50 +0000 UTC to 2026-03-29 17:01:51 +0000 UTC (now=2026-02-27 17:01:51.960134581 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960704 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1772211711\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1772211711\\\\\\\\\\\\\\\" (2026-02-27 16:01:51 +0000 UTC to 2027-02-27 16:01:51 +0000 UTC (now=2026-02-27 17:01:51.960667784 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960745 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0227 17:01:51.960790 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0227 17:01:51.960837 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960878 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960925 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\"\\\\nI0227 17:01:51.961071 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0227 17:01:51.958239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:01:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.732372 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.750022 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b45511-b94c-479f-98db-fd2c4eceec46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bddf1a7b7264d36174b90f1d0d4e49f029781b1acfce88582aa5b2575dcdd09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ef7c37b61de7a039756248f222bd1769b774703c406de5ad11abd9fb2f46dd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f9dq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.767091 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-s8k74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5288317-4810-4321-baa0-31a118a4dc36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-s8k74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.783092 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f425b79-78da-4f9e-a1ad-205a09466052\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3baf8927284b91f3edeec1dc906c2928955badb19062087790d720936b6f7624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd29e6d703945e0f366f2d109b7c35fdd6360c77091b165616e9832dddac1c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cxnv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.793399 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.793454 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.793502 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.793532 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.793550 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:11Z","lastTransitionTime":"2026-02-27T17:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.807755 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.807808 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.807820 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.807841 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.807857 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:11Z","lastTransitionTime":"2026-02-27T17:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:11 crc kubenswrapper[4700]: E0227 17:02:11.829108 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.834792 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.834857 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.834877 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.834905 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.834926 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:11Z","lastTransitionTime":"2026-02-27T17:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:11 crc kubenswrapper[4700]: E0227 17:02:11.850435 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.855903 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.855963 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.855983 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.856012 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.856033 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:11Z","lastTransitionTime":"2026-02-27T17:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:11 crc kubenswrapper[4700]: E0227 17:02:11.870850 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.877127 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.877174 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.877188 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.877212 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.877231 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:11Z","lastTransitionTime":"2026-02-27T17:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:11 crc kubenswrapper[4700]: E0227 17:02:11.896660 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.902229 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.902276 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.902290 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.902313 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.902329 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:11Z","lastTransitionTime":"2026-02-27T17:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:11 crc kubenswrapper[4700]: E0227 17:02:11.920378 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:11 crc kubenswrapper[4700]: E0227 17:02:11.920618 4700 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.922992 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.923042 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.923059 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.923082 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.923096 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:11Z","lastTransitionTime":"2026-02-27T17:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.980504 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.980633 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.980738 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:11 crc kubenswrapper[4700]: E0227 17:02:11.980649 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:02:11 crc kubenswrapper[4700]: E0227 17:02:11.980885 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:02:11 crc kubenswrapper[4700]: E0227 17:02:11.981018 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:02:11 crc kubenswrapper[4700]: I0227 17:02:11.981147 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:02:11 crc kubenswrapper[4700]: E0227 17:02:11.981230 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.026085 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.026153 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.026171 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.026190 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.026215 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:12Z","lastTransitionTime":"2026-02-27T17:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.128987 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.129065 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.129085 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.129117 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.129139 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:12Z","lastTransitionTime":"2026-02-27T17:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.231933 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.231989 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.232009 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.232032 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.232055 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:12Z","lastTransitionTime":"2026-02-27T17:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.334688 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.334723 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.334732 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.334746 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.334757 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:12Z","lastTransitionTime":"2026-02-27T17:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.437943 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.438004 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.438022 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.438049 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.438069 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:12Z","lastTransitionTime":"2026-02-27T17:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.468887 4700 generic.go:334] "Generic (PLEG): container finished" podID="47cf6a8e-df81-4ccb-992b-abb030862331" containerID="99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209" exitCode=0 Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.468950 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" event={"ID":"47cf6a8e-df81-4ccb-992b-abb030862331","Type":"ContainerDied","Data":"99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209"} Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.491017 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:12Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.513602 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c501ee5970cba15e45e1f1c8f0e33768019511f94dbbe509a7bbfea2c5b4a064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f4cd0745c61d33ac717a4a655f00713e2bd935c3cbba23c4ca0913859f6016f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:12Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.541633 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:12Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.542889 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.542960 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.542980 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.543039 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.543058 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:12Z","lastTransitionTime":"2026-02-27T17:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.559366 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dnkln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e772eecd-127f-4b26-8021-010fb507ce0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6df6de510d7e55695db4c89cbe9db84f91aca0e4bfbf2ed852407047c32822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v92s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dnkln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:12Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.583675 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47cf6a8e-df81-4ccb-992b-abb030862331\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8n2wb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:12Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.601621 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a336c2f16e12b411f111f6291366aa1b466f6b405896c152d1667b832a589b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:12Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.635869 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a34c0e3-2513-4e81-b6b1-80d1230475fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jtbqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:12Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.648587 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.648632 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.648647 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.648671 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.648688 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:12Z","lastTransitionTime":"2026-02-27T17:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.667761 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ee6013b-8a2d-4651-ae5f-98aaac016d22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af0992b3383953e562bf540bf7659d41e60aeee5d6103105eaab164797486dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a14230f224fb3b7b9b0ced93634313b66ac3abc376dca228bc5802890b0c393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f09ed7c87894bc0473c7015f0e0d55a28f893f5a2e8bf9085c86654c464371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa2391deb934395fe4743249887832731bf7c6ea457d66484bcb6b556e4ab33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64972d043accab95774784cc95dc1144e70c855eff777664c66cb121d519d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:12Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.691074 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a2d1b57-7fc1-4614-aac5-7144cdd812d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"27 17:01:51.956663 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 17:01:51.956671 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 17:01:51.956678 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 17:01:51.956715 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0227 17:01:51.960396 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1772211711\\\\\\\\\\\\\\\" (2026-02-27 17:01:50 +0000 UTC to 2026-03-29 17:01:51 +0000 UTC (now=2026-02-27 17:01:51.960134581 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960704 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1772211711\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1772211711\\\\\\\\\\\\\\\" (2026-02-27 16:01:51 +0000 UTC to 2027-02-27 16:01:51 +0000 UTC (now=2026-02-27 17:01:51.960667784 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960745 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0227 17:01:51.960790 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0227 17:01:51.960837 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960878 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960925 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\"\\\\nI0227 17:01:51.961071 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0227 17:01:51.958239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:01:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:12Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.714745 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:12Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.738771 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b45511-b94c-479f-98db-fd2c4eceec46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bddf1a7b7264d36174b90f1d0d4e49f029781b1acfce88582aa5b2575dcdd09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ef7c37b61de7a039756248f222bd1769b774703c406de5ad11abd9fb2f46dd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f9dq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:12Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.750851 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.750922 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.750942 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.750974 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.750997 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:12Z","lastTransitionTime":"2026-02-27T17:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.756309 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-s8k74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5288317-4810-4321-baa0-31a118a4dc36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-s8k74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:12Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.770555 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f425b79-78da-4f9e-a1ad-205a09466052\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3baf8927284b91f3edeec1dc906c2928955badb19062087790d720936b6f7624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd29e6d703945e0f366f2d109b7c35fdd6360c77091b165616e9832dddac1c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cxnv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:12Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.783956 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fcde75bb1f9dd52a2ac77ffaa7d60d8c785c399bbe45a7cbff39d1b926d5541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:12Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.797244 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-g66q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d4f1eeba71eeb576c1df386f5acaf2bc354e7ca08aa9aac0971f61db72f244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b6cx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-g66q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:12Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.813846 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rbptz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66eb22fb-e593-40ff-9ff3-5bb32912972d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f58f4814976a1e25bef53555d7f5d14fa1cc800c8c5220801960758d5262b552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2dtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rbptz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:12Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.853605 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.853654 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.853668 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.853688 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.853703 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:12Z","lastTransitionTime":"2026-02-27T17:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.957187 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.957252 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.957272 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.957302 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:12 crc kubenswrapper[4700]: I0227 17:02:12.957323 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:12Z","lastTransitionTime":"2026-02-27T17:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.004879 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.061688 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.061852 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.061886 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.062118 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.062216 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:13Z","lastTransitionTime":"2026-02-27T17:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.165031 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.165072 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.165082 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.165102 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.165112 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:13Z","lastTransitionTime":"2026-02-27T17:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.267832 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.267870 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.267881 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.267898 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.267911 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:13Z","lastTransitionTime":"2026-02-27T17:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.370309 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.370348 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.370358 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.370372 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.370384 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:13Z","lastTransitionTime":"2026-02-27T17:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.478313 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.478391 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.478426 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.478526 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.478551 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:13Z","lastTransitionTime":"2026-02-27T17:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.490270 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" event={"ID":"3a34c0e3-2513-4e81-b6b1-80d1230475fd","Type":"ContainerStarted","Data":"f77fe43a0dedd716026a95d501d3356910604728834ee1a167bfa5db37783fe7"} Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.490401 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.490595 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.490752 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.499642 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" event={"ID":"47cf6a8e-df81-4ccb-992b-abb030862331","Type":"ContainerStarted","Data":"857e486c0ff289c59f3de9933d1e0301be2f25b65e2b3f1fae6c4f149f900a4f"} Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.521531 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rbptz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66eb22fb-e593-40ff-9ff3-5bb32912972d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f58f4814976a1e25bef53555d7f5d14fa1cc800c8c5220801960758d5262b552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2dtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rbptz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:13Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.529906 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.534531 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.542107 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"032e16e8-09b3-456f-bebb-09d4b07bcebf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eabca65587272a0e2ce09d13f103fbbe5374fb6d5ed49bfa0e26de8336d9e683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:13Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.561478 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fcde75bb1f9dd52a2ac77ffaa7d60d8c785c399bbe45a7cbff39d1b926d5541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:13Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.577629 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-g66q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d4f1eeba71eeb576c1df386f5acaf2bc354e7ca08aa9aac0971f61db72f244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b6cx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-g66q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:13Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.582339 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.582394 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.582417 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.582446 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.582493 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:13Z","lastTransitionTime":"2026-02-27T17:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.599367 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:13Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.624255 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dnkln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e772eecd-127f-4b26-8021-010fb507ce0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6df6de510d7e55695db4c89cbe9db84f91aca0e4bfbf2ed852407047c32822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v92s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dnkln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:13Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.645923 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47cf6a8e-df81-4ccb-992b-abb030862331\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8n2wb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:13Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.658259 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:02:13 crc kubenswrapper[4700]: E0227 17:02:13.658539 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:02:21.658505591 +0000 UTC m=+101.643818348 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.658666 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.658749 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:02:13 crc kubenswrapper[4700]: E0227 17:02:13.658998 4700 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 17:02:13 crc kubenswrapper[4700]: E0227 17:02:13.659061 4700 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 17:02:13 crc kubenswrapper[4700]: E0227 17:02:13.659085 4700 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 17:02:13 crc kubenswrapper[4700]: E0227 17:02:13.659184 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-27 17:02:21.659144558 +0000 UTC m=+101.644457345 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 17:02:13 crc kubenswrapper[4700]: E0227 17:02:13.659297 4700 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 17:02:13 crc kubenswrapper[4700]: E0227 17:02:13.659361 4700 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 17:02:13 crc kubenswrapper[4700]: E0227 17:02:13.659385 4700 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.659512 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:13 crc kubenswrapper[4700]: E0227 17:02:13.659601 4700 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 17:02:13 crc kubenswrapper[4700]: E0227 17:02:13.659669 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-27 17:02:21.659529568 +0000 UTC m=+101.644842325 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.659697 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:13 crc kubenswrapper[4700]: E0227 17:02:13.659736 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 17:02:21.659727083 +0000 UTC m=+101.645039840 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 17:02:13 crc kubenswrapper[4700]: E0227 17:02:13.659752 4700 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 17:02:13 crc kubenswrapper[4700]: E0227 17:02:13.659801 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 17:02:21.659790255 +0000 UTC m=+101.645103012 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.671926 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:13Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.685562 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.685631 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.685642 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.685724 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.685738 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:13Z","lastTransitionTime":"2026-02-27T17:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.693474 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c501ee5970cba15e45e1f1c8f0e33768019511f94dbbe509a7bbfea2c5b4a064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f4cd0745c61d33ac717a4a655f00713e2bd935c3cbba23c4ca0913859f6016f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:13Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.715006 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a336c2f16e12b411f111f6291366aa1b466f6b405896c152d1667b832a589b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:13Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.745358 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a34c0e3-2513-4e81-b6b1-80d1230475fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f77fe43a0dedd716026a95d501d3356910604728834ee1a167bfa5db37783fe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jtbqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:13Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.761027 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f5288317-4810-4321-baa0-31a118a4dc36-metrics-certs\") pod \"network-metrics-daemon-s8k74\" (UID: \"f5288317-4810-4321-baa0-31a118a4dc36\") " pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:13 crc kubenswrapper[4700]: E0227 17:02:13.761280 4700 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 17:02:13 crc kubenswrapper[4700]: E0227 17:02:13.761445 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f5288317-4810-4321-baa0-31a118a4dc36-metrics-certs podName:f5288317-4810-4321-baa0-31a118a4dc36 nodeName:}" failed. No retries permitted until 2026-02-27 17:02:21.761419597 +0000 UTC m=+101.746732354 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f5288317-4810-4321-baa0-31a118a4dc36-metrics-certs") pod "network-metrics-daemon-s8k74" (UID: "f5288317-4810-4321-baa0-31a118a4dc36") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.764424 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b45511-b94c-479f-98db-fd2c4eceec46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bddf1a7b7264d36174b90f1d0d4e49f029781b1acfce88582aa5b2575dcdd09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ef7c37b61de7a039756248f222bd1769b774703c406de5ad11abd9fb2f46dd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f9dq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:13Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.783303 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-s8k74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5288317-4810-4321-baa0-31a118a4dc36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-s8k74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:13Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.788669 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.788796 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.788872 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.788961 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.789054 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:13Z","lastTransitionTime":"2026-02-27T17:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.805552 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f425b79-78da-4f9e-a1ad-205a09466052\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3baf8927284b91f3edeec1dc906c2928955badb19062087790d720936b6f7624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd29e6d703945e0f366f2d109b7c35fdd6360c77091b165616e9832dddac1c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cxnv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:13Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.841404 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ee6013b-8a2d-4651-ae5f-98aaac016d22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af0992b3383953e562bf540bf7659d41e60aeee5d6103105eaab164797486dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a14230f224fb3b7b9b0ced93634313b66ac3abc376dca228bc5802890b0c393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f09ed7c87894bc0473c7015f0e0d55a28f893f5a2e8bf9085c86654c464371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa2391deb934395fe4743249887832731bf7c6ea457d66484bcb6b556e4ab33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64972d043accab95774784cc95dc1144e70c855eff777664c66cb121d519d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:13Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.866565 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a2d1b57-7fc1-4614-aac5-7144cdd812d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"27 17:01:51.956663 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 17:01:51.956671 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 17:01:51.956678 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 17:01:51.956715 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0227 17:01:51.960396 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1772211711\\\\\\\\\\\\\\\" (2026-02-27 17:01:50 +0000 UTC to 2026-03-29 17:01:51 +0000 UTC (now=2026-02-27 17:01:51.960134581 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960704 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1772211711\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1772211711\\\\\\\\\\\\\\\" (2026-02-27 16:01:51 +0000 UTC to 2027-02-27 16:01:51 +0000 UTC (now=2026-02-27 17:01:51.960667784 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960745 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0227 17:01:51.960790 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0227 17:01:51.960837 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960878 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960925 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\"\\\\nI0227 17:01:51.961071 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0227 17:01:51.958239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:01:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:13Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.886728 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:13Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.892294 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.892571 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.892701 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.892829 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.892943 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:13Z","lastTransitionTime":"2026-02-27T17:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.902852 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"032e16e8-09b3-456f-bebb-09d4b07bcebf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eabca65587272a0e2ce09d13f103fbbe5374fb6d5ed49bfa0e26de8336d9e683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:13Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.922574 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fcde75bb1f9dd52a2ac77ffaa7d60d8c785c399bbe45a7cbff39d1b926d5541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:13Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.935227 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-g66q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d4f1eeba71eeb576c1df386f5acaf2bc354e7ca08aa9aac0971f61db72f244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b6cx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-g66q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:13Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.952381 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rbptz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66eb22fb-e593-40ff-9ff3-5bb32912972d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f58f4814976a1e25bef53555d7f5d14fa1cc800c8c5220801960758d5262b552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2dtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rbptz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:13Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.969509 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:13Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.981194 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.981300 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.981339 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.981198 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:02:13 crc kubenswrapper[4700]: E0227 17:02:13.981691 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:02:13 crc kubenswrapper[4700]: E0227 17:02:13.981820 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:02:13 crc kubenswrapper[4700]: E0227 17:02:13.981922 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:02:13 crc kubenswrapper[4700]: E0227 17:02:13.982009 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.991439 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c501ee5970cba15e45e1f1c8f0e33768019511f94dbbe509a7bbfea2c5b4a064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f4cd0745c61d33ac717a4a655f00713e2bd935c3cbba23c4ca0913859f6016f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:13Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.995609 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.995837 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.996077 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.996240 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:13 crc kubenswrapper[4700]: I0227 17:02:13.996394 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:13Z","lastTransitionTime":"2026-02-27T17:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.010379 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:14Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.025601 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dnkln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e772eecd-127f-4b26-8021-010fb507ce0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6df6de510d7e55695db4c89cbe9db84f91aca0e4bfbf2ed852407047c32822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v92s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dnkln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:14Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.047695 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47cf6a8e-df81-4ccb-992b-abb030862331\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857e486c0ff289c59f3de9933d1e0301be2f25b65e2b3f1fae6c4f149f900a4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8n2wb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:14Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.064786 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a336c2f16e12b411f111f6291366aa1b466f6b405896c152d1667b832a589b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:14Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.086069 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a34c0e3-2513-4e81-b6b1-80d1230475fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f77fe43a0dedd716026a95d501d3356910604728834ee1a167bfa5db37783fe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jtbqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:14Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.099917 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.099989 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.100281 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.100357 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.100674 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:14Z","lastTransitionTime":"2026-02-27T17:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.110715 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ee6013b-8a2d-4651-ae5f-98aaac016d22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af0992b3383953e562bf540bf7659d41e60aeee5d6103105eaab164797486dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a14230f224fb3b7b9b0ced93634313b66ac3abc376dca228bc5802890b0c393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f09ed7c87894bc0473c7015f0e0d55a28f893f5a2e8bf9085c86654c464371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa2391deb934395fe4743249887832731bf7c6ea457d66484bcb6b556e4ab33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64972d043accab95774784cc95dc1144e70c855eff777664c66cb121d519d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:14Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.130683 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a2d1b57-7fc1-4614-aac5-7144cdd812d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"27 17:01:51.956663 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 17:01:51.956671 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 17:01:51.956678 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 17:01:51.956715 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0227 17:01:51.960396 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1772211711\\\\\\\\\\\\\\\" (2026-02-27 17:01:50 +0000 UTC to 2026-03-29 17:01:51 +0000 UTC (now=2026-02-27 17:01:51.960134581 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960704 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1772211711\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1772211711\\\\\\\\\\\\\\\" (2026-02-27 16:01:51 +0000 UTC to 2027-02-27 16:01:51 +0000 UTC (now=2026-02-27 17:01:51.960667784 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960745 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0227 17:01:51.960790 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0227 17:01:51.960837 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960878 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960925 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\"\\\\nI0227 17:01:51.961071 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0227 17:01:51.958239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:01:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:14Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.145304 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:14Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.157635 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b45511-b94c-479f-98db-fd2c4eceec46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bddf1a7b7264d36174b90f1d0d4e49f029781b1acfce88582aa5b2575dcdd09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ef7c37b61de7a039756248f222bd1769b774703c406de5ad11abd9fb2f46dd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f9dq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:14Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.170767 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-s8k74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5288317-4810-4321-baa0-31a118a4dc36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-s8k74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:14Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.184967 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f425b79-78da-4f9e-a1ad-205a09466052\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3baf8927284b91f3edeec1dc906c2928955badb19062087790d720936b6f7624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd29e6d703945e0f366f2d109b7c35fdd6360c77091b165616e9832dddac1c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cxnv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:14Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.203739 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.203789 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.203802 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.203823 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.203836 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:14Z","lastTransitionTime":"2026-02-27T17:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.307878 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.307936 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.307954 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.307979 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.307998 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:14Z","lastTransitionTime":"2026-02-27T17:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.410352 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.410422 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.410447 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.410530 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.410556 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:14Z","lastTransitionTime":"2026-02-27T17:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.513018 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.513087 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.513113 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.513146 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.513170 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:14Z","lastTransitionTime":"2026-02-27T17:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.616685 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.617207 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.617235 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.617264 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.617286 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:14Z","lastTransitionTime":"2026-02-27T17:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.721073 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.721122 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.721137 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.721158 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.721174 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:14Z","lastTransitionTime":"2026-02-27T17:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.823126 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.823162 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.823173 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.823191 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.823201 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:14Z","lastTransitionTime":"2026-02-27T17:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.926020 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.926069 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.926089 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.926113 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:14 crc kubenswrapper[4700]: I0227 17:02:14.926130 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:14Z","lastTransitionTime":"2026-02-27T17:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.028639 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.028697 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.028714 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.028739 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.028756 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:15Z","lastTransitionTime":"2026-02-27T17:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.131779 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.131824 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.131841 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.131862 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.131881 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:15Z","lastTransitionTime":"2026-02-27T17:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.236291 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.236335 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.236354 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.236378 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.236394 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:15Z","lastTransitionTime":"2026-02-27T17:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.339066 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.339107 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.339117 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.339135 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.339146 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:15Z","lastTransitionTime":"2026-02-27T17:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.441870 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.441927 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.441944 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.441970 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.441986 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:15Z","lastTransitionTime":"2026-02-27T17:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.544161 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.544208 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.544221 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.544241 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.544253 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:15Z","lastTransitionTime":"2026-02-27T17:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.647202 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.647254 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.647269 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.647291 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.647306 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:15Z","lastTransitionTime":"2026-02-27T17:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.750359 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.750405 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.750418 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.750438 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.750449 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:15Z","lastTransitionTime":"2026-02-27T17:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.853961 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.854015 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.854033 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.854058 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.854075 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:15Z","lastTransitionTime":"2026-02-27T17:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.957487 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.957535 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.957548 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.957570 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.957584 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:15Z","lastTransitionTime":"2026-02-27T17:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.980587 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.980697 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.980586 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:02:15 crc kubenswrapper[4700]: E0227 17:02:15.980774 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:02:15 crc kubenswrapper[4700]: I0227 17:02:15.980591 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:02:15 crc kubenswrapper[4700]: E0227 17:02:15.980902 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:02:15 crc kubenswrapper[4700]: E0227 17:02:15.981010 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:02:15 crc kubenswrapper[4700]: E0227 17:02:15.981080 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.061786 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.061843 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.061863 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.061889 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.061905 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:16Z","lastTransitionTime":"2026-02-27T17:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.164861 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.164938 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.164960 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.164987 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.165005 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:16Z","lastTransitionTime":"2026-02-27T17:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.267592 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.267660 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.267691 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.267717 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.267734 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:16Z","lastTransitionTime":"2026-02-27T17:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.371200 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.371257 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.371273 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.371297 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.371313 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:16Z","lastTransitionTime":"2026-02-27T17:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.473714 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.473762 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.473782 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.473806 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.473824 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:16Z","lastTransitionTime":"2026-02-27T17:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.512311 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jtbqn_3a34c0e3-2513-4e81-b6b1-80d1230475fd/ovnkube-controller/0.log" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.516806 4700 generic.go:334] "Generic (PLEG): container finished" podID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerID="f77fe43a0dedd716026a95d501d3356910604728834ee1a167bfa5db37783fe7" exitCode=1 Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.516850 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" event={"ID":"3a34c0e3-2513-4e81-b6b1-80d1230475fd","Type":"ContainerDied","Data":"f77fe43a0dedd716026a95d501d3356910604728834ee1a167bfa5db37783fe7"} Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.518027 4700 scope.go:117] "RemoveContainer" containerID="f77fe43a0dedd716026a95d501d3356910604728834ee1a167bfa5db37783fe7" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.544082 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a2d1b57-7fc1-4614-aac5-7144cdd812d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"27 17:01:51.956663 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 17:01:51.956671 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 17:01:51.956678 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 17:01:51.956715 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0227 17:01:51.960396 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1772211711\\\\\\\\\\\\\\\" (2026-02-27 17:01:50 +0000 UTC to 2026-03-29 17:01:51 +0000 UTC (now=2026-02-27 17:01:51.960134581 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960704 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1772211711\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1772211711\\\\\\\\\\\\\\\" (2026-02-27 16:01:51 +0000 UTC to 2027-02-27 16:01:51 +0000 UTC (now=2026-02-27 17:01:51.960667784 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960745 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0227 17:01:51.960790 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0227 17:01:51.960837 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960878 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960925 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\"\\\\nI0227 17:01:51.961071 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0227 17:01:51.958239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:01:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:16Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.566768 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:16Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.576705 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.576758 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.576781 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.576809 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.576832 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:16Z","lastTransitionTime":"2026-02-27T17:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.581417 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b45511-b94c-479f-98db-fd2c4eceec46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bddf1a7b7264d36174b90f1d0d4e49f029781b1acfce88582aa5b2575dcdd09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ef7c37b61de7a039756248f222bd1769b774703c406de5ad11abd9fb2f46dd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f9dq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:16Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.595637 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-s8k74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5288317-4810-4321-baa0-31a118a4dc36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-s8k74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:16Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.611396 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f425b79-78da-4f9e-a1ad-205a09466052\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3baf8927284b91f3edeec1dc906c2928955badb19062087790d720936b6f7624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd29e6d703945e0f366f2d109b7c35fdd6360c77091b165616e9832dddac1c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cxnv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:16Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.643626 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ee6013b-8a2d-4651-ae5f-98aaac016d22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af0992b3383953e562bf540bf7659d41e60aeee5d6103105eaab164797486dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a14230f224fb3b7b9b0ced93634313b66ac3abc376dca228bc5802890b0c393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f09ed7c87894bc0473c7015f0e0d55a28f893f5a2e8bf9085c86654c464371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa2391deb934395fe4743249887832731bf7c6ea457d66484bcb6b556e4ab33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64972d043accab95774784cc95dc1144e70c855eff777664c66cb121d519d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:16Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.657058 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fcde75bb1f9dd52a2ac77ffaa7d60d8c785c399bbe45a7cbff39d1b926d5541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:16Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.671119 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-g66q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d4f1eeba71eeb576c1df386f5acaf2bc354e7ca08aa9aac0971f61db72f244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b6cx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-g66q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:16Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.682051 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.682098 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.682112 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.682132 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.682166 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:16Z","lastTransitionTime":"2026-02-27T17:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.689654 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rbptz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66eb22fb-e593-40ff-9ff3-5bb32912972d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f58f4814976a1e25bef53555d7f5d14fa1cc800c8c5220801960758d5262b552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2dtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rbptz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:16Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.703163 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"032e16e8-09b3-456f-bebb-09d4b07bcebf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eabca65587272a0e2ce09d13f103fbbe5374fb6d5ed49bfa0e26de8336d9e683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:16Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.714406 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:16Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.728571 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c501ee5970cba15e45e1f1c8f0e33768019511f94dbbe509a7bbfea2c5b4a064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f4cd0745c61d33ac717a4a655f00713e2bd935c3cbba23c4ca0913859f6016f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:16Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.743157 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:16Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.756740 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dnkln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e772eecd-127f-4b26-8021-010fb507ce0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6df6de510d7e55695db4c89cbe9db84f91aca0e4bfbf2ed852407047c32822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v92s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dnkln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:16Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.775046 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47cf6a8e-df81-4ccb-992b-abb030862331\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857e486c0ff289c59f3de9933d1e0301be2f25b65e2b3f1fae6c4f149f900a4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8n2wb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:16Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.789923 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.789969 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.789983 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.790005 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.790058 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:16Z","lastTransitionTime":"2026-02-27T17:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.796351 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a336c2f16e12b411f111f6291366aa1b466f6b405896c152d1667b832a589b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:16Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.815348 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a34c0e3-2513-4e81-b6b1-80d1230475fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f77fe43a0dedd716026a95d501d3356910604728834ee1a167bfa5db37783fe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f77fe43a0dedd716026a95d501d3356910604728834ee1a167bfa5db37783fe7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T17:02:15Z\\\",\\\"message\\\":\\\"m/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:15.756429 6559 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0227 17:02:15.756519 6559 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 17:02:15.756551 6559 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:15.756778 6559 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:15.756869 6559 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 17:02:15.756961 6559 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 17:02:15.758084 6559 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0227 17:02:15.758163 6559 factory.go:656] Stopping watch factory\\\\nI0227 17:02:15.758193 6559 ovnkube.go:599] Stopped ovnkube\\\\nI0227 17:02:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jtbqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:16Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.893789 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.893854 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.893872 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.893900 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.893916 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:16Z","lastTransitionTime":"2026-02-27T17:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.997000 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.997033 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.997042 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.997054 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:16 crc kubenswrapper[4700]: I0227 17:02:16.997062 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:16Z","lastTransitionTime":"2026-02-27T17:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.100308 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.100356 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.100368 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.100387 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.100399 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:17Z","lastTransitionTime":"2026-02-27T17:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.203240 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.203279 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.203288 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.203304 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.203313 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:17Z","lastTransitionTime":"2026-02-27T17:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.305307 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.305368 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.305386 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.305411 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.305428 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:17Z","lastTransitionTime":"2026-02-27T17:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.408553 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.408613 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.408632 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.408656 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.408675 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:17Z","lastTransitionTime":"2026-02-27T17:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.511225 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.511271 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.511290 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.511310 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.511322 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:17Z","lastTransitionTime":"2026-02-27T17:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.524008 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jtbqn_3a34c0e3-2513-4e81-b6b1-80d1230475fd/ovnkube-controller/0.log" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.527668 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" event={"ID":"3a34c0e3-2513-4e81-b6b1-80d1230475fd","Type":"ContainerStarted","Data":"95d94117a4d368a7846f104eb0fb339ec01a13ed85efda95050747db38da0320"} Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.528213 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.543019 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"032e16e8-09b3-456f-bebb-09d4b07bcebf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eabca65587272a0e2ce09d13f103fbbe5374fb6d5ed49bfa0e26de8336d9e683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:17Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.557899 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fcde75bb1f9dd52a2ac77ffaa7d60d8c785c399bbe45a7cbff39d1b926d5541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:17Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.570870 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-g66q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d4f1eeba71eeb576c1df386f5acaf2bc354e7ca08aa9aac0971f61db72f244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b6cx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-g66q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:17Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.591313 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rbptz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66eb22fb-e593-40ff-9ff3-5bb32912972d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f58f4814976a1e25bef53555d7f5d14fa1cc800c8c5220801960758d5262b552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2dtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rbptz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:17Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.608331 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:17Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.613027 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.613087 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.613106 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.613133 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.613152 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:17Z","lastTransitionTime":"2026-02-27T17:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.626289 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c501ee5970cba15e45e1f1c8f0e33768019511f94dbbe509a7bbfea2c5b4a064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f4cd0745c61d33ac717a4a655f00713e2bd935c3cbba23c4ca0913859f6016f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:17Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.640616 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:17Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.653128 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dnkln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e772eecd-127f-4b26-8021-010fb507ce0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6df6de510d7e55695db4c89cbe9db84f91aca0e4bfbf2ed852407047c32822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v92s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dnkln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:17Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.666265 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47cf6a8e-df81-4ccb-992b-abb030862331\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857e486c0ff289c59f3de9933d1e0301be2f25b65e2b3f1fae6c4f149f900a4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8n2wb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:17Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.687716 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a336c2f16e12b411f111f6291366aa1b466f6b405896c152d1667b832a589b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:17Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.708199 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a34c0e3-2513-4e81-b6b1-80d1230475fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95d94117a4d368a7846f104eb0fb339ec01a13ed85efda95050747db38da0320\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f77fe43a0dedd716026a95d501d3356910604728834ee1a167bfa5db37783fe7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T17:02:15Z\\\",\\\"message\\\":\\\"m/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:15.756429 6559 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0227 17:02:15.756519 6559 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 17:02:15.756551 6559 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:15.756778 6559 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:15.756869 6559 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 17:02:15.756961 6559 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 17:02:15.758084 6559 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0227 17:02:15.758163 6559 factory.go:656] Stopping watch factory\\\\nI0227 17:02:15.758193 6559 ovnkube.go:599] Stopped ovnkube\\\\nI0227 17:02:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jtbqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:17Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.715681 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.715732 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.715750 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.715775 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.715791 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:17Z","lastTransitionTime":"2026-02-27T17:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.730097 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ee6013b-8a2d-4651-ae5f-98aaac016d22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af0992b3383953e562bf540bf7659d41e60aeee5d6103105eaab164797486dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a14230f224fb3b7b9b0ced93634313b66ac3abc376dca228bc5802890b0c393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f09ed7c87894bc0473c7015f0e0d55a28f893f5a2e8bf9085c86654c464371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa2391deb934395fe4743249887832731bf7c6ea457d66484bcb6b556e4ab33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64972d043accab95774784cc95dc1144e70c855eff777664c66cb121d519d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:17Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.747568 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a2d1b57-7fc1-4614-aac5-7144cdd812d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"27 17:01:51.956663 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 17:01:51.956671 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 17:01:51.956678 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 17:01:51.956715 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0227 17:01:51.960396 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1772211711\\\\\\\\\\\\\\\" (2026-02-27 17:01:50 +0000 UTC to 2026-03-29 17:01:51 +0000 UTC (now=2026-02-27 17:01:51.960134581 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960704 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1772211711\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1772211711\\\\\\\\\\\\\\\" (2026-02-27 16:01:51 +0000 UTC to 2027-02-27 16:01:51 +0000 UTC (now=2026-02-27 17:01:51.960667784 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960745 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0227 17:01:51.960790 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0227 17:01:51.960837 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960878 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960925 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\"\\\\nI0227 17:01:51.961071 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0227 17:01:51.958239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:01:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:17Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.763121 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:17Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.775315 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b45511-b94c-479f-98db-fd2c4eceec46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bddf1a7b7264d36174b90f1d0d4e49f029781b1acfce88582aa5b2575dcdd09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ef7c37b61de7a039756248f222bd1769b774703c406de5ad11abd9fb2f46dd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f9dq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:17Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.785129 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-s8k74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5288317-4810-4321-baa0-31a118a4dc36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-s8k74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:17Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.805838 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f425b79-78da-4f9e-a1ad-205a09466052\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3baf8927284b91f3edeec1dc906c2928955badb19062087790d720936b6f7624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd29e6d703945e0f366f2d109b7c35fdd6360c77091b165616e9832dddac1c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cxnv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:17Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.818739 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.818782 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.818793 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.818818 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.818831 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:17Z","lastTransitionTime":"2026-02-27T17:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.921712 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.921774 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.921792 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.921819 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.921836 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:17Z","lastTransitionTime":"2026-02-27T17:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.980845 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.980922 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.980933 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:02:17 crc kubenswrapper[4700]: E0227 17:02:17.981046 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:02:17 crc kubenswrapper[4700]: I0227 17:02:17.981117 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:17 crc kubenswrapper[4700]: E0227 17:02:17.981218 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:02:17 crc kubenswrapper[4700]: E0227 17:02:17.981432 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:02:17 crc kubenswrapper[4700]: E0227 17:02:17.981605 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.024524 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.024571 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.024588 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.024615 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.024633 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:18Z","lastTransitionTime":"2026-02-27T17:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.127184 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.127288 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.127311 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.127341 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.127360 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:18Z","lastTransitionTime":"2026-02-27T17:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.229753 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.229833 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.229859 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.229884 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.229900 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:18Z","lastTransitionTime":"2026-02-27T17:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.332895 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.332964 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.332989 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.333024 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.333048 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:18Z","lastTransitionTime":"2026-02-27T17:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.436417 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.436524 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.436543 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.436568 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.436585 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:18Z","lastTransitionTime":"2026-02-27T17:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.536439 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jtbqn_3a34c0e3-2513-4e81-b6b1-80d1230475fd/ovnkube-controller/1.log" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.537577 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jtbqn_3a34c0e3-2513-4e81-b6b1-80d1230475fd/ovnkube-controller/0.log" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.538394 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.538435 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.538449 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.538492 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.538511 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:18Z","lastTransitionTime":"2026-02-27T17:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.541377 4700 generic.go:334] "Generic (PLEG): container finished" podID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerID="95d94117a4d368a7846f104eb0fb339ec01a13ed85efda95050747db38da0320" exitCode=1 Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.541427 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" event={"ID":"3a34c0e3-2513-4e81-b6b1-80d1230475fd","Type":"ContainerDied","Data":"95d94117a4d368a7846f104eb0fb339ec01a13ed85efda95050747db38da0320"} Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.541499 4700 scope.go:117] "RemoveContainer" containerID="f77fe43a0dedd716026a95d501d3356910604728834ee1a167bfa5db37783fe7" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.542589 4700 scope.go:117] "RemoveContainer" containerID="95d94117a4d368a7846f104eb0fb339ec01a13ed85efda95050747db38da0320" Feb 27 17:02:18 crc kubenswrapper[4700]: E0227 17:02:18.542891 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jtbqn_openshift-ovn-kubernetes(3a34c0e3-2513-4e81-b6b1-80d1230475fd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.570942 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:18Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.585208 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dnkln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e772eecd-127f-4b26-8021-010fb507ce0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6df6de510d7e55695db4c89cbe9db84f91aca0e4bfbf2ed852407047c32822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v92s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dnkln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:18Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.604245 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47cf6a8e-df81-4ccb-992b-abb030862331\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857e486c0ff289c59f3de9933d1e0301be2f25b65e2b3f1fae6c4f149f900a4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8n2wb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:18Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.619670 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:18Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.633290 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c501ee5970cba15e45e1f1c8f0e33768019511f94dbbe509a7bbfea2c5b4a064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f4cd0745c61d33ac717a4a655f00713e2bd935c3cbba23c4ca0913859f6016f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:18Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.641119 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.641149 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.641158 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.641174 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.641184 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:18Z","lastTransitionTime":"2026-02-27T17:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.647002 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a336c2f16e12b411f111f6291366aa1b466f6b405896c152d1667b832a589b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:18Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.663345 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a34c0e3-2513-4e81-b6b1-80d1230475fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95d94117a4d368a7846f104eb0fb339ec01a13ed85efda95050747db38da0320\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f77fe43a0dedd716026a95d501d3356910604728834ee1a167bfa5db37783fe7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T17:02:15Z\\\",\\\"message\\\":\\\"m/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:15.756429 6559 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0227 17:02:15.756519 6559 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 17:02:15.756551 6559 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:15.756778 6559 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:15.756869 6559 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 17:02:15.756961 6559 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 17:02:15.758084 6559 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0227 17:02:15.758163 6559 factory.go:656] Stopping watch factory\\\\nI0227 17:02:15.758193 6559 ovnkube.go:599] Stopped ovnkube\\\\nI0227 17:02:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95d94117a4d368a7846f104eb0fb339ec01a13ed85efda95050747db38da0320\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T17:02:17Z\\\",\\\"message\\\":\\\" syncing service certified-operators on namespace openshift-marketplace for network=default : 1.427497ms\\\\nI0227 17:02:17.571627 6694 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-apiserver/apiserver\\\\\\\"}\\\\nI0227 17:02:17.572157 6694 services_controller.go:360] Finished syncing service apiserver on namespace openshift-kube-apiserver for network=default : 2.941716ms\\\\nI0227 17:02:17.571720 6694 services_controller.go:360] Finished syncing service network-check-source on namespace openshift-network-diagnostics for network=default : 47.942µs\\\\nI0227 17:02:17.572342 6694 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console-operator/metrics\\\\\\\"}\\\\nI0227 17:02:17.572362 6694 services_controller.go:360] Finished syncing service metrics on namespace openshift-console-operator for network=default : 1.241712ms\\\\nI0227 17:02:17.572505 6694 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0227 17:02:17.572650 6694 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0227 17:02:17.572694 6694 ovnkube.go:599] Stopped ovnkube\\\\nI0227 17:02:17.572767 6694 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0227 17:02:17.572875 6694 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jtbqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:18Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.672584 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b45511-b94c-479f-98db-fd2c4eceec46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bddf1a7b7264d36174b90f1d0d4e49f029781b1acfce88582aa5b2575dcdd09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ef7c37b61de7a039756248f222bd1769b774703c406de5ad11abd9fb2f46dd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f9dq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:18Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.683632 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-s8k74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5288317-4810-4321-baa0-31a118a4dc36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-s8k74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:18Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.695240 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f425b79-78da-4f9e-a1ad-205a09466052\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3baf8927284b91f3edeec1dc906c2928955badb19062087790d720936b6f7624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd29e6d703945e0f366f2d109b7c35fdd6360c77091b165616e9832dddac1c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cxnv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:18Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.723115 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ee6013b-8a2d-4651-ae5f-98aaac016d22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af0992b3383953e562bf540bf7659d41e60aeee5d6103105eaab164797486dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a14230f224fb3b7b9b0ced93634313b66ac3abc376dca228bc5802890b0c393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f09ed7c87894bc0473c7015f0e0d55a28f893f5a2e8bf9085c86654c464371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa2391deb934395fe4743249887832731bf7c6ea457d66484bcb6b556e4ab33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64972d043accab95774784cc95dc1144e70c855eff777664c66cb121d519d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:18Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.739387 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a2d1b57-7fc1-4614-aac5-7144cdd812d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"27 17:01:51.956663 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 17:01:51.956671 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 17:01:51.956678 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 17:01:51.956715 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0227 17:01:51.960396 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1772211711\\\\\\\\\\\\\\\" (2026-02-27 17:01:50 +0000 UTC to 2026-03-29 17:01:51 +0000 UTC (now=2026-02-27 17:01:51.960134581 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960704 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1772211711\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1772211711\\\\\\\\\\\\\\\" (2026-02-27 16:01:51 +0000 UTC to 2027-02-27 16:01:51 +0000 UTC (now=2026-02-27 17:01:51.960667784 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960745 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0227 17:01:51.960790 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0227 17:01:51.960837 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960878 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960925 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\"\\\\nI0227 17:01:51.961071 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0227 17:01:51.958239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:01:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:18Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.743772 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.743839 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.743860 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.743892 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.743916 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:18Z","lastTransitionTime":"2026-02-27T17:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.761169 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:18Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.778929 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rbptz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66eb22fb-e593-40ff-9ff3-5bb32912972d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f58f4814976a1e25bef53555d7f5d14fa1cc800c8c5220801960758d5262b552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2dtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rbptz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:18Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.789484 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"032e16e8-09b3-456f-bebb-09d4b07bcebf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eabca65587272a0e2ce09d13f103fbbe5374fb6d5ed49bfa0e26de8336d9e683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:18Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.807924 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fcde75bb1f9dd52a2ac77ffaa7d60d8c785c399bbe45a7cbff39d1b926d5541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:18Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.820231 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-g66q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d4f1eeba71eeb576c1df386f5acaf2bc354e7ca08aa9aac0971f61db72f244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b6cx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-g66q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:18Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.847793 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.847837 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.847855 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.847881 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.847900 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:18Z","lastTransitionTime":"2026-02-27T17:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.950452 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.950534 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.950559 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.950590 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:18 crc kubenswrapper[4700]: I0227 17:02:18.950612 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:18Z","lastTransitionTime":"2026-02-27T17:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.053322 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.053360 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.053376 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.053397 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.053414 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:19Z","lastTransitionTime":"2026-02-27T17:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.156038 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.156091 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.156109 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.156134 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.156152 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:19Z","lastTransitionTime":"2026-02-27T17:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.258861 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.258919 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.258938 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.259000 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.259030 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:19Z","lastTransitionTime":"2026-02-27T17:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.361626 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.361669 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.361685 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.361708 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.361729 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:19Z","lastTransitionTime":"2026-02-27T17:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.464994 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.465096 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.465118 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.465140 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.465157 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:19Z","lastTransitionTime":"2026-02-27T17:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.548986 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jtbqn_3a34c0e3-2513-4e81-b6b1-80d1230475fd/ovnkube-controller/1.log" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.555172 4700 scope.go:117] "RemoveContainer" containerID="95d94117a4d368a7846f104eb0fb339ec01a13ed85efda95050747db38da0320" Feb 27 17:02:19 crc kubenswrapper[4700]: E0227 17:02:19.555516 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jtbqn_openshift-ovn-kubernetes(3a34c0e3-2513-4e81-b6b1-80d1230475fd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.572098 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.572163 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.572188 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.572217 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.572240 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:19Z","lastTransitionTime":"2026-02-27T17:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.576371 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f425b79-78da-4f9e-a1ad-205a09466052\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3baf8927284b91f3edeec1dc906c2928955badb19062087790d720936b6f7624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd29e6d703945e0f366f2d109b7c35fdd6360c77091b165616e9832dddac1c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cxnv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:19Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.611271 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ee6013b-8a2d-4651-ae5f-98aaac016d22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af0992b3383953e562bf540bf7659d41e60aeee5d6103105eaab164797486dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a14230f224fb3b7b9b0ced93634313b66ac3abc376dca228bc5802890b0c393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f09ed7c87894bc0473c7015f0e0d55a28f893f5a2e8bf9085c86654c464371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa2391deb934395fe4743249887832731bf7c6ea457d66484bcb6b556e4ab33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64972d043accab95774784cc95dc1144e70c855eff777664c66cb121d519d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:19Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.634633 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a2d1b57-7fc1-4614-aac5-7144cdd812d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"27 17:01:51.956663 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 17:01:51.956671 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 17:01:51.956678 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 17:01:51.956715 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0227 17:01:51.960396 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1772211711\\\\\\\\\\\\\\\" (2026-02-27 17:01:50 +0000 UTC to 2026-03-29 17:01:51 +0000 UTC (now=2026-02-27 17:01:51.960134581 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960704 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1772211711\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1772211711\\\\\\\\\\\\\\\" (2026-02-27 16:01:51 +0000 UTC to 2027-02-27 16:01:51 +0000 UTC (now=2026-02-27 17:01:51.960667784 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960745 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0227 17:01:51.960790 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0227 17:01:51.960837 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960878 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960925 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\"\\\\nI0227 17:01:51.961071 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0227 17:01:51.958239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:01:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:19Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.655125 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:19Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.675253 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.675301 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.675318 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.675341 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.675357 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:19Z","lastTransitionTime":"2026-02-27T17:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.676302 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b45511-b94c-479f-98db-fd2c4eceec46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bddf1a7b7264d36174b90f1d0d4e49f029781b1acfce88582aa5b2575dcdd09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ef7c37b61de7a039756248f222bd1769b774703c406de5ad11abd9fb2f46dd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f9dq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:19Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.690082 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-s8k74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5288317-4810-4321-baa0-31a118a4dc36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-s8k74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:19Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.704897 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"032e16e8-09b3-456f-bebb-09d4b07bcebf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eabca65587272a0e2ce09d13f103fbbe5374fb6d5ed49bfa0e26de8336d9e683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:19Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.720415 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fcde75bb1f9dd52a2ac77ffaa7d60d8c785c399bbe45a7cbff39d1b926d5541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:19Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.732777 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-g66q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d4f1eeba71eeb576c1df386f5acaf2bc354e7ca08aa9aac0971f61db72f244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b6cx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-g66q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:19Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.749129 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rbptz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66eb22fb-e593-40ff-9ff3-5bb32912972d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f58f4814976a1e25bef53555d7f5d14fa1cc800c8c5220801960758d5262b552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2dtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rbptz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:19Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.768586 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47cf6a8e-df81-4ccb-992b-abb030862331\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857e486c0ff289c59f3de9933d1e0301be2f25b65e2b3f1fae6c4f149f900a4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8n2wb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:19Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.778387 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.778454 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.778531 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.778564 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.778587 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:19Z","lastTransitionTime":"2026-02-27T17:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.788235 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:19Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.807071 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c501ee5970cba15e45e1f1c8f0e33768019511f94dbbe509a7bbfea2c5b4a064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f4cd0745c61d33ac717a4a655f00713e2bd935c3cbba23c4ca0913859f6016f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:19Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.821390 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:19Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.835142 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dnkln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e772eecd-127f-4b26-8021-010fb507ce0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6df6de510d7e55695db4c89cbe9db84f91aca0e4bfbf2ed852407047c32822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v92s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dnkln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:19Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.854712 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a336c2f16e12b411f111f6291366aa1b466f6b405896c152d1667b832a589b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:19Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.881142 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.881191 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.881207 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.881230 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.881246 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:19Z","lastTransitionTime":"2026-02-27T17:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.886432 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a34c0e3-2513-4e81-b6b1-80d1230475fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95d94117a4d368a7846f104eb0fb339ec01a13ed85efda95050747db38da0320\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95d94117a4d368a7846f104eb0fb339ec01a13ed85efda95050747db38da0320\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T17:02:17Z\\\",\\\"message\\\":\\\" syncing service certified-operators on namespace openshift-marketplace for network=default : 1.427497ms\\\\nI0227 17:02:17.571627 6694 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-apiserver/apiserver\\\\\\\"}\\\\nI0227 17:02:17.572157 6694 services_controller.go:360] Finished syncing service apiserver on namespace openshift-kube-apiserver for network=default : 2.941716ms\\\\nI0227 17:02:17.571720 6694 services_controller.go:360] Finished syncing service network-check-source on namespace openshift-network-diagnostics for network=default : 47.942µs\\\\nI0227 17:02:17.572342 6694 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console-operator/metrics\\\\\\\"}\\\\nI0227 17:02:17.572362 6694 services_controller.go:360] Finished syncing service metrics on namespace openshift-console-operator for network=default : 1.241712ms\\\\nI0227 17:02:17.572505 6694 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0227 17:02:17.572650 6694 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0227 17:02:17.572694 6694 ovnkube.go:599] Stopped ovnkube\\\\nI0227 17:02:17.572767 6694 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0227 17:02:17.572875 6694 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jtbqn_openshift-ovn-kubernetes(3a34c0e3-2513-4e81-b6b1-80d1230475fd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jtbqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:19Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.980919 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.981006 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.980972 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.980919 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:19 crc kubenswrapper[4700]: E0227 17:02:19.981124 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:02:19 crc kubenswrapper[4700]: E0227 17:02:19.981206 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:02:19 crc kubenswrapper[4700]: E0227 17:02:19.981389 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:02:19 crc kubenswrapper[4700]: E0227 17:02:19.981560 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.983969 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.984023 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.984043 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.984063 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:19 crc kubenswrapper[4700]: I0227 17:02:19.984080 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:19Z","lastTransitionTime":"2026-02-27T17:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.088099 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.088168 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.088187 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.088214 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.088231 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:20Z","lastTransitionTime":"2026-02-27T17:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.198308 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.198376 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.198394 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.198421 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.198439 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:20Z","lastTransitionTime":"2026-02-27T17:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.302684 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.303668 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.303734 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.303766 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.303788 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:20Z","lastTransitionTime":"2026-02-27T17:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.406904 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.406973 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.406993 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.407020 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.407040 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:20Z","lastTransitionTime":"2026-02-27T17:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.510806 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.510882 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.510895 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.510916 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.510931 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:20Z","lastTransitionTime":"2026-02-27T17:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.614496 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.614564 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.614589 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.614622 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.614642 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:20Z","lastTransitionTime":"2026-02-27T17:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.718292 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.718352 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.718369 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.718434 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.718452 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:20Z","lastTransitionTime":"2026-02-27T17:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.821557 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.821626 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.821648 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.821682 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.821706 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:20Z","lastTransitionTime":"2026-02-27T17:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.925272 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.925355 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.925378 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.925413 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:20 crc kubenswrapper[4700]: I0227 17:02:20.925435 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:20Z","lastTransitionTime":"2026-02-27T17:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.001171 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:20Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.025648 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c501ee5970cba15e45e1f1c8f0e33768019511f94dbbe509a7bbfea2c5b4a064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f4cd0745c61d33ac717a4a655f00713e2bd935c3cbba23c4ca0913859f6016f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:21Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.028304 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.028375 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.028397 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.028425 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.028443 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:21Z","lastTransitionTime":"2026-02-27T17:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.045044 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:21Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.061999 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dnkln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e772eecd-127f-4b26-8021-010fb507ce0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6df6de510d7e55695db4c89cbe9db84f91aca0e4bfbf2ed852407047c32822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v92s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dnkln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:21Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.088396 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47cf6a8e-df81-4ccb-992b-abb030862331\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857e486c0ff289c59f3de9933d1e0301be2f25b65e2b3f1fae6c4f149f900a4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8n2wb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:21Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.115955 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a336c2f16e12b411f111f6291366aa1b466f6b405896c152d1667b832a589b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:21Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.131587 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.131627 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.131646 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.131671 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.131690 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:21Z","lastTransitionTime":"2026-02-27T17:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.147114 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a34c0e3-2513-4e81-b6b1-80d1230475fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95d94117a4d368a7846f104eb0fb339ec01a13ed85efda95050747db38da0320\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95d94117a4d368a7846f104eb0fb339ec01a13ed85efda95050747db38da0320\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T17:02:17Z\\\",\\\"message\\\":\\\" syncing service certified-operators on namespace openshift-marketplace for network=default : 1.427497ms\\\\nI0227 17:02:17.571627 6694 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-apiserver/apiserver\\\\\\\"}\\\\nI0227 17:02:17.572157 6694 services_controller.go:360] Finished syncing service apiserver on namespace openshift-kube-apiserver for network=default : 2.941716ms\\\\nI0227 17:02:17.571720 6694 services_controller.go:360] Finished syncing service network-check-source on namespace openshift-network-diagnostics for network=default : 47.942µs\\\\nI0227 17:02:17.572342 6694 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console-operator/metrics\\\\\\\"}\\\\nI0227 17:02:17.572362 6694 services_controller.go:360] Finished syncing service metrics on namespace openshift-console-operator for network=default : 1.241712ms\\\\nI0227 17:02:17.572505 6694 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0227 17:02:17.572650 6694 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0227 17:02:17.572694 6694 ovnkube.go:599] Stopped ovnkube\\\\nI0227 17:02:17.572767 6694 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0227 17:02:17.572875 6694 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jtbqn_openshift-ovn-kubernetes(3a34c0e3-2513-4e81-b6b1-80d1230475fd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jtbqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:21Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.182328 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ee6013b-8a2d-4651-ae5f-98aaac016d22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af0992b3383953e562bf540bf7659d41e60aeee5d6103105eaab164797486dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a14230f224fb3b7b9b0ced93634313b66ac3abc376dca228bc5802890b0c393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f09ed7c87894bc0473c7015f0e0d55a28f893f5a2e8bf9085c86654c464371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa2391deb934395fe4743249887832731bf7c6ea457d66484bcb6b556e4ab33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64972d043accab95774784cc95dc1144e70c855eff777664c66cb121d519d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:21Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.209208 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a2d1b57-7fc1-4614-aac5-7144cdd812d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"27 17:01:51.956663 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 17:01:51.956671 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 17:01:51.956678 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 17:01:51.956715 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0227 17:01:51.960396 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1772211711\\\\\\\\\\\\\\\" (2026-02-27 17:01:50 +0000 UTC to 2026-03-29 17:01:51 +0000 UTC (now=2026-02-27 17:01:51.960134581 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960704 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1772211711\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1772211711\\\\\\\\\\\\\\\" (2026-02-27 16:01:51 +0000 UTC to 2027-02-27 16:01:51 +0000 UTC (now=2026-02-27 17:01:51.960667784 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960745 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0227 17:01:51.960790 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0227 17:01:51.960837 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960878 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960925 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\"\\\\nI0227 17:01:51.961071 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0227 17:01:51.958239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:01:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:21Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.231327 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:21Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.234323 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.234360 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.234373 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.234392 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.234405 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:21Z","lastTransitionTime":"2026-02-27T17:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.254263 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b45511-b94c-479f-98db-fd2c4eceec46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bddf1a7b7264d36174b90f1d0d4e49f029781b1acfce88582aa5b2575dcdd09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ef7c37b61de7a039756248f222bd1769b774703c406de5ad11abd9fb2f46dd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f9dq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:21Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.272765 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-s8k74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5288317-4810-4321-baa0-31a118a4dc36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-s8k74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:21Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.289838 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f425b79-78da-4f9e-a1ad-205a09466052\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3baf8927284b91f3edeec1dc906c2928955badb19062087790d720936b6f7624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd29e6d703945e0f366f2d109b7c35fdd6360c77091b165616e9832dddac1c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cxnv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:21Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.307337 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"032e16e8-09b3-456f-bebb-09d4b07bcebf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eabca65587272a0e2ce09d13f103fbbe5374fb6d5ed49bfa0e26de8336d9e683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:21Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.326385 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fcde75bb1f9dd52a2ac77ffaa7d60d8c785c399bbe45a7cbff39d1b926d5541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:21Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.337265 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.337311 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.337328 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.337351 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.337369 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:21Z","lastTransitionTime":"2026-02-27T17:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.343777 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-g66q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d4f1eeba71eeb576c1df386f5acaf2bc354e7ca08aa9aac0971f61db72f244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b6cx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-g66q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:21Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.365802 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rbptz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66eb22fb-e593-40ff-9ff3-5bb32912972d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f58f4814976a1e25bef53555d7f5d14fa1cc800c8c5220801960758d5262b552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2dtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rbptz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:21Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.440894 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.440967 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.440986 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.441011 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.441031 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:21Z","lastTransitionTime":"2026-02-27T17:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.544265 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.544335 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.544347 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.544367 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.544382 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:21Z","lastTransitionTime":"2026-02-27T17:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.647284 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.647346 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.647359 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.647378 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.647389 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:21Z","lastTransitionTime":"2026-02-27T17:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.750369 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.750412 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.750424 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.750443 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.750485 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:21Z","lastTransitionTime":"2026-02-27T17:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.752030 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.752176 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:02:21 crc kubenswrapper[4700]: E0227 17:02:21.752196 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:02:37.75217391 +0000 UTC m=+117.737486667 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.752236 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.752305 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.752356 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:21 crc kubenswrapper[4700]: E0227 17:02:21.752370 4700 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 17:02:21 crc kubenswrapper[4700]: E0227 17:02:21.752397 4700 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 17:02:21 crc kubenswrapper[4700]: E0227 17:02:21.752411 4700 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 17:02:21 crc kubenswrapper[4700]: E0227 17:02:21.752494 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-27 17:02:37.752477208 +0000 UTC m=+117.737789965 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 17:02:21 crc kubenswrapper[4700]: E0227 17:02:21.752520 4700 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 17:02:21 crc kubenswrapper[4700]: E0227 17:02:21.752527 4700 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 17:02:21 crc kubenswrapper[4700]: E0227 17:02:21.752577 4700 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 17:02:21 crc kubenswrapper[4700]: E0227 17:02:21.752605 4700 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 17:02:21 crc kubenswrapper[4700]: E0227 17:02:21.752645 4700 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 17:02:21 crc kubenswrapper[4700]: E0227 17:02:21.752580 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 17:02:37.752561761 +0000 UTC m=+117.737874548 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 17:02:21 crc kubenswrapper[4700]: E0227 17:02:21.752706 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 17:02:37.752694034 +0000 UTC m=+117.738006791 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 17:02:21 crc kubenswrapper[4700]: E0227 17:02:21.752732 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-27 17:02:37.752724455 +0000 UTC m=+117.738037212 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.853383 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f5288317-4810-4321-baa0-31a118a4dc36-metrics-certs\") pod \"network-metrics-daemon-s8k74\" (UID: \"f5288317-4810-4321-baa0-31a118a4dc36\") " pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:21 crc kubenswrapper[4700]: E0227 17:02:21.853701 4700 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.853790 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.854266 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.854286 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.854311 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.854330 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:21Z","lastTransitionTime":"2026-02-27T17:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:21 crc kubenswrapper[4700]: E0227 17:02:21.854226 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f5288317-4810-4321-baa0-31a118a4dc36-metrics-certs podName:f5288317-4810-4321-baa0-31a118a4dc36 nodeName:}" failed. No retries permitted until 2026-02-27 17:02:37.854199153 +0000 UTC m=+117.839511940 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f5288317-4810-4321-baa0-31a118a4dc36-metrics-certs") pod "network-metrics-daemon-s8k74" (UID: "f5288317-4810-4321-baa0-31a118a4dc36") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.957862 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.957931 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.957956 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.957987 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.958008 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:21Z","lastTransitionTime":"2026-02-27T17:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.980617 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.980678 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.980716 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:21 crc kubenswrapper[4700]: I0227 17:02:21.980677 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:21 crc kubenswrapper[4700]: E0227 17:02:21.980858 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:02:21 crc kubenswrapper[4700]: E0227 17:02:21.981033 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:02:21 crc kubenswrapper[4700]: E0227 17:02:21.981162 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:02:21 crc kubenswrapper[4700]: E0227 17:02:21.981226 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.029880 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.029932 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.029944 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.029966 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.029978 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:22Z","lastTransitionTime":"2026-02-27T17:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:22 crc kubenswrapper[4700]: E0227 17:02:22.046832 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:22Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.051268 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.051305 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.051321 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.051348 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.051364 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:22Z","lastTransitionTime":"2026-02-27T17:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:22 crc kubenswrapper[4700]: E0227 17:02:22.068509 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:22Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.073247 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.073287 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.073301 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.073316 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.073331 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:22Z","lastTransitionTime":"2026-02-27T17:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:22 crc kubenswrapper[4700]: E0227 17:02:22.090147 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:22Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.094871 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.094930 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.094953 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.094983 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.095005 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:22Z","lastTransitionTime":"2026-02-27T17:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:22 crc kubenswrapper[4700]: E0227 17:02:22.113859 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:22Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.120162 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.120209 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.120225 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.120250 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.120269 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:22Z","lastTransitionTime":"2026-02-27T17:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:22 crc kubenswrapper[4700]: E0227 17:02:22.139368 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:22Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:22 crc kubenswrapper[4700]: E0227 17:02:22.139635 4700 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.141906 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.141952 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.141967 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.141992 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.142010 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:22Z","lastTransitionTime":"2026-02-27T17:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.245577 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.245646 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.245663 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.245686 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.245707 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:22Z","lastTransitionTime":"2026-02-27T17:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.348970 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.349039 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.349058 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.349084 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.349102 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:22Z","lastTransitionTime":"2026-02-27T17:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.452362 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.452417 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.452434 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.452492 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.452512 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:22Z","lastTransitionTime":"2026-02-27T17:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.555739 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.555819 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.555845 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.555879 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.555905 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:22Z","lastTransitionTime":"2026-02-27T17:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.658875 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.658971 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.658997 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.659032 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.659057 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:22Z","lastTransitionTime":"2026-02-27T17:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.762933 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.762993 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.763010 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.763039 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.763058 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:22Z","lastTransitionTime":"2026-02-27T17:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.866159 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.866237 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.866261 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.866291 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.866313 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:22Z","lastTransitionTime":"2026-02-27T17:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.969175 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.969243 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.969269 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.969301 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:22 crc kubenswrapper[4700]: I0227 17:02:22.969325 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:22Z","lastTransitionTime":"2026-02-27T17:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.072133 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.072211 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.072236 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.072263 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.072281 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:23Z","lastTransitionTime":"2026-02-27T17:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.175881 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.175949 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.175967 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.175994 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.176023 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:23Z","lastTransitionTime":"2026-02-27T17:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.280457 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.280584 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.280608 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.280639 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.280656 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:23Z","lastTransitionTime":"2026-02-27T17:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.384085 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.384172 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.384199 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.384234 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.384257 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:23Z","lastTransitionTime":"2026-02-27T17:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.487782 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.487862 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.487878 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.487922 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.487946 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:23Z","lastTransitionTime":"2026-02-27T17:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.590963 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.591027 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.591040 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.591056 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.591068 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:23Z","lastTransitionTime":"2026-02-27T17:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.694074 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.694138 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.694160 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.694189 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.694211 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:23Z","lastTransitionTime":"2026-02-27T17:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.797256 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.797337 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.797364 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.797400 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.797425 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:23Z","lastTransitionTime":"2026-02-27T17:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.901376 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.901498 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.901527 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.901562 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.901589 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:23Z","lastTransitionTime":"2026-02-27T17:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.980968 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.981047 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.981120 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:23 crc kubenswrapper[4700]: E0227 17:02:23.981116 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:02:23 crc kubenswrapper[4700]: E0227 17:02:23.981224 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.981254 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:23 crc kubenswrapper[4700]: E0227 17:02:23.981372 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:02:23 crc kubenswrapper[4700]: E0227 17:02:23.981547 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:02:23 crc kubenswrapper[4700]: I0227 17:02:23.982338 4700 scope.go:117] "RemoveContainer" containerID="26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d" Feb 27 17:02:23 crc kubenswrapper[4700]: E0227 17:02:23.982641 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.003841 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.003956 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.003979 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.004002 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.004021 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:24Z","lastTransitionTime":"2026-02-27T17:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.107690 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.107756 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.107772 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.107796 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.107812 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:24Z","lastTransitionTime":"2026-02-27T17:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.211333 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.211391 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.211408 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.211432 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.211453 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:24Z","lastTransitionTime":"2026-02-27T17:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.314026 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.314091 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.314113 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.314149 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.314183 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:24Z","lastTransitionTime":"2026-02-27T17:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.416867 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.416906 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.416923 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.416943 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.416957 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:24Z","lastTransitionTime":"2026-02-27T17:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.519763 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.519826 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.519843 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.519867 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.519886 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:24Z","lastTransitionTime":"2026-02-27T17:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.623265 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.623327 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.623345 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.623372 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.623389 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:24Z","lastTransitionTime":"2026-02-27T17:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.726398 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.726487 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.726507 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.726532 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.726552 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:24Z","lastTransitionTime":"2026-02-27T17:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.829675 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.829723 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.829767 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.829790 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.829807 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:24Z","lastTransitionTime":"2026-02-27T17:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.934314 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.934371 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.934387 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.934412 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:24 crc kubenswrapper[4700]: I0227 17:02:24.934430 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:24Z","lastTransitionTime":"2026-02-27T17:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.038183 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.038282 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.038346 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.038378 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.038437 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:25Z","lastTransitionTime":"2026-02-27T17:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.141559 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.141718 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.141740 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.141764 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.141816 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:25Z","lastTransitionTime":"2026-02-27T17:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.245198 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.245290 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.245308 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.245330 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.245383 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:25Z","lastTransitionTime":"2026-02-27T17:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.348523 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.348587 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.348603 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.348628 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.348646 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:25Z","lastTransitionTime":"2026-02-27T17:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.452447 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.452563 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.452583 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.452605 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.452617 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:25Z","lastTransitionTime":"2026-02-27T17:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.555851 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.556004 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.556024 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.556051 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.556111 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:25Z","lastTransitionTime":"2026-02-27T17:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.659779 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.659842 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.659861 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.659885 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.659903 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:25Z","lastTransitionTime":"2026-02-27T17:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.763336 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.764224 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.764266 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.764296 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.764317 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:25Z","lastTransitionTime":"2026-02-27T17:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.867664 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.867824 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.867847 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.867871 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.867888 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:25Z","lastTransitionTime":"2026-02-27T17:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.971816 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.971883 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.971903 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.971930 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.971950 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:25Z","lastTransitionTime":"2026-02-27T17:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.980448 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.980525 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.980587 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:25 crc kubenswrapper[4700]: I0227 17:02:25.980681 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:02:25 crc kubenswrapper[4700]: E0227 17:02:25.980674 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:02:25 crc kubenswrapper[4700]: E0227 17:02:25.980826 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:02:25 crc kubenswrapper[4700]: E0227 17:02:25.980967 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:02:25 crc kubenswrapper[4700]: E0227 17:02:25.981188 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.075281 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.075340 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.075361 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.075385 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.075403 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:26Z","lastTransitionTime":"2026-02-27T17:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.177625 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.177687 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.177728 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.177760 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.177781 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:26Z","lastTransitionTime":"2026-02-27T17:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.281999 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.282072 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.282097 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.282131 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.282152 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:26Z","lastTransitionTime":"2026-02-27T17:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.385144 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.385203 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.385229 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.385258 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.385281 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:26Z","lastTransitionTime":"2026-02-27T17:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.488202 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.488281 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.488307 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.488335 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.488352 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:26Z","lastTransitionTime":"2026-02-27T17:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.591974 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.592419 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.592442 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.592552 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.592580 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:26Z","lastTransitionTime":"2026-02-27T17:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.695187 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.695252 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.695270 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.695296 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.695315 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:26Z","lastTransitionTime":"2026-02-27T17:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.798215 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.798265 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.798282 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.798319 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.798337 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:26Z","lastTransitionTime":"2026-02-27T17:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.901823 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.901924 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.901944 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.901972 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:26 crc kubenswrapper[4700]: I0227 17:02:26.901989 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:26Z","lastTransitionTime":"2026-02-27T17:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.005816 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.005865 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.005882 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.005906 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.005924 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:27Z","lastTransitionTime":"2026-02-27T17:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.108863 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.108951 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.108968 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.108990 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.109008 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:27Z","lastTransitionTime":"2026-02-27T17:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.211593 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.211642 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.211664 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.211687 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.211704 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:27Z","lastTransitionTime":"2026-02-27T17:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.314068 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.314131 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.314153 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.314187 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.314210 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:27Z","lastTransitionTime":"2026-02-27T17:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.417387 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.417496 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.417522 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.417553 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.417577 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:27Z","lastTransitionTime":"2026-02-27T17:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.520358 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.520412 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.520427 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.520450 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.520502 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:27Z","lastTransitionTime":"2026-02-27T17:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.624208 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.624258 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.624274 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.624298 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.624316 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:27Z","lastTransitionTime":"2026-02-27T17:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.727246 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.727311 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.727335 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.727364 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.727388 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:27Z","lastTransitionTime":"2026-02-27T17:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.830926 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.831002 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.831026 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.831056 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.831082 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:27Z","lastTransitionTime":"2026-02-27T17:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.933727 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.933781 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.933798 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.933822 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.933840 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:27Z","lastTransitionTime":"2026-02-27T17:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.980533 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.980579 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.980550 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:27 crc kubenswrapper[4700]: I0227 17:02:27.980550 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:27 crc kubenswrapper[4700]: E0227 17:02:27.980699 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:02:27 crc kubenswrapper[4700]: E0227 17:02:27.980858 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:02:27 crc kubenswrapper[4700]: E0227 17:02:27.980982 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:02:27 crc kubenswrapper[4700]: E0227 17:02:27.981152 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.036444 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.036534 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.036552 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.036575 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.036591 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:28Z","lastTransitionTime":"2026-02-27T17:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.140097 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.140157 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.140173 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.140197 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.140214 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:28Z","lastTransitionTime":"2026-02-27T17:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.243145 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.243208 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.243226 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.243250 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.243266 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:28Z","lastTransitionTime":"2026-02-27T17:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.346500 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.346551 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.346568 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.346593 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.346612 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:28Z","lastTransitionTime":"2026-02-27T17:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.449335 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.449403 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.449420 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.449446 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.449496 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:28Z","lastTransitionTime":"2026-02-27T17:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.553716 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.553773 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.553790 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.553813 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.553833 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:28Z","lastTransitionTime":"2026-02-27T17:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.656007 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.656080 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.656104 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.656136 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.656157 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:28Z","lastTransitionTime":"2026-02-27T17:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.759794 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.759870 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.759895 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.759928 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.759951 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:28Z","lastTransitionTime":"2026-02-27T17:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.862720 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.862795 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.862816 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.862841 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.862862 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:28Z","lastTransitionTime":"2026-02-27T17:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.966217 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.966266 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.966284 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.966306 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:28 crc kubenswrapper[4700]: I0227 17:02:28.966322 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:28Z","lastTransitionTime":"2026-02-27T17:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.069211 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.069271 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.069290 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.069313 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.069332 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:29Z","lastTransitionTime":"2026-02-27T17:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.172434 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.172574 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.172596 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.172623 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.172638 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:29Z","lastTransitionTime":"2026-02-27T17:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.275981 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.276035 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.276052 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.276073 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.276089 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:29Z","lastTransitionTime":"2026-02-27T17:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.379108 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.379193 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.379280 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.379323 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.379342 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:29Z","lastTransitionTime":"2026-02-27T17:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.483240 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.483299 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.483316 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.483340 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.483357 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:29Z","lastTransitionTime":"2026-02-27T17:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.586751 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.586816 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.586839 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.586870 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.586893 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:29Z","lastTransitionTime":"2026-02-27T17:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.689584 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.689624 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.689632 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.689645 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.689653 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:29Z","lastTransitionTime":"2026-02-27T17:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.792980 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.793056 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.793073 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.793097 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.793116 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:29Z","lastTransitionTime":"2026-02-27T17:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.895642 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.895717 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.895738 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.895767 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.895828 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:29Z","lastTransitionTime":"2026-02-27T17:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.980978 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.981061 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.981062 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:02:29 crc kubenswrapper[4700]: E0227 17:02:29.981184 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.981207 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:29 crc kubenswrapper[4700]: E0227 17:02:29.981349 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:02:29 crc kubenswrapper[4700]: E0227 17:02:29.981572 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:02:29 crc kubenswrapper[4700]: E0227 17:02:29.981635 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.999550 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.999614 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.999636 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.999663 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:29 crc kubenswrapper[4700]: I0227 17:02:29.999680 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:29Z","lastTransitionTime":"2026-02-27T17:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.103502 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.104122 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.104296 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.104515 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.104687 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:30Z","lastTransitionTime":"2026-02-27T17:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.207971 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.208043 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.208061 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.208088 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.208106 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:30Z","lastTransitionTime":"2026-02-27T17:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.310773 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.310840 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.310858 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.310884 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.310902 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:30Z","lastTransitionTime":"2026-02-27T17:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.414594 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.414664 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.414683 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.414707 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.414723 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:30Z","lastTransitionTime":"2026-02-27T17:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.517391 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.517482 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.517505 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.517538 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.517561 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:30Z","lastTransitionTime":"2026-02-27T17:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.620520 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.620578 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.620598 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.620622 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.620641 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:30Z","lastTransitionTime":"2026-02-27T17:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.723284 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.723344 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.723361 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.723385 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.723405 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:30Z","lastTransitionTime":"2026-02-27T17:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.826560 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.826626 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.826642 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.826666 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.826685 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:30Z","lastTransitionTime":"2026-02-27T17:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.930293 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.930354 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.930370 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.930395 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.930414 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:30Z","lastTransitionTime":"2026-02-27T17:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:30 crc kubenswrapper[4700]: I0227 17:02:30.981718 4700 scope.go:117] "RemoveContainer" containerID="95d94117a4d368a7846f104eb0fb339ec01a13ed85efda95050747db38da0320" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.001313 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"032e16e8-09b3-456f-bebb-09d4b07bcebf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eabca65587272a0e2ce09d13f103fbbe5374fb6d5ed49bfa0e26de8336d9e683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:30Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.023940 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fcde75bb1f9dd52a2ac77ffaa7d60d8c785c399bbe45a7cbff39d1b926d5541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:31Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.033420 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.033490 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.033508 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.033528 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.033543 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:31Z","lastTransitionTime":"2026-02-27T17:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.044796 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-g66q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d4f1eeba71eeb576c1df386f5acaf2bc354e7ca08aa9aac0971f61db72f244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b6cx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-g66q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:31Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.067107 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rbptz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66eb22fb-e593-40ff-9ff3-5bb32912972d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f58f4814976a1e25bef53555d7f5d14fa1cc800c8c5220801960758d5262b552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2dtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rbptz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:31Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.090567 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:31Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.122957 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c501ee5970cba15e45e1f1c8f0e33768019511f94dbbe509a7bbfea2c5b4a064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f4cd0745c61d33ac717a4a655f00713e2bd935c3cbba23c4ca0913859f6016f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:31Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.136654 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.136734 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.136751 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.136786 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.136808 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:31Z","lastTransitionTime":"2026-02-27T17:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.140356 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:31Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.153359 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dnkln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e772eecd-127f-4b26-8021-010fb507ce0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6df6de510d7e55695db4c89cbe9db84f91aca0e4bfbf2ed852407047c32822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v92s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dnkln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:31Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.172324 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47cf6a8e-df81-4ccb-992b-abb030862331\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857e486c0ff289c59f3de9933d1e0301be2f25b65e2b3f1fae6c4f149f900a4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8n2wb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:31Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.196408 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a336c2f16e12b411f111f6291366aa1b466f6b405896c152d1667b832a589b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:31Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.226638 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a34c0e3-2513-4e81-b6b1-80d1230475fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95d94117a4d368a7846f104eb0fb339ec01a13ed85efda95050747db38da0320\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95d94117a4d368a7846f104eb0fb339ec01a13ed85efda95050747db38da0320\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T17:02:17Z\\\",\\\"message\\\":\\\" syncing service certified-operators on namespace openshift-marketplace for network=default : 1.427497ms\\\\nI0227 17:02:17.571627 6694 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-apiserver/apiserver\\\\\\\"}\\\\nI0227 17:02:17.572157 6694 services_controller.go:360] Finished syncing service apiserver on namespace openshift-kube-apiserver for network=default : 2.941716ms\\\\nI0227 17:02:17.571720 6694 services_controller.go:360] Finished syncing service network-check-source on namespace openshift-network-diagnostics for network=default : 47.942µs\\\\nI0227 17:02:17.572342 6694 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console-operator/metrics\\\\\\\"}\\\\nI0227 17:02:17.572362 6694 services_controller.go:360] Finished syncing service metrics on namespace openshift-console-operator for network=default : 1.241712ms\\\\nI0227 17:02:17.572505 6694 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0227 17:02:17.572650 6694 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0227 17:02:17.572694 6694 ovnkube.go:599] Stopped ovnkube\\\\nI0227 17:02:17.572767 6694 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0227 17:02:17.572875 6694 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jtbqn_openshift-ovn-kubernetes(3a34c0e3-2513-4e81-b6b1-80d1230475fd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jtbqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:31Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.239688 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.239740 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.239760 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.239785 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.239801 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:31Z","lastTransitionTime":"2026-02-27T17:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.250410 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ee6013b-8a2d-4651-ae5f-98aaac016d22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af0992b3383953e562bf540bf7659d41e60aeee5d6103105eaab164797486dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a14230f224fb3b7b9b0ced93634313b66ac3abc376dca228bc5802890b0c393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f09ed7c87894bc0473c7015f0e0d55a28f893f5a2e8bf9085c86654c464371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa2391deb934395fe4743249887832731bf7c6ea457d66484bcb6b556e4ab33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64972d043accab95774784cc95dc1144e70c855eff777664c66cb121d519d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:31Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.265780 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a2d1b57-7fc1-4614-aac5-7144cdd812d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"27 17:01:51.956663 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 17:01:51.956671 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 17:01:51.956678 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 17:01:51.956715 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0227 17:01:51.960396 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1772211711\\\\\\\\\\\\\\\" (2026-02-27 17:01:50 +0000 UTC to 2026-03-29 17:01:51 +0000 UTC (now=2026-02-27 17:01:51.960134581 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960704 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1772211711\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1772211711\\\\\\\\\\\\\\\" (2026-02-27 16:01:51 +0000 UTC to 2027-02-27 16:01:51 +0000 UTC (now=2026-02-27 17:01:51.960667784 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960745 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0227 17:01:51.960790 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0227 17:01:51.960837 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960878 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960925 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\"\\\\nI0227 17:01:51.961071 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0227 17:01:51.958239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:01:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:31Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.283490 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:31Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.298911 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b45511-b94c-479f-98db-fd2c4eceec46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bddf1a7b7264d36174b90f1d0d4e49f029781b1acfce88582aa5b2575dcdd09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ef7c37b61de7a039756248f222bd1769b774703c406de5ad11abd9fb2f46dd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f9dq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:31Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.311454 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-s8k74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5288317-4810-4321-baa0-31a118a4dc36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-s8k74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:31Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.323516 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f425b79-78da-4f9e-a1ad-205a09466052\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3baf8927284b91f3edeec1dc906c2928955badb19062087790d720936b6f7624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd29e6d703945e0f366f2d109b7c35fdd6360c77091b165616e9832dddac1c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cxnv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:31Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.342146 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.342182 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.342194 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.342211 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.342223 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:31Z","lastTransitionTime":"2026-02-27T17:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.445099 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.445166 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.445193 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.445220 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.445235 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:31Z","lastTransitionTime":"2026-02-27T17:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.548355 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.548403 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.548414 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.548437 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.548452 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:31Z","lastTransitionTime":"2026-02-27T17:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.604756 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jtbqn_3a34c0e3-2513-4e81-b6b1-80d1230475fd/ovnkube-controller/1.log" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.608309 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" event={"ID":"3a34c0e3-2513-4e81-b6b1-80d1230475fd","Type":"ContainerStarted","Data":"c2eb3fcd96bb1fdabe12481cd07fcde6f19f153b6f52d0b48b3788ef4a0cd491"} Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.609876 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.629226 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"032e16e8-09b3-456f-bebb-09d4b07bcebf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eabca65587272a0e2ce09d13f103fbbe5374fb6d5ed49bfa0e26de8336d9e683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:31Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.648550 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fcde75bb1f9dd52a2ac77ffaa7d60d8c785c399bbe45a7cbff39d1b926d5541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:31Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.650736 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.650817 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.650840 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.650877 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.650898 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:31Z","lastTransitionTime":"2026-02-27T17:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.664171 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-g66q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d4f1eeba71eeb576c1df386f5acaf2bc354e7ca08aa9aac0971f61db72f244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b6cx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-g66q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:31Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.685920 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rbptz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66eb22fb-e593-40ff-9ff3-5bb32912972d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f58f4814976a1e25bef53555d7f5d14fa1cc800c8c5220801960758d5262b552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2dtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rbptz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:31Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.702859 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:31Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.723676 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c501ee5970cba15e45e1f1c8f0e33768019511f94dbbe509a7bbfea2c5b4a064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f4cd0745c61d33ac717a4a655f00713e2bd935c3cbba23c4ca0913859f6016f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:31Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.739570 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:31Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.753786 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.753864 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.753887 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.753913 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.753931 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:31Z","lastTransitionTime":"2026-02-27T17:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.754102 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dnkln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e772eecd-127f-4b26-8021-010fb507ce0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6df6de510d7e55695db4c89cbe9db84f91aca0e4bfbf2ed852407047c32822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v92s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dnkln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:31Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.771701 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47cf6a8e-df81-4ccb-992b-abb030862331\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857e486c0ff289c59f3de9933d1e0301be2f25b65e2b3f1fae6c4f149f900a4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8n2wb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:31Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.786448 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a336c2f16e12b411f111f6291366aa1b466f6b405896c152d1667b832a589b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:31Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.808137 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a34c0e3-2513-4e81-b6b1-80d1230475fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb3fcd96bb1fdabe12481cd07fcde6f19f153b6f52d0b48b3788ef4a0cd491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95d94117a4d368a7846f104eb0fb339ec01a13ed85efda95050747db38da0320\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T17:02:17Z\\\",\\\"message\\\":\\\" syncing service certified-operators on namespace openshift-marketplace for network=default : 1.427497ms\\\\nI0227 17:02:17.571627 6694 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-apiserver/apiserver\\\\\\\"}\\\\nI0227 17:02:17.572157 6694 services_controller.go:360] Finished syncing service apiserver on namespace openshift-kube-apiserver for network=default : 2.941716ms\\\\nI0227 17:02:17.571720 6694 services_controller.go:360] Finished syncing service network-check-source on namespace openshift-network-diagnostics for network=default : 47.942µs\\\\nI0227 17:02:17.572342 6694 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console-operator/metrics\\\\\\\"}\\\\nI0227 17:02:17.572362 6694 services_controller.go:360] Finished syncing service metrics on namespace openshift-console-operator for network=default : 1.241712ms\\\\nI0227 17:02:17.572505 6694 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0227 17:02:17.572650 6694 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0227 17:02:17.572694 6694 ovnkube.go:599] Stopped ovnkube\\\\nI0227 17:02:17.572767 6694 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0227 17:02:17.572875 6694 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jtbqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:31Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.841288 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ee6013b-8a2d-4651-ae5f-98aaac016d22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af0992b3383953e562bf540bf7659d41e60aeee5d6103105eaab164797486dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a14230f224fb3b7b9b0ced93634313b66ac3abc376dca228bc5802890b0c393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f09ed7c87894bc0473c7015f0e0d55a28f893f5a2e8bf9085c86654c464371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa2391deb934395fe4743249887832731bf7c6ea457d66484bcb6b556e4ab33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64972d043accab95774784cc95dc1144e70c855eff777664c66cb121d519d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:31Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.857443 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.857520 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.857535 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.857558 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.857573 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:31Z","lastTransitionTime":"2026-02-27T17:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.864406 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a2d1b57-7fc1-4614-aac5-7144cdd812d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"27 17:01:51.956663 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 17:01:51.956671 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 17:01:51.956678 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 17:01:51.956715 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0227 17:01:51.960396 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1772211711\\\\\\\\\\\\\\\" (2026-02-27 17:01:50 +0000 UTC to 2026-03-29 17:01:51 +0000 UTC (now=2026-02-27 17:01:51.960134581 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960704 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1772211711\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1772211711\\\\\\\\\\\\\\\" (2026-02-27 16:01:51 +0000 UTC to 2027-02-27 16:01:51 +0000 UTC (now=2026-02-27 17:01:51.960667784 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960745 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0227 17:01:51.960790 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0227 17:01:51.960837 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960878 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960925 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\"\\\\nI0227 17:01:51.961071 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0227 17:01:51.958239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:01:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:31Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.881354 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:31Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.893819 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b45511-b94c-479f-98db-fd2c4eceec46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bddf1a7b7264d36174b90f1d0d4e49f029781b1acfce88582aa5b2575dcdd09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ef7c37b61de7a039756248f222bd1769b774703c406de5ad11abd9fb2f46dd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f9dq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:31Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.907343 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-s8k74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5288317-4810-4321-baa0-31a118a4dc36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-s8k74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:31Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.919577 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f425b79-78da-4f9e-a1ad-205a09466052\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3baf8927284b91f3edeec1dc906c2928955badb19062087790d720936b6f7624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd29e6d703945e0f366f2d109b7c35fdd6360c77091b165616e9832dddac1c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cxnv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:31Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.960395 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.960447 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.960507 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.960535 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.960553 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:31Z","lastTransitionTime":"2026-02-27T17:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.980223 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.980294 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.980301 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:02:31 crc kubenswrapper[4700]: I0227 17:02:31.980250 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:02:31 crc kubenswrapper[4700]: E0227 17:02:31.980395 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:02:31 crc kubenswrapper[4700]: E0227 17:02:31.980678 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:02:31 crc kubenswrapper[4700]: E0227 17:02:31.980805 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:02:31 crc kubenswrapper[4700]: E0227 17:02:31.980915 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.063423 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.063485 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.063495 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.063512 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.063520 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:32Z","lastTransitionTime":"2026-02-27T17:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.166793 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.166917 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.166945 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.167019 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.167048 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:32Z","lastTransitionTime":"2026-02-27T17:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.190511 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.190584 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.190607 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.190637 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.190660 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:32Z","lastTransitionTime":"2026-02-27T17:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:32 crc kubenswrapper[4700]: E0227 17:02:32.217395 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:32Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.222725 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.222784 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.222802 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.222825 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.222843 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:32Z","lastTransitionTime":"2026-02-27T17:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:32 crc kubenswrapper[4700]: E0227 17:02:32.244660 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:32Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.250189 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.250256 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.250276 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.250304 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.250322 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:32Z","lastTransitionTime":"2026-02-27T17:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:32 crc kubenswrapper[4700]: E0227 17:02:32.270912 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:32Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.275702 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.275743 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.275755 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.275773 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.275785 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:32Z","lastTransitionTime":"2026-02-27T17:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:32 crc kubenswrapper[4700]: E0227 17:02:32.294755 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:32Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.299292 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.299564 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.299586 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.299636 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.299656 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:32Z","lastTransitionTime":"2026-02-27T17:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:32 crc kubenswrapper[4700]: E0227 17:02:32.321165 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:32Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:32 crc kubenswrapper[4700]: E0227 17:02:32.321389 4700 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.323260 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.323310 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.323326 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.323353 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.323370 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:32Z","lastTransitionTime":"2026-02-27T17:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.426902 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.426960 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.426977 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.427001 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.427019 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:32Z","lastTransitionTime":"2026-02-27T17:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.530323 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.530383 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.530400 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.530427 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.530446 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:32Z","lastTransitionTime":"2026-02-27T17:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.615057 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jtbqn_3a34c0e3-2513-4e81-b6b1-80d1230475fd/ovnkube-controller/2.log" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.616041 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jtbqn_3a34c0e3-2513-4e81-b6b1-80d1230475fd/ovnkube-controller/1.log" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.620031 4700 generic.go:334] "Generic (PLEG): container finished" podID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerID="c2eb3fcd96bb1fdabe12481cd07fcde6f19f153b6f52d0b48b3788ef4a0cd491" exitCode=1 Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.620107 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" event={"ID":"3a34c0e3-2513-4e81-b6b1-80d1230475fd","Type":"ContainerDied","Data":"c2eb3fcd96bb1fdabe12481cd07fcde6f19f153b6f52d0b48b3788ef4a0cd491"} Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.620175 4700 scope.go:117] "RemoveContainer" containerID="95d94117a4d368a7846f104eb0fb339ec01a13ed85efda95050747db38da0320" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.621221 4700 scope.go:117] "RemoveContainer" containerID="c2eb3fcd96bb1fdabe12481cd07fcde6f19f153b6f52d0b48b3788ef4a0cd491" Feb 27 17:02:32 crc kubenswrapper[4700]: E0227 17:02:32.621516 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jtbqn_openshift-ovn-kubernetes(3a34c0e3-2513-4e81-b6b1-80d1230475fd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.637155 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.637206 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.637224 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.637250 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.637269 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:32Z","lastTransitionTime":"2026-02-27T17:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.641984 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"032e16e8-09b3-456f-bebb-09d4b07bcebf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eabca65587272a0e2ce09d13f103fbbe5374fb6d5ed49bfa0e26de8336d9e683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:32Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.661926 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fcde75bb1f9dd52a2ac77ffaa7d60d8c785c399bbe45a7cbff39d1b926d5541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:32Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.678596 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-g66q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d4f1eeba71eeb576c1df386f5acaf2bc354e7ca08aa9aac0971f61db72f244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b6cx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-g66q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:32Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.698533 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rbptz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66eb22fb-e593-40ff-9ff3-5bb32912972d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f58f4814976a1e25bef53555d7f5d14fa1cc800c8c5220801960758d5262b552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2dtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rbptz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:32Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.721274 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47cf6a8e-df81-4ccb-992b-abb030862331\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857e486c0ff289c59f3de9933d1e0301be2f25b65e2b3f1fae6c4f149f900a4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8n2wb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:32Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.740199 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:32Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.740584 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.740633 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.740651 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.740674 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.740691 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:32Z","lastTransitionTime":"2026-02-27T17:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.760833 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c501ee5970cba15e45e1f1c8f0e33768019511f94dbbe509a7bbfea2c5b4a064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f4cd0745c61d33ac717a4a655f00713e2bd935c3cbba23c4ca0913859f6016f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:32Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.779590 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:32Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.798281 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dnkln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e772eecd-127f-4b26-8021-010fb507ce0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6df6de510d7e55695db4c89cbe9db84f91aca0e4bfbf2ed852407047c32822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v92s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dnkln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:32Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.819085 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a336c2f16e12b411f111f6291366aa1b466f6b405896c152d1667b832a589b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:32Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.847064 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.847134 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.847251 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.848318 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.848373 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:32Z","lastTransitionTime":"2026-02-27T17:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.854054 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a34c0e3-2513-4e81-b6b1-80d1230475fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb3fcd96bb1fdabe12481cd07fcde6f19f153b6f52d0b48b3788ef4a0cd491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95d94117a4d368a7846f104eb0fb339ec01a13ed85efda95050747db38da0320\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T17:02:17Z\\\",\\\"message\\\":\\\" syncing service certified-operators on namespace openshift-marketplace for network=default : 1.427497ms\\\\nI0227 17:02:17.571627 6694 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-apiserver/apiserver\\\\\\\"}\\\\nI0227 17:02:17.572157 6694 services_controller.go:360] Finished syncing service apiserver on namespace openshift-kube-apiserver for network=default : 2.941716ms\\\\nI0227 17:02:17.571720 6694 services_controller.go:360] Finished syncing service network-check-source on namespace openshift-network-diagnostics for network=default : 47.942µs\\\\nI0227 17:02:17.572342 6694 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console-operator/metrics\\\\\\\"}\\\\nI0227 17:02:17.572362 6694 services_controller.go:360] Finished syncing service metrics on namespace openshift-console-operator for network=default : 1.241712ms\\\\nI0227 17:02:17.572505 6694 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0227 17:02:17.572650 6694 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0227 17:02:17.572694 6694 ovnkube.go:599] Stopped ovnkube\\\\nI0227 17:02:17.572767 6694 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0227 17:02:17.572875 6694 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2eb3fcd96bb1fdabe12481cd07fcde6f19f153b6f52d0b48b3788ef4a0cd491\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"message\\\":\\\"or.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992312 6876 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992349 6876 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992495 6876 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 17:02:31.992823 6876 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992944 6876 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 17:02:31.993180 6876 ovnkube.go:599] Stopped ovnkube\\\\nI0227 17:02:31.993236 6876 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0227 17:02:31.993318 6876 ovnkube.go:137] failed to run ovnkube: [failed to start network contr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jtbqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:32Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.874065 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f425b79-78da-4f9e-a1ad-205a09466052\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3baf8927284b91f3edeec1dc906c2928955badb19062087790d720936b6f7624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd29e6d703945e0f366f2d109b7c35fdd6360c77091b165616e9832dddac1c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cxnv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:32Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.908654 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ee6013b-8a2d-4651-ae5f-98aaac016d22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af0992b3383953e562bf540bf7659d41e60aeee5d6103105eaab164797486dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a14230f224fb3b7b9b0ced93634313b66ac3abc376dca228bc5802890b0c393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f09ed7c87894bc0473c7015f0e0d55a28f893f5a2e8bf9085c86654c464371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa2391deb934395fe4743249887832731bf7c6ea457d66484bcb6b556e4ab33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64972d043accab95774784cc95dc1144e70c855eff777664c66cb121d519d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:32Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.930027 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a2d1b57-7fc1-4614-aac5-7144cdd812d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"27 17:01:51.956663 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 17:01:51.956671 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 17:01:51.956678 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 17:01:51.956715 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0227 17:01:51.960396 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1772211711\\\\\\\\\\\\\\\" (2026-02-27 17:01:50 +0000 UTC to 2026-03-29 17:01:51 +0000 UTC (now=2026-02-27 17:01:51.960134581 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960704 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1772211711\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1772211711\\\\\\\\\\\\\\\" (2026-02-27 16:01:51 +0000 UTC to 2027-02-27 16:01:51 +0000 UTC (now=2026-02-27 17:01:51.960667784 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960745 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0227 17:01:51.960790 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0227 17:01:51.960837 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960878 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960925 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\"\\\\nI0227 17:01:51.961071 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0227 17:01:51.958239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:01:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:32Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.949987 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:32Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.951069 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.951145 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.951164 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.951189 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.951211 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:32Z","lastTransitionTime":"2026-02-27T17:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.967271 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b45511-b94c-479f-98db-fd2c4eceec46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bddf1a7b7264d36174b90f1d0d4e49f029781b1acfce88582aa5b2575dcdd09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ef7c37b61de7a039756248f222bd1769b774703c406de5ad11abd9fb2f46dd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f9dq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:32Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:32 crc kubenswrapper[4700]: I0227 17:02:32.984255 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-s8k74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5288317-4810-4321-baa0-31a118a4dc36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-s8k74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:32Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.054259 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.054314 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.054332 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.054353 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.054369 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:33Z","lastTransitionTime":"2026-02-27T17:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.157024 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.157109 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.157127 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.157169 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.157186 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:33Z","lastTransitionTime":"2026-02-27T17:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.259791 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.259826 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.259838 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.259855 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.259866 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:33Z","lastTransitionTime":"2026-02-27T17:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.362743 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.362801 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.362818 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.362841 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.362859 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:33Z","lastTransitionTime":"2026-02-27T17:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.465927 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.465984 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.466003 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.466028 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.466045 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:33Z","lastTransitionTime":"2026-02-27T17:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.569193 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.569332 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.569351 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.569375 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.569394 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:33Z","lastTransitionTime":"2026-02-27T17:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.627014 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jtbqn_3a34c0e3-2513-4e81-b6b1-80d1230475fd/ovnkube-controller/2.log" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.632656 4700 scope.go:117] "RemoveContainer" containerID="c2eb3fcd96bb1fdabe12481cd07fcde6f19f153b6f52d0b48b3788ef4a0cd491" Feb 27 17:02:33 crc kubenswrapper[4700]: E0227 17:02:33.632902 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jtbqn_openshift-ovn-kubernetes(3a34c0e3-2513-4e81-b6b1-80d1230475fd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.654918 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a336c2f16e12b411f111f6291366aa1b466f6b405896c152d1667b832a589b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:33Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.672751 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.672814 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.672831 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.672854 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.672871 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:33Z","lastTransitionTime":"2026-02-27T17:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.686279 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a34c0e3-2513-4e81-b6b1-80d1230475fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb3fcd96bb1fdabe12481cd07fcde6f19f153b6f52d0b48b3788ef4a0cd491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2eb3fcd96bb1fdabe12481cd07fcde6f19f153b6f52d0b48b3788ef4a0cd491\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"message\\\":\\\"or.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992312 6876 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992349 6876 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992495 6876 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 17:02:31.992823 6876 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992944 6876 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 17:02:31.993180 6876 ovnkube.go:599] Stopped ovnkube\\\\nI0227 17:02:31.993236 6876 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0227 17:02:31.993318 6876 ovnkube.go:137] failed to run ovnkube: [failed to start network contr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jtbqn_openshift-ovn-kubernetes(3a34c0e3-2513-4e81-b6b1-80d1230475fd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jtbqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:33Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.717554 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ee6013b-8a2d-4651-ae5f-98aaac016d22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af0992b3383953e562bf540bf7659d41e60aeee5d6103105eaab164797486dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a14230f224fb3b7b9b0ced93634313b66ac3abc376dca228bc5802890b0c393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f09ed7c87894bc0473c7015f0e0d55a28f893f5a2e8bf9085c86654c464371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa2391deb934395fe4743249887832731bf7c6ea457d66484bcb6b556e4ab33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64972d043accab95774784cc95dc1144e70c855eff777664c66cb121d519d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:33Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.739660 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a2d1b57-7fc1-4614-aac5-7144cdd812d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"27 17:01:51.956663 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 17:01:51.956671 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 17:01:51.956678 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 17:01:51.956715 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0227 17:01:51.960396 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1772211711\\\\\\\\\\\\\\\" (2026-02-27 17:01:50 +0000 UTC to 2026-03-29 17:01:51 +0000 UTC (now=2026-02-27 17:01:51.960134581 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960704 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1772211711\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1772211711\\\\\\\\\\\\\\\" (2026-02-27 16:01:51 +0000 UTC to 2027-02-27 16:01:51 +0000 UTC (now=2026-02-27 17:01:51.960667784 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960745 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0227 17:01:51.960790 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0227 17:01:51.960837 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960878 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960925 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\"\\\\nI0227 17:01:51.961071 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0227 17:01:51.958239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:01:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:33Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.756842 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:33Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.775249 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b45511-b94c-479f-98db-fd2c4eceec46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bddf1a7b7264d36174b90f1d0d4e49f029781b1acfce88582aa5b2575dcdd09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ef7c37b61de7a039756248f222bd1769b774703c406de5ad11abd9fb2f46dd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f9dq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:33Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.776045 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.776178 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.776202 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.776229 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.776248 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:33Z","lastTransitionTime":"2026-02-27T17:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.794074 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-s8k74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5288317-4810-4321-baa0-31a118a4dc36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-s8k74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:33Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.812336 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f425b79-78da-4f9e-a1ad-205a09466052\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3baf8927284b91f3edeec1dc906c2928955badb19062087790d720936b6f7624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd29e6d703945e0f366f2d109b7c35fdd6360c77091b165616e9832dddac1c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cxnv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:33Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.828883 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"032e16e8-09b3-456f-bebb-09d4b07bcebf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eabca65587272a0e2ce09d13f103fbbe5374fb6d5ed49bfa0e26de8336d9e683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:33Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.848224 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fcde75bb1f9dd52a2ac77ffaa7d60d8c785c399bbe45a7cbff39d1b926d5541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:33Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.865734 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-g66q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d4f1eeba71eeb576c1df386f5acaf2bc354e7ca08aa9aac0971f61db72f244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b6cx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-g66q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:33Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.880550 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.880639 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.880659 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.880720 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.880745 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:33Z","lastTransitionTime":"2026-02-27T17:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.887375 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rbptz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66eb22fb-e593-40ff-9ff3-5bb32912972d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f58f4814976a1e25bef53555d7f5d14fa1cc800c8c5220801960758d5262b552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2dtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rbptz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:33Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.908346 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:33Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.929790 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c501ee5970cba15e45e1f1c8f0e33768019511f94dbbe509a7bbfea2c5b4a064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f4cd0745c61d33ac717a4a655f00713e2bd935c3cbba23c4ca0913859f6016f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:33Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.946808 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:33Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.961971 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dnkln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e772eecd-127f-4b26-8021-010fb507ce0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6df6de510d7e55695db4c89cbe9db84f91aca0e4bfbf2ed852407047c32822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v92s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dnkln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:33Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.981072 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.981119 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.981092 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.981195 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:33 crc kubenswrapper[4700]: E0227 17:02:33.981319 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:02:33 crc kubenswrapper[4700]: E0227 17:02:33.981600 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:02:33 crc kubenswrapper[4700]: E0227 17:02:33.981815 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:02:33 crc kubenswrapper[4700]: E0227 17:02:33.981902 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.983511 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.983605 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.983629 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.983651 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.983710 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:33Z","lastTransitionTime":"2026-02-27T17:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:33 crc kubenswrapper[4700]: I0227 17:02:33.990794 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47cf6a8e-df81-4ccb-992b-abb030862331\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857e486c0ff289c59f3de9933d1e0301be2f25b65e2b3f1fae6c4f149f900a4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8n2wb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:33Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.086865 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.086936 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.086956 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.086979 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.086996 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:34Z","lastTransitionTime":"2026-02-27T17:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.190619 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.190685 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.190703 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.190727 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.190749 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:34Z","lastTransitionTime":"2026-02-27T17:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.294356 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.294428 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.294452 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.294517 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.294543 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:34Z","lastTransitionTime":"2026-02-27T17:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.397732 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.397792 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.397809 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.397834 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.397852 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:34Z","lastTransitionTime":"2026-02-27T17:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.500371 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.500420 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.500826 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.500851 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.500870 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:34Z","lastTransitionTime":"2026-02-27T17:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.604660 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.604728 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.604746 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.604774 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.604794 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:34Z","lastTransitionTime":"2026-02-27T17:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.708052 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.708117 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.708135 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.708160 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.708178 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:34Z","lastTransitionTime":"2026-02-27T17:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.812366 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.812428 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.812448 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.812495 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.812513 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:34Z","lastTransitionTime":"2026-02-27T17:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.915293 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.915361 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.915377 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.915404 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:34 crc kubenswrapper[4700]: I0227 17:02:34.915421 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:34Z","lastTransitionTime":"2026-02-27T17:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.018898 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.019553 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.019584 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.019616 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.019633 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:35Z","lastTransitionTime":"2026-02-27T17:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.122981 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.123061 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.123101 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.123130 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.123148 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:35Z","lastTransitionTime":"2026-02-27T17:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.226069 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.226141 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.226159 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.226190 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.226215 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:35Z","lastTransitionTime":"2026-02-27T17:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.329919 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.329993 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.330015 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.330043 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.330066 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:35Z","lastTransitionTime":"2026-02-27T17:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.432879 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.432970 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.432987 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.433015 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.433033 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:35Z","lastTransitionTime":"2026-02-27T17:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.535849 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.535923 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.535941 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.535970 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.535992 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:35Z","lastTransitionTime":"2026-02-27T17:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.638990 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.639065 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.639083 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.639107 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.639128 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:35Z","lastTransitionTime":"2026-02-27T17:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.742182 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.742247 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.742266 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.742295 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.742314 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:35Z","lastTransitionTime":"2026-02-27T17:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.847379 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.847545 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.847604 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.847641 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.847664 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:35Z","lastTransitionTime":"2026-02-27T17:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.950862 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.950930 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.950950 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.950984 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.951004 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:35Z","lastTransitionTime":"2026-02-27T17:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.980508 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.980545 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:02:35 crc kubenswrapper[4700]: E0227 17:02:35.980676 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.980756 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:02:35 crc kubenswrapper[4700]: I0227 17:02:35.980780 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:35 crc kubenswrapper[4700]: E0227 17:02:35.980912 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:02:35 crc kubenswrapper[4700]: E0227 17:02:35.981047 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:02:35 crc kubenswrapper[4700]: E0227 17:02:35.981162 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.054084 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.054124 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.054153 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.054171 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.054185 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:36Z","lastTransitionTime":"2026-02-27T17:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.157248 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.157305 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.157321 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.157344 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.157360 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:36Z","lastTransitionTime":"2026-02-27T17:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.260727 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.260825 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.260845 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.260871 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.260892 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:36Z","lastTransitionTime":"2026-02-27T17:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.365188 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.365256 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.365275 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.365300 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.365318 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:36Z","lastTransitionTime":"2026-02-27T17:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.468649 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.468703 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.468717 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.469093 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.469132 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:36Z","lastTransitionTime":"2026-02-27T17:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.572662 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.572698 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.572712 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.572726 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.572736 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:36Z","lastTransitionTime":"2026-02-27T17:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.676022 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.676089 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.676107 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.676134 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.676157 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:36Z","lastTransitionTime":"2026-02-27T17:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.779543 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.779608 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.779621 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.779640 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.779654 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:36Z","lastTransitionTime":"2026-02-27T17:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.882305 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.882365 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.882386 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.882412 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.882431 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:36Z","lastTransitionTime":"2026-02-27T17:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.985073 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.985195 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.985229 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.985271 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:36 crc kubenswrapper[4700]: I0227 17:02:36.985298 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:36Z","lastTransitionTime":"2026-02-27T17:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.088326 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.088379 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.088397 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.088423 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.088440 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:37Z","lastTransitionTime":"2026-02-27T17:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.191422 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.191522 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.191540 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.191568 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.191593 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:37Z","lastTransitionTime":"2026-02-27T17:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.294990 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.295072 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.295099 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.295128 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.295149 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:37Z","lastTransitionTime":"2026-02-27T17:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.398000 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.398069 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.398086 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.398113 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.398131 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:37Z","lastTransitionTime":"2026-02-27T17:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.500367 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.500419 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.500435 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.500491 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.500509 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:37Z","lastTransitionTime":"2026-02-27T17:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.604604 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.604664 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.604682 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.604706 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.604723 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:37Z","lastTransitionTime":"2026-02-27T17:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.708127 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.708183 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.708200 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.708222 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.708238 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:37Z","lastTransitionTime":"2026-02-27T17:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.811500 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.811588 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.811614 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.811647 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.811666 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:37Z","lastTransitionTime":"2026-02-27T17:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.832893 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:02:37 crc kubenswrapper[4700]: E0227 17:02:37.833115 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:03:09.833070984 +0000 UTC m=+149.818383771 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.833560 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:37 crc kubenswrapper[4700]: E0227 17:02:37.833750 4700 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 17:02:37 crc kubenswrapper[4700]: E0227 17:02:37.833835 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 17:03:09.833814023 +0000 UTC m=+149.819126800 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 17:02:37 crc kubenswrapper[4700]: E0227 17:02:37.834010 4700 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 17:02:37 crc kubenswrapper[4700]: E0227 17:02:37.834118 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 17:03:09.83409328 +0000 UTC m=+149.819406057 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.834297 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.834556 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.834782 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:02:37 crc kubenswrapper[4700]: E0227 17:02:37.834812 4700 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 17:02:37 crc kubenswrapper[4700]: E0227 17:02:37.835194 4700 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 17:02:37 crc kubenswrapper[4700]: E0227 17:02:37.835342 4700 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 17:02:37 crc kubenswrapper[4700]: E0227 17:02:37.834893 4700 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 17:02:37 crc kubenswrapper[4700]: E0227 17:02:37.835680 4700 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 17:02:37 crc kubenswrapper[4700]: E0227 17:02:37.835709 4700 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 17:02:37 crc kubenswrapper[4700]: E0227 17:02:37.835589 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-27 17:03:09.835557128 +0000 UTC m=+149.820869905 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 17:02:37 crc kubenswrapper[4700]: E0227 17:02:37.835824 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-27 17:03:09.835796784 +0000 UTC m=+149.821109571 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.915245 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.915293 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.915311 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.915342 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.915365 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:37Z","lastTransitionTime":"2026-02-27T17:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.936550 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f5288317-4810-4321-baa0-31a118a4dc36-metrics-certs\") pod \"network-metrics-daemon-s8k74\" (UID: \"f5288317-4810-4321-baa0-31a118a4dc36\") " pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:37 crc kubenswrapper[4700]: E0227 17:02:37.936838 4700 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 17:02:37 crc kubenswrapper[4700]: E0227 17:02:37.936980 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f5288317-4810-4321-baa0-31a118a4dc36-metrics-certs podName:f5288317-4810-4321-baa0-31a118a4dc36 nodeName:}" failed. No retries permitted until 2026-02-27 17:03:09.936948264 +0000 UTC m=+149.922261051 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f5288317-4810-4321-baa0-31a118a4dc36-metrics-certs") pod "network-metrics-daemon-s8k74" (UID: "f5288317-4810-4321-baa0-31a118a4dc36") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.981101 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.981168 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.981174 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:37 crc kubenswrapper[4700]: E0227 17:02:37.981296 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:02:37 crc kubenswrapper[4700]: E0227 17:02:37.981354 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:02:37 crc kubenswrapper[4700]: E0227 17:02:37.981414 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:02:37 crc kubenswrapper[4700]: I0227 17:02:37.981703 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:02:37 crc kubenswrapper[4700]: E0227 17:02:37.981788 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.018559 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.018605 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.018616 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.018632 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.018646 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:38Z","lastTransitionTime":"2026-02-27T17:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.122351 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.122410 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.122421 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.122440 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.122453 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:38Z","lastTransitionTime":"2026-02-27T17:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.226500 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.226565 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.226583 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.226609 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.226627 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:38Z","lastTransitionTime":"2026-02-27T17:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.330424 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.330523 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.330542 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.330570 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.330592 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:38Z","lastTransitionTime":"2026-02-27T17:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.434088 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.434150 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.434167 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.434193 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.434218 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:38Z","lastTransitionTime":"2026-02-27T17:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.544072 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.544141 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.544158 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.544186 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.544204 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:38Z","lastTransitionTime":"2026-02-27T17:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.647705 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.647766 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.647785 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.647810 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.647830 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:38Z","lastTransitionTime":"2026-02-27T17:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.750782 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.750842 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.750854 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.750876 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.750889 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:38Z","lastTransitionTime":"2026-02-27T17:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.853678 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.853749 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.853768 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.853796 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.853814 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:38Z","lastTransitionTime":"2026-02-27T17:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.957529 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.957593 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.957612 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.957637 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.957655 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:38Z","lastTransitionTime":"2026-02-27T17:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:38 crc kubenswrapper[4700]: I0227 17:02:38.981756 4700 scope.go:117] "RemoveContainer" containerID="26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.060891 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.060962 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.060986 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.061038 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.061067 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:39Z","lastTransitionTime":"2026-02-27T17:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.164798 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.164859 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.164880 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.164908 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.164926 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:39Z","lastTransitionTime":"2026-02-27T17:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.268133 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.268610 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.268627 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.268652 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.268672 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:39Z","lastTransitionTime":"2026-02-27T17:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.372611 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.372701 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.372722 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.372749 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.372769 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:39Z","lastTransitionTime":"2026-02-27T17:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.476097 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.476175 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.476194 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.476225 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.476249 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:39Z","lastTransitionTime":"2026-02-27T17:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.579680 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.579762 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.579780 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.579809 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.579829 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:39Z","lastTransitionTime":"2026-02-27T17:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.657132 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.660524 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4d2d2c9065b008ff0936c19bcce2ce6a574580911a8188084c5f978e08177b69"} Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.661327 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.682237 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.682303 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.682320 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.682350 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.682371 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:39Z","lastTransitionTime":"2026-02-27T17:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.686493 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a336c2f16e12b411f111f6291366aa1b466f6b405896c152d1667b832a589b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:39Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.724305 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a34c0e3-2513-4e81-b6b1-80d1230475fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb3fcd96bb1fdabe12481cd07fcde6f19f153b6f52d0b48b3788ef4a0cd491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2eb3fcd96bb1fdabe12481cd07fcde6f19f153b6f52d0b48b3788ef4a0cd491\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"message\\\":\\\"or.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992312 6876 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992349 6876 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992495 6876 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 17:02:31.992823 6876 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992944 6876 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 17:02:31.993180 6876 ovnkube.go:599] Stopped ovnkube\\\\nI0227 17:02:31.993236 6876 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0227 17:02:31.993318 6876 ovnkube.go:137] failed to run ovnkube: [failed to start network contr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jtbqn_openshift-ovn-kubernetes(3a34c0e3-2513-4e81-b6b1-80d1230475fd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jtbqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:39Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.743774 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b45511-b94c-479f-98db-fd2c4eceec46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bddf1a7b7264d36174b90f1d0d4e49f029781b1acfce88582aa5b2575dcdd09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ef7c37b61de7a039756248f222bd1769b774703c406de5ad11abd9fb2f46dd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f9dq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:39Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.759282 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-s8k74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5288317-4810-4321-baa0-31a118a4dc36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-s8k74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:39Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.776982 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f425b79-78da-4f9e-a1ad-205a09466052\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3baf8927284b91f3edeec1dc906c2928955badb19062087790d720936b6f7624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd29e6d703945e0f366f2d109b7c35fdd6360c77091b165616e9832dddac1c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cxnv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:39Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.785531 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.785583 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.785603 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.785627 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.785645 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:39Z","lastTransitionTime":"2026-02-27T17:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.813142 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ee6013b-8a2d-4651-ae5f-98aaac016d22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af0992b3383953e562bf540bf7659d41e60aeee5d6103105eaab164797486dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a14230f224fb3b7b9b0ced93634313b66ac3abc376dca228bc5802890b0c393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f09ed7c87894bc0473c7015f0e0d55a28f893f5a2e8bf9085c86654c464371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa2391deb934395fe4743249887832731bf7c6ea457d66484bcb6b556e4ab33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64972d043accab95774784cc95dc1144e70c855eff777664c66cb121d519d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:39Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.838831 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a2d1b57-7fc1-4614-aac5-7144cdd812d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d2d2c9065b008ff0936c19bcce2ce6a574580911a8188084c5f978e08177b69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"27 17:01:51.956663 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 17:01:51.956671 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 17:01:51.956678 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 17:01:51.956715 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0227 17:01:51.960396 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1772211711\\\\\\\\\\\\\\\" (2026-02-27 17:01:50 +0000 UTC to 2026-03-29 17:01:51 +0000 UTC (now=2026-02-27 17:01:51.960134581 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960704 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1772211711\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1772211711\\\\\\\\\\\\\\\" (2026-02-27 16:01:51 +0000 UTC to 2027-02-27 16:01:51 +0000 UTC (now=2026-02-27 17:01:51.960667784 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960745 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0227 17:01:51.960790 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0227 17:01:51.960837 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960878 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960925 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\"\\\\nI0227 17:01:51.961071 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0227 17:01:51.958239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:01:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:39Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.862279 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:39Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.885399 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rbptz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66eb22fb-e593-40ff-9ff3-5bb32912972d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f58f4814976a1e25bef53555d7f5d14fa1cc800c8c5220801960758d5262b552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2dtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rbptz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:39Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.888324 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.888368 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.888381 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.888413 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.888426 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:39Z","lastTransitionTime":"2026-02-27T17:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.903408 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"032e16e8-09b3-456f-bebb-09d4b07bcebf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eabca65587272a0e2ce09d13f103fbbe5374fb6d5ed49bfa0e26de8336d9e683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:39Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.925632 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fcde75bb1f9dd52a2ac77ffaa7d60d8c785c399bbe45a7cbff39d1b926d5541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:39Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.945134 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-g66q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d4f1eeba71eeb576c1df386f5acaf2bc354e7ca08aa9aac0971f61db72f244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b6cx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-g66q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:39Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.967975 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:39Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.980112 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.980213 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:39 crc kubenswrapper[4700]: E0227 17:02:39.980348 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.980490 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.980528 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:39 crc kubenswrapper[4700]: E0227 17:02:39.980684 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:02:39 crc kubenswrapper[4700]: E0227 17:02:39.980949 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:02:39 crc kubenswrapper[4700]: E0227 17:02:39.980991 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.993046 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.993080 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.993091 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.993108 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.993122 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:39Z","lastTransitionTime":"2026-02-27T17:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:39 crc kubenswrapper[4700]: I0227 17:02:39.994430 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dnkln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e772eecd-127f-4b26-8021-010fb507ce0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6df6de510d7e55695db4c89cbe9db84f91aca0e4bfbf2ed852407047c32822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v92s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dnkln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:39Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.020026 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47cf6a8e-df81-4ccb-992b-abb030862331\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857e486c0ff289c59f3de9933d1e0301be2f25b65e2b3f1fae6c4f149f900a4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8n2wb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:40Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.039576 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:40Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.057193 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c501ee5970cba15e45e1f1c8f0e33768019511f94dbbe509a7bbfea2c5b4a064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f4cd0745c61d33ac717a4a655f00713e2bd935c3cbba23c4ca0913859f6016f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:40Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.094679 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.094747 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.094764 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.095255 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.096519 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:40Z","lastTransitionTime":"2026-02-27T17:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.200338 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.200391 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.200407 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.200429 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.200445 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:40Z","lastTransitionTime":"2026-02-27T17:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.303233 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.303293 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.303309 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.303332 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.303349 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:40Z","lastTransitionTime":"2026-02-27T17:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.406995 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.407047 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.407064 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.407086 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.407103 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:40Z","lastTransitionTime":"2026-02-27T17:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.510581 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.510719 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.510764 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.510808 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.510824 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:40Z","lastTransitionTime":"2026-02-27T17:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.613337 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.613375 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.613387 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.613402 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.613415 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:40Z","lastTransitionTime":"2026-02-27T17:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.716716 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.716762 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.716775 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.716793 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.716805 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:40Z","lastTransitionTime":"2026-02-27T17:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.819980 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.820686 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.820859 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.821028 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:40 crc kubenswrapper[4700]: I0227 17:02:40.821182 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:40Z","lastTransitionTime":"2026-02-27T17:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:40 crc kubenswrapper[4700]: E0227 17:02:40.921813 4700 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Feb 27 17:02:41 crc kubenswrapper[4700]: I0227 17:02:41.003617 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a336c2f16e12b411f111f6291366aa1b466f6b405896c152d1667b832a589b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:41Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:41 crc kubenswrapper[4700]: I0227 17:02:41.037610 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a34c0e3-2513-4e81-b6b1-80d1230475fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb3fcd96bb1fdabe12481cd07fcde6f19f153b6f52d0b48b3788ef4a0cd491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2eb3fcd96bb1fdabe12481cd07fcde6f19f153b6f52d0b48b3788ef4a0cd491\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"message\\\":\\\"or.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992312 6876 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992349 6876 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992495 6876 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 17:02:31.992823 6876 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992944 6876 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 17:02:31.993180 6876 ovnkube.go:599] Stopped ovnkube\\\\nI0227 17:02:31.993236 6876 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0227 17:02:31.993318 6876 ovnkube.go:137] failed to run ovnkube: [failed to start network contr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jtbqn_openshift-ovn-kubernetes(3a34c0e3-2513-4e81-b6b1-80d1230475fd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jtbqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:41Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:41 crc kubenswrapper[4700]: I0227 17:02:41.056562 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f425b79-78da-4f9e-a1ad-205a09466052\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3baf8927284b91f3edeec1dc906c2928955badb19062087790d720936b6f7624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd29e6d703945e0f366f2d109b7c35fdd6360c77091b165616e9832dddac1c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cxnv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:41Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:41 crc kubenswrapper[4700]: I0227 17:02:41.092136 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ee6013b-8a2d-4651-ae5f-98aaac016d22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af0992b3383953e562bf540bf7659d41e60aeee5d6103105eaab164797486dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a14230f224fb3b7b9b0ced93634313b66ac3abc376dca228bc5802890b0c393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f09ed7c87894bc0473c7015f0e0d55a28f893f5a2e8bf9085c86654c464371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa2391deb934395fe4743249887832731bf7c6ea457d66484bcb6b556e4ab33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64972d043accab95774784cc95dc1144e70c855eff777664c66cb121d519d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:41Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:41 crc kubenswrapper[4700]: E0227 17:02:41.101199 4700 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 17:02:41 crc kubenswrapper[4700]: I0227 17:02:41.117245 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a2d1b57-7fc1-4614-aac5-7144cdd812d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d2d2c9065b008ff0936c19bcce2ce6a574580911a8188084c5f978e08177b69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"27 17:01:51.956663 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 17:01:51.956671 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 17:01:51.956678 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 17:01:51.956715 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0227 17:01:51.960396 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1772211711\\\\\\\\\\\\\\\" (2026-02-27 17:01:50 +0000 UTC to 2026-03-29 17:01:51 +0000 UTC (now=2026-02-27 17:01:51.960134581 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960704 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1772211711\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1772211711\\\\\\\\\\\\\\\" (2026-02-27 16:01:51 +0000 UTC to 2027-02-27 16:01:51 +0000 UTC (now=2026-02-27 17:01:51.960667784 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960745 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0227 17:01:51.960790 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0227 17:01:51.960837 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960878 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960925 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\"\\\\nI0227 17:01:51.961071 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0227 17:01:51.958239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:01:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:41Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:41 crc kubenswrapper[4700]: I0227 17:02:41.137983 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:41Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:41 crc kubenswrapper[4700]: I0227 17:02:41.157640 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b45511-b94c-479f-98db-fd2c4eceec46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bddf1a7b7264d36174b90f1d0d4e49f029781b1acfce88582aa5b2575dcdd09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ef7c37b61de7a039756248f222bd1769b774703c406de5ad11abd9fb2f46dd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f9dq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:41Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:41 crc kubenswrapper[4700]: I0227 17:02:41.188542 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-s8k74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5288317-4810-4321-baa0-31a118a4dc36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-s8k74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:41Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:41 crc kubenswrapper[4700]: I0227 17:02:41.208388 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"032e16e8-09b3-456f-bebb-09d4b07bcebf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eabca65587272a0e2ce09d13f103fbbe5374fb6d5ed49bfa0e26de8336d9e683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:41Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:41 crc kubenswrapper[4700]: I0227 17:02:41.226899 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fcde75bb1f9dd52a2ac77ffaa7d60d8c785c399bbe45a7cbff39d1b926d5541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:41Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:41 crc kubenswrapper[4700]: I0227 17:02:41.242113 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-g66q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d4f1eeba71eeb576c1df386f5acaf2bc354e7ca08aa9aac0971f61db72f244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b6cx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-g66q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:41Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:41 crc kubenswrapper[4700]: I0227 17:02:41.261199 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rbptz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66eb22fb-e593-40ff-9ff3-5bb32912972d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f58f4814976a1e25bef53555d7f5d14fa1cc800c8c5220801960758d5262b552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2dtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rbptz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:41Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:41 crc kubenswrapper[4700]: I0227 17:02:41.276276 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47cf6a8e-df81-4ccb-992b-abb030862331\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857e486c0ff289c59f3de9933d1e0301be2f25b65e2b3f1fae6c4f149f900a4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8n2wb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:41Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:41 crc kubenswrapper[4700]: I0227 17:02:41.289969 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:41Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:41 crc kubenswrapper[4700]: I0227 17:02:41.369780 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c501ee5970cba15e45e1f1c8f0e33768019511f94dbbe509a7bbfea2c5b4a064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f4cd0745c61d33ac717a4a655f00713e2bd935c3cbba23c4ca0913859f6016f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:41Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:41 crc kubenswrapper[4700]: I0227 17:02:41.383962 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:41Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:41 crc kubenswrapper[4700]: I0227 17:02:41.396019 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dnkln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e772eecd-127f-4b26-8021-010fb507ce0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6df6de510d7e55695db4c89cbe9db84f91aca0e4bfbf2ed852407047c32822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v92s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dnkln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:41Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:41 crc kubenswrapper[4700]: I0227 17:02:41.981098 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:02:41 crc kubenswrapper[4700]: I0227 17:02:41.981172 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:02:41 crc kubenswrapper[4700]: I0227 17:02:41.981109 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:41 crc kubenswrapper[4700]: E0227 17:02:41.981231 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:02:41 crc kubenswrapper[4700]: I0227 17:02:41.981109 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:41 crc kubenswrapper[4700]: E0227 17:02:41.981307 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:02:41 crc kubenswrapper[4700]: E0227 17:02:41.981405 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:02:41 crc kubenswrapper[4700]: E0227 17:02:41.981496 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:02:42 crc kubenswrapper[4700]: I0227 17:02:42.362691 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:42 crc kubenswrapper[4700]: I0227 17:02:42.362734 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:42 crc kubenswrapper[4700]: I0227 17:02:42.362763 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:42 crc kubenswrapper[4700]: I0227 17:02:42.362781 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:42 crc kubenswrapper[4700]: I0227 17:02:42.362797 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:42Z","lastTransitionTime":"2026-02-27T17:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:42 crc kubenswrapper[4700]: E0227 17:02:42.383171 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:42Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:42 crc kubenswrapper[4700]: I0227 17:02:42.387483 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:42 crc kubenswrapper[4700]: I0227 17:02:42.387530 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:42 crc kubenswrapper[4700]: I0227 17:02:42.387545 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:42 crc kubenswrapper[4700]: I0227 17:02:42.387568 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:42 crc kubenswrapper[4700]: I0227 17:02:42.387583 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:42Z","lastTransitionTime":"2026-02-27T17:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:42 crc kubenswrapper[4700]: E0227 17:02:42.407259 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:42Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:42 crc kubenswrapper[4700]: I0227 17:02:42.411341 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:42 crc kubenswrapper[4700]: I0227 17:02:42.411386 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:42 crc kubenswrapper[4700]: I0227 17:02:42.411401 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:42 crc kubenswrapper[4700]: I0227 17:02:42.411420 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:42 crc kubenswrapper[4700]: I0227 17:02:42.411434 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:42Z","lastTransitionTime":"2026-02-27T17:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:42 crc kubenswrapper[4700]: E0227 17:02:42.429849 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:42Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:42 crc kubenswrapper[4700]: I0227 17:02:42.433950 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:42 crc kubenswrapper[4700]: I0227 17:02:42.434008 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:42 crc kubenswrapper[4700]: I0227 17:02:42.434025 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:42 crc kubenswrapper[4700]: I0227 17:02:42.434049 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:42 crc kubenswrapper[4700]: I0227 17:02:42.434067 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:42Z","lastTransitionTime":"2026-02-27T17:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:42 crc kubenswrapper[4700]: E0227 17:02:42.449690 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:42Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:42 crc kubenswrapper[4700]: I0227 17:02:42.453814 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:42 crc kubenswrapper[4700]: I0227 17:02:42.453858 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:42 crc kubenswrapper[4700]: I0227 17:02:42.453869 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:42 crc kubenswrapper[4700]: I0227 17:02:42.453886 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:42 crc kubenswrapper[4700]: I0227 17:02:42.453897 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:42Z","lastTransitionTime":"2026-02-27T17:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:42 crc kubenswrapper[4700]: E0227 17:02:42.470250 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:42Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:42 crc kubenswrapper[4700]: E0227 17:02:42.470627 4700 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 27 17:02:43 crc kubenswrapper[4700]: I0227 17:02:43.980291 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:02:43 crc kubenswrapper[4700]: I0227 17:02:43.980317 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:43 crc kubenswrapper[4700]: I0227 17:02:43.980359 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:02:43 crc kubenswrapper[4700]: I0227 17:02:43.980501 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:43 crc kubenswrapper[4700]: E0227 17:02:43.980609 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:02:43 crc kubenswrapper[4700]: E0227 17:02:43.980747 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:02:43 crc kubenswrapper[4700]: E0227 17:02:43.980881 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:02:43 crc kubenswrapper[4700]: E0227 17:02:43.980930 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:02:44 crc kubenswrapper[4700]: I0227 17:02:44.995098 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Feb 27 17:02:45 crc kubenswrapper[4700]: I0227 17:02:45.980586 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:02:45 crc kubenswrapper[4700]: I0227 17:02:45.980601 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:02:45 crc kubenswrapper[4700]: E0227 17:02:45.981313 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:02:45 crc kubenswrapper[4700]: I0227 17:02:45.980634 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:45 crc kubenswrapper[4700]: I0227 17:02:45.980607 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:45 crc kubenswrapper[4700]: E0227 17:02:45.981591 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:02:45 crc kubenswrapper[4700]: E0227 17:02:45.982264 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:02:45 crc kubenswrapper[4700]: E0227 17:02:45.982399 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:02:46 crc kubenswrapper[4700]: E0227 17:02:46.102666 4700 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 17:02:46 crc kubenswrapper[4700]: I0227 17:02:46.981540 4700 scope.go:117] "RemoveContainer" containerID="c2eb3fcd96bb1fdabe12481cd07fcde6f19f153b6f52d0b48b3788ef4a0cd491" Feb 27 17:02:46 crc kubenswrapper[4700]: E0227 17:02:46.981820 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jtbqn_openshift-ovn-kubernetes(3a34c0e3-2513-4e81-b6b1-80d1230475fd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" Feb 27 17:02:47 crc kubenswrapper[4700]: I0227 17:02:47.980801 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:02:47 crc kubenswrapper[4700]: I0227 17:02:47.980929 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:47 crc kubenswrapper[4700]: E0227 17:02:47.981592 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:02:47 crc kubenswrapper[4700]: I0227 17:02:47.981034 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:02:47 crc kubenswrapper[4700]: I0227 17:02:47.980937 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:47 crc kubenswrapper[4700]: E0227 17:02:47.981919 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:02:47 crc kubenswrapper[4700]: E0227 17:02:47.982081 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:02:47 crc kubenswrapper[4700]: E0227 17:02:47.981767 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:02:49 crc kubenswrapper[4700]: I0227 17:02:49.980086 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:02:49 crc kubenswrapper[4700]: I0227 17:02:49.980137 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:49 crc kubenswrapper[4700]: E0227 17:02:49.980247 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:02:49 crc kubenswrapper[4700]: E0227 17:02:49.980435 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:02:49 crc kubenswrapper[4700]: I0227 17:02:49.980102 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:02:49 crc kubenswrapper[4700]: E0227 17:02:49.980638 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:02:49 crc kubenswrapper[4700]: I0227 17:02:49.980856 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:49 crc kubenswrapper[4700]: E0227 17:02:49.980990 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.001316 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c501ee5970cba15e45e1f1c8f0e33768019511f94dbbe509a7bbfea2c5b4a064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f4cd0745c61d33ac717a4a655f00713e2bd935c3cbba23c4ca0913859f6016f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:50Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.020760 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:51Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.036088 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dnkln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e772eecd-127f-4b26-8021-010fb507ce0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6df6de510d7e55695db4c89cbe9db84f91aca0e4bfbf2ed852407047c32822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v92s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dnkln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:51Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.057995 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47cf6a8e-df81-4ccb-992b-abb030862331\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857e486c0ff289c59f3de9933d1e0301be2f25b65e2b3f1fae6c4f149f900a4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8n2wb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:51Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.079297 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"309d44d0-45a4-4483-b588-7937902ee1fb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9da83072905af757ff9f9cecda0d94b2acfcd847aeeba7ca36ac291c4ff6a3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5a0dc977696434dce7529115ee1d6cce69cbd4c501bbfb77ed0d240d6340d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf6d6f90f1a5acb192ed932732bab2c2a2bebe9f03fddc79c5dfe6423fab012c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b6649f7710d4c1435bfd703120b70e4514f95b4750b247b8ad0a69ed6952a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b6649f7710d4c1435bfd703120b70e4514f95b4750b247b8ad0a69ed6952a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:51Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.101777 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:51Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:51 crc kubenswrapper[4700]: E0227 17:02:51.104175 4700 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.137453 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a34c0e3-2513-4e81-b6b1-80d1230475fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb3fcd96bb1fdabe12481cd07fcde6f19f153b6f52d0b48b3788ef4a0cd491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2eb3fcd96bb1fdabe12481cd07fcde6f19f153b6f52d0b48b3788ef4a0cd491\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"message\\\":\\\"or.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992312 6876 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992349 6876 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992495 6876 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 17:02:31.992823 6876 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992944 6876 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 17:02:31.993180 6876 ovnkube.go:599] Stopped ovnkube\\\\nI0227 17:02:31.993236 6876 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0227 17:02:31.993318 6876 ovnkube.go:137] failed to run ovnkube: [failed to start network contr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jtbqn_openshift-ovn-kubernetes(3a34c0e3-2513-4e81-b6b1-80d1230475fd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jtbqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:51Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.164962 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a336c2f16e12b411f111f6291366aa1b466f6b405896c152d1667b832a589b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:51Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.186207 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:51Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.205491 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b45511-b94c-479f-98db-fd2c4eceec46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bddf1a7b7264d36174b90f1d0d4e49f029781b1acfce88582aa5b2575dcdd09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ef7c37b61de7a039756248f222bd1769b774703c406de5ad11abd9fb2f46dd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f9dq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:51Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.223373 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-s8k74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5288317-4810-4321-baa0-31a118a4dc36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-s8k74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:51Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.239132 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f425b79-78da-4f9e-a1ad-205a09466052\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3baf8927284b91f3edeec1dc906c2928955badb19062087790d720936b6f7624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd29e6d703945e0f366f2d109b7c35fdd6360c77091b165616e9832dddac1c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cxnv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:51Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.265251 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ee6013b-8a2d-4651-ae5f-98aaac016d22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af0992b3383953e562bf540bf7659d41e60aeee5d6103105eaab164797486dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a14230f224fb3b7b9b0ced93634313b66ac3abc376dca228bc5802890b0c393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f09ed7c87894bc0473c7015f0e0d55a28f893f5a2e8bf9085c86654c464371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa2391deb934395fe4743249887832731bf7c6ea457d66484bcb6b556e4ab33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64972d043accab95774784cc95dc1144e70c855eff777664c66cb121d519d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:51Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.283936 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a2d1b57-7fc1-4614-aac5-7144cdd812d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d2d2c9065b008ff0936c19bcce2ce6a574580911a8188084c5f978e08177b69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"27 17:01:51.956663 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 17:01:51.956671 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 17:01:51.956678 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 17:01:51.956715 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0227 17:01:51.960396 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1772211711\\\\\\\\\\\\\\\" (2026-02-27 17:01:50 +0000 UTC to 2026-03-29 17:01:51 +0000 UTC (now=2026-02-27 17:01:51.960134581 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960704 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1772211711\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1772211711\\\\\\\\\\\\\\\" (2026-02-27 16:01:51 +0000 UTC to 2027-02-27 16:01:51 +0000 UTC (now=2026-02-27 17:01:51.960667784 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960745 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0227 17:01:51.960790 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0227 17:01:51.960837 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960878 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960925 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\"\\\\nI0227 17:01:51.961071 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0227 17:01:51.958239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:01:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:51Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.299230 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-g66q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d4f1eeba71eeb576c1df386f5acaf2bc354e7ca08aa9aac0971f61db72f244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b6cx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-g66q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:51Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.314623 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rbptz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66eb22fb-e593-40ff-9ff3-5bb32912972d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f58f4814976a1e25bef53555d7f5d14fa1cc800c8c5220801960758d5262b552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2dtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rbptz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:51Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.329702 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"032e16e8-09b3-456f-bebb-09d4b07bcebf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eabca65587272a0e2ce09d13f103fbbe5374fb6d5ed49bfa0e26de8336d9e683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:51Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.344935 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fcde75bb1f9dd52a2ac77ffaa7d60d8c785c399bbe45a7cbff39d1b926d5541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:51Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.525526 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.539363 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"032e16e8-09b3-456f-bebb-09d4b07bcebf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eabca65587272a0e2ce09d13f103fbbe5374fb6d5ed49bfa0e26de8336d9e683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:51Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.553408 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fcde75bb1f9dd52a2ac77ffaa7d60d8c785c399bbe45a7cbff39d1b926d5541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:51Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.566188 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-g66q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d4f1eeba71eeb576c1df386f5acaf2bc354e7ca08aa9aac0971f61db72f244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b6cx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-g66q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:51Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.579870 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rbptz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66eb22fb-e593-40ff-9ff3-5bb32912972d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f58f4814976a1e25bef53555d7f5d14fa1cc800c8c5220801960758d5262b552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2dtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rbptz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:51Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.593308 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dnkln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e772eecd-127f-4b26-8021-010fb507ce0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6df6de510d7e55695db4c89cbe9db84f91aca0e4bfbf2ed852407047c32822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v92s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dnkln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:51Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.613283 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47cf6a8e-df81-4ccb-992b-abb030862331\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857e486c0ff289c59f3de9933d1e0301be2f25b65e2b3f1fae6c4f149f900a4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8n2wb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:51Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.631840 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"309d44d0-45a4-4483-b588-7937902ee1fb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9da83072905af757ff9f9cecda0d94b2acfcd847aeeba7ca36ac291c4ff6a3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5a0dc977696434dce7529115ee1d6cce69cbd4c501bbfb77ed0d240d6340d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf6d6f90f1a5acb192ed932732bab2c2a2bebe9f03fddc79c5dfe6423fab012c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b6649f7710d4c1435bfd703120b70e4514f95b4750b247b8ad0a69ed6952a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b6649f7710d4c1435bfd703120b70e4514f95b4750b247b8ad0a69ed6952a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:51Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.649852 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:51Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.680305 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c501ee5970cba15e45e1f1c8f0e33768019511f94dbbe509a7bbfea2c5b4a064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f4cd0745c61d33ac717a4a655f00713e2bd935c3cbba23c4ca0913859f6016f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:51Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.697005 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:51Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.717177 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a336c2f16e12b411f111f6291366aa1b466f6b405896c152d1667b832a589b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:51Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.739547 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a34c0e3-2513-4e81-b6b1-80d1230475fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb3fcd96bb1fdabe12481cd07fcde6f19f153b6f52d0b48b3788ef4a0cd491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2eb3fcd96bb1fdabe12481cd07fcde6f19f153b6f52d0b48b3788ef4a0cd491\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"message\\\":\\\"or.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992312 6876 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992349 6876 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992495 6876 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 17:02:31.992823 6876 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992944 6876 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 17:02:31.993180 6876 ovnkube.go:599] Stopped ovnkube\\\\nI0227 17:02:31.993236 6876 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0227 17:02:31.993318 6876 ovnkube.go:137] failed to run ovnkube: [failed to start network contr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jtbqn_openshift-ovn-kubernetes(3a34c0e3-2513-4e81-b6b1-80d1230475fd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jtbqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:51Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.754802 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-s8k74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5288317-4810-4321-baa0-31a118a4dc36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-s8k74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:51Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.771317 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f425b79-78da-4f9e-a1ad-205a09466052\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3baf8927284b91f3edeec1dc906c2928955badb19062087790d720936b6f7624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd29e6d703945e0f366f2d109b7c35fdd6360c77091b165616e9832dddac1c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cxnv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:51Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.797404 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ee6013b-8a2d-4651-ae5f-98aaac016d22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af0992b3383953e562bf540bf7659d41e60aeee5d6103105eaab164797486dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a14230f224fb3b7b9b0ced93634313b66ac3abc376dca228bc5802890b0c393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f09ed7c87894bc0473c7015f0e0d55a28f893f5a2e8bf9085c86654c464371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa2391deb934395fe4743249887832731bf7c6ea457d66484bcb6b556e4ab33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64972d043accab95774784cc95dc1144e70c855eff777664c66cb121d519d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:51Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.815050 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a2d1b57-7fc1-4614-aac5-7144cdd812d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d2d2c9065b008ff0936c19bcce2ce6a574580911a8188084c5f978e08177b69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"27 17:01:51.956663 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 17:01:51.956671 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 17:01:51.956678 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 17:01:51.956715 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0227 17:01:51.960396 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1772211711\\\\\\\\\\\\\\\" (2026-02-27 17:01:50 +0000 UTC to 2026-03-29 17:01:51 +0000 UTC (now=2026-02-27 17:01:51.960134581 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960704 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1772211711\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1772211711\\\\\\\\\\\\\\\" (2026-02-27 16:01:51 +0000 UTC to 2027-02-27 16:01:51 +0000 UTC (now=2026-02-27 17:01:51.960667784 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960745 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0227 17:01:51.960790 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0227 17:01:51.960837 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960878 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960925 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\"\\\\nI0227 17:01:51.961071 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0227 17:01:51.958239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:01:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:51Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.832445 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:51Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.846909 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b45511-b94c-479f-98db-fd2c4eceec46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bddf1a7b7264d36174b90f1d0d4e49f029781b1acfce88582aa5b2575dcdd09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ef7c37b61de7a039756248f222bd1769b774703c406de5ad11abd9fb2f46dd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f9dq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:51Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.981023 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.981060 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:02:51 crc kubenswrapper[4700]: E0227 17:02:51.981172 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.981296 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:02:51 crc kubenswrapper[4700]: E0227 17:02:51.981621 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:02:51 crc kubenswrapper[4700]: I0227 17:02:51.981813 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:51 crc kubenswrapper[4700]: E0227 17:02:51.981905 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:02:51 crc kubenswrapper[4700]: E0227 17:02:51.982340 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:02:52 crc kubenswrapper[4700]: I0227 17:02:52.578766 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:52 crc kubenswrapper[4700]: I0227 17:02:52.578818 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:52 crc kubenswrapper[4700]: I0227 17:02:52.578855 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:52 crc kubenswrapper[4700]: I0227 17:02:52.578871 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:52 crc kubenswrapper[4700]: I0227 17:02:52.578884 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:52Z","lastTransitionTime":"2026-02-27T17:02:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:52 crc kubenswrapper[4700]: E0227 17:02:52.598723 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:52Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:52 crc kubenswrapper[4700]: I0227 17:02:52.605677 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:52 crc kubenswrapper[4700]: I0227 17:02:52.605765 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:52 crc kubenswrapper[4700]: I0227 17:02:52.605793 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:52 crc kubenswrapper[4700]: I0227 17:02:52.605826 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:52 crc kubenswrapper[4700]: I0227 17:02:52.605860 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:52Z","lastTransitionTime":"2026-02-27T17:02:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:52 crc kubenswrapper[4700]: E0227 17:02:52.629894 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:52Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:52 crc kubenswrapper[4700]: I0227 17:02:52.635716 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:52 crc kubenswrapper[4700]: I0227 17:02:52.635790 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:52 crc kubenswrapper[4700]: I0227 17:02:52.635941 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:52 crc kubenswrapper[4700]: I0227 17:02:52.635976 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:52 crc kubenswrapper[4700]: I0227 17:02:52.636000 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:52Z","lastTransitionTime":"2026-02-27T17:02:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:52 crc kubenswrapper[4700]: E0227 17:02:52.658927 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:52Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:52 crc kubenswrapper[4700]: I0227 17:02:52.663675 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:52 crc kubenswrapper[4700]: I0227 17:02:52.663735 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:52 crc kubenswrapper[4700]: I0227 17:02:52.663758 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:52 crc kubenswrapper[4700]: I0227 17:02:52.663788 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:52 crc kubenswrapper[4700]: I0227 17:02:52.663809 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:52Z","lastTransitionTime":"2026-02-27T17:02:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:52 crc kubenswrapper[4700]: E0227 17:02:52.684693 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:52Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:52 crc kubenswrapper[4700]: I0227 17:02:52.689725 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:02:52 crc kubenswrapper[4700]: I0227 17:02:52.689796 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:02:52 crc kubenswrapper[4700]: I0227 17:02:52.689814 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:02:52 crc kubenswrapper[4700]: I0227 17:02:52.689836 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:02:52 crc kubenswrapper[4700]: I0227 17:02:52.689852 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:02:52Z","lastTransitionTime":"2026-02-27T17:02:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:02:52 crc kubenswrapper[4700]: E0227 17:02:52.709518 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:52Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:52 crc kubenswrapper[4700]: E0227 17:02:52.709734 4700 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 27 17:02:53 crc kubenswrapper[4700]: I0227 17:02:53.711001 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-rbptz_66eb22fb-e593-40ff-9ff3-5bb32912972d/kube-multus/0.log" Feb 27 17:02:53 crc kubenswrapper[4700]: I0227 17:02:53.711063 4700 generic.go:334] "Generic (PLEG): container finished" podID="66eb22fb-e593-40ff-9ff3-5bb32912972d" containerID="f58f4814976a1e25bef53555d7f5d14fa1cc800c8c5220801960758d5262b552" exitCode=1 Feb 27 17:02:53 crc kubenswrapper[4700]: I0227 17:02:53.711095 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-rbptz" event={"ID":"66eb22fb-e593-40ff-9ff3-5bb32912972d","Type":"ContainerDied","Data":"f58f4814976a1e25bef53555d7f5d14fa1cc800c8c5220801960758d5262b552"} Feb 27 17:02:53 crc kubenswrapper[4700]: I0227 17:02:53.711543 4700 scope.go:117] "RemoveContainer" containerID="f58f4814976a1e25bef53555d7f5d14fa1cc800c8c5220801960758d5262b552" Feb 27 17:02:53 crc kubenswrapper[4700]: I0227 17:02:53.736689 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:53Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:53 crc kubenswrapper[4700]: I0227 17:02:53.752122 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dnkln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e772eecd-127f-4b26-8021-010fb507ce0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6df6de510d7e55695db4c89cbe9db84f91aca0e4bfbf2ed852407047c32822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v92s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dnkln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:53Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:53 crc kubenswrapper[4700]: I0227 17:02:53.772355 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47cf6a8e-df81-4ccb-992b-abb030862331\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857e486c0ff289c59f3de9933d1e0301be2f25b65e2b3f1fae6c4f149f900a4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8n2wb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:53Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:53 crc kubenswrapper[4700]: I0227 17:02:53.788739 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"309d44d0-45a4-4483-b588-7937902ee1fb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9da83072905af757ff9f9cecda0d94b2acfcd847aeeba7ca36ac291c4ff6a3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5a0dc977696434dce7529115ee1d6cce69cbd4c501bbfb77ed0d240d6340d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf6d6f90f1a5acb192ed932732bab2c2a2bebe9f03fddc79c5dfe6423fab012c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b6649f7710d4c1435bfd703120b70e4514f95b4750b247b8ad0a69ed6952a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b6649f7710d4c1435bfd703120b70e4514f95b4750b247b8ad0a69ed6952a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:53Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:53 crc kubenswrapper[4700]: I0227 17:02:53.806858 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:53Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:53 crc kubenswrapper[4700]: I0227 17:02:53.820790 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c501ee5970cba15e45e1f1c8f0e33768019511f94dbbe509a7bbfea2c5b4a064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f4cd0745c61d33ac717a4a655f00713e2bd935c3cbba23c4ca0913859f6016f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:53Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:53 crc kubenswrapper[4700]: I0227 17:02:53.833622 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a336c2f16e12b411f111f6291366aa1b466f6b405896c152d1667b832a589b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:53Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:53 crc kubenswrapper[4700]: I0227 17:02:53.861551 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a34c0e3-2513-4e81-b6b1-80d1230475fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb3fcd96bb1fdabe12481cd07fcde6f19f153b6f52d0b48b3788ef4a0cd491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2eb3fcd96bb1fdabe12481cd07fcde6f19f153b6f52d0b48b3788ef4a0cd491\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"message\\\":\\\"or.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992312 6876 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992349 6876 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992495 6876 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 17:02:31.992823 6876 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992944 6876 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 17:02:31.993180 6876 ovnkube.go:599] Stopped ovnkube\\\\nI0227 17:02:31.993236 6876 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0227 17:02:31.993318 6876 ovnkube.go:137] failed to run ovnkube: [failed to start network contr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jtbqn_openshift-ovn-kubernetes(3a34c0e3-2513-4e81-b6b1-80d1230475fd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jtbqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:53Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:53 crc kubenswrapper[4700]: I0227 17:02:53.877529 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b45511-b94c-479f-98db-fd2c4eceec46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bddf1a7b7264d36174b90f1d0d4e49f029781b1acfce88582aa5b2575dcdd09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ef7c37b61de7a039756248f222bd1769b774703c406de5ad11abd9fb2f46dd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f9dq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:53Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:53 crc kubenswrapper[4700]: I0227 17:02:53.888268 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-s8k74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5288317-4810-4321-baa0-31a118a4dc36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-s8k74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:53Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:53 crc kubenswrapper[4700]: I0227 17:02:53.899159 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f425b79-78da-4f9e-a1ad-205a09466052\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3baf8927284b91f3edeec1dc906c2928955badb19062087790d720936b6f7624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd29e6d703945e0f366f2d109b7c35fdd6360c77091b165616e9832dddac1c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cxnv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:53Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:53 crc kubenswrapper[4700]: I0227 17:02:53.920219 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ee6013b-8a2d-4651-ae5f-98aaac016d22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af0992b3383953e562bf540bf7659d41e60aeee5d6103105eaab164797486dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a14230f224fb3b7b9b0ced93634313b66ac3abc376dca228bc5802890b0c393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f09ed7c87894bc0473c7015f0e0d55a28f893f5a2e8bf9085c86654c464371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa2391deb934395fe4743249887832731bf7c6ea457d66484bcb6b556e4ab33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64972d043accab95774784cc95dc1144e70c855eff777664c66cb121d519d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:53Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:53 crc kubenswrapper[4700]: I0227 17:02:53.939428 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a2d1b57-7fc1-4614-aac5-7144cdd812d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d2d2c9065b008ff0936c19bcce2ce6a574580911a8188084c5f978e08177b69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"27 17:01:51.956663 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 17:01:51.956671 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 17:01:51.956678 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 17:01:51.956715 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0227 17:01:51.960396 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1772211711\\\\\\\\\\\\\\\" (2026-02-27 17:01:50 +0000 UTC to 2026-03-29 17:01:51 +0000 UTC (now=2026-02-27 17:01:51.960134581 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960704 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1772211711\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1772211711\\\\\\\\\\\\\\\" (2026-02-27 16:01:51 +0000 UTC to 2027-02-27 16:01:51 +0000 UTC (now=2026-02-27 17:01:51.960667784 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960745 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0227 17:01:51.960790 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0227 17:01:51.960837 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960878 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960925 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\"\\\\nI0227 17:01:51.961071 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0227 17:01:51.958239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:01:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:53Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:53 crc kubenswrapper[4700]: I0227 17:02:53.955656 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:53Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:53 crc kubenswrapper[4700]: I0227 17:02:53.976648 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rbptz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66eb22fb-e593-40ff-9ff3-5bb32912972d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f58f4814976a1e25bef53555d7f5d14fa1cc800c8c5220801960758d5262b552\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f58f4814976a1e25bef53555d7f5d14fa1cc800c8c5220801960758d5262b552\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T17:02:53Z\\\",\\\"message\\\":\\\"2026-02-27T17:02:07+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7a0028f6-f07c-4b10-a5a8-60005439ad4f\\\\n2026-02-27T17:02:07+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7a0028f6-f07c-4b10-a5a8-60005439ad4f to /host/opt/cni/bin/\\\\n2026-02-27T17:02:08Z [verbose] multus-daemon started\\\\n2026-02-27T17:02:08Z [verbose] Readiness Indicator file check\\\\n2026-02-27T17:02:53Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2dtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rbptz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:53Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:53 crc kubenswrapper[4700]: I0227 17:02:53.981062 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:02:53 crc kubenswrapper[4700]: I0227 17:02:53.981082 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:53 crc kubenswrapper[4700]: E0227 17:02:53.981150 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:02:53 crc kubenswrapper[4700]: I0227 17:02:53.981198 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:53 crc kubenswrapper[4700]: I0227 17:02:53.981089 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:02:53 crc kubenswrapper[4700]: E0227 17:02:53.981307 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:02:53 crc kubenswrapper[4700]: E0227 17:02:53.981415 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:02:53 crc kubenswrapper[4700]: E0227 17:02:53.981493 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:02:53 crc kubenswrapper[4700]: I0227 17:02:53.991366 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"032e16e8-09b3-456f-bebb-09d4b07bcebf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eabca65587272a0e2ce09d13f103fbbe5374fb6d5ed49bfa0e26de8336d9e683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:53Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:54 crc kubenswrapper[4700]: I0227 17:02:54.008697 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fcde75bb1f9dd52a2ac77ffaa7d60d8c785c399bbe45a7cbff39d1b926d5541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:54Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:54 crc kubenswrapper[4700]: I0227 17:02:54.024695 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-g66q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d4f1eeba71eeb576c1df386f5acaf2bc354e7ca08aa9aac0971f61db72f244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b6cx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-g66q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:54Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:54 crc kubenswrapper[4700]: I0227 17:02:54.726153 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-rbptz_66eb22fb-e593-40ff-9ff3-5bb32912972d/kube-multus/0.log" Feb 27 17:02:54 crc kubenswrapper[4700]: I0227 17:02:54.726263 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-rbptz" event={"ID":"66eb22fb-e593-40ff-9ff3-5bb32912972d","Type":"ContainerStarted","Data":"9c675574fe85ad3f55e0431b8d36fd1b83523ee502e675165aafad8551ba4560"} Feb 27 17:02:54 crc kubenswrapper[4700]: I0227 17:02:54.761589 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ee6013b-8a2d-4651-ae5f-98aaac016d22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af0992b3383953e562bf540bf7659d41e60aeee5d6103105eaab164797486dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a14230f224fb3b7b9b0ced93634313b66ac3abc376dca228bc5802890b0c393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f09ed7c87894bc0473c7015f0e0d55a28f893f5a2e8bf9085c86654c464371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa2391deb934395fe4743249887832731bf7c6ea457d66484bcb6b556e4ab33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64972d043accab95774784cc95dc1144e70c855eff777664c66cb121d519d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:54Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:54 crc kubenswrapper[4700]: I0227 17:02:54.786266 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a2d1b57-7fc1-4614-aac5-7144cdd812d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d2d2c9065b008ff0936c19bcce2ce6a574580911a8188084c5f978e08177b69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"27 17:01:51.956663 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 17:01:51.956671 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 17:01:51.956678 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 17:01:51.956715 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0227 17:01:51.960396 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1772211711\\\\\\\\\\\\\\\" (2026-02-27 17:01:50 +0000 UTC to 2026-03-29 17:01:51 +0000 UTC (now=2026-02-27 17:01:51.960134581 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960704 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1772211711\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1772211711\\\\\\\\\\\\\\\" (2026-02-27 16:01:51 +0000 UTC to 2027-02-27 16:01:51 +0000 UTC (now=2026-02-27 17:01:51.960667784 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960745 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0227 17:01:51.960790 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0227 17:01:51.960837 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960878 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960925 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\"\\\\nI0227 17:01:51.961071 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0227 17:01:51.958239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:01:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:54Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:54 crc kubenswrapper[4700]: I0227 17:02:54.805907 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:54Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:54 crc kubenswrapper[4700]: I0227 17:02:54.825101 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b45511-b94c-479f-98db-fd2c4eceec46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bddf1a7b7264d36174b90f1d0d4e49f029781b1acfce88582aa5b2575dcdd09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ef7c37b61de7a039756248f222bd1769b774703c406de5ad11abd9fb2f46dd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f9dq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:54Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:54 crc kubenswrapper[4700]: I0227 17:02:54.843829 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-s8k74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5288317-4810-4321-baa0-31a118a4dc36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-s8k74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:54Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:54 crc kubenswrapper[4700]: I0227 17:02:54.862759 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f425b79-78da-4f9e-a1ad-205a09466052\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3baf8927284b91f3edeec1dc906c2928955badb19062087790d720936b6f7624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd29e6d703945e0f366f2d109b7c35fdd6360c77091b165616e9832dddac1c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cxnv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:54Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:54 crc kubenswrapper[4700]: I0227 17:02:54.879630 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"032e16e8-09b3-456f-bebb-09d4b07bcebf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eabca65587272a0e2ce09d13f103fbbe5374fb6d5ed49bfa0e26de8336d9e683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:54Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:54 crc kubenswrapper[4700]: I0227 17:02:54.897840 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fcde75bb1f9dd52a2ac77ffaa7d60d8c785c399bbe45a7cbff39d1b926d5541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:54Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:54 crc kubenswrapper[4700]: I0227 17:02:54.913129 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-g66q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d4f1eeba71eeb576c1df386f5acaf2bc354e7ca08aa9aac0971f61db72f244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b6cx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-g66q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:54Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:54 crc kubenswrapper[4700]: I0227 17:02:54.934869 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rbptz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66eb22fb-e593-40ff-9ff3-5bb32912972d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c675574fe85ad3f55e0431b8d36fd1b83523ee502e675165aafad8551ba4560\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f58f4814976a1e25bef53555d7f5d14fa1cc800c8c5220801960758d5262b552\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T17:02:53Z\\\",\\\"message\\\":\\\"2026-02-27T17:02:07+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7a0028f6-f07c-4b10-a5a8-60005439ad4f\\\\n2026-02-27T17:02:07+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7a0028f6-f07c-4b10-a5a8-60005439ad4f to /host/opt/cni/bin/\\\\n2026-02-27T17:02:08Z [verbose] multus-daemon started\\\\n2026-02-27T17:02:08Z [verbose] Readiness Indicator file check\\\\n2026-02-27T17:02:53Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2dtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rbptz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:54Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:54 crc kubenswrapper[4700]: I0227 17:02:54.954357 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"309d44d0-45a4-4483-b588-7937902ee1fb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9da83072905af757ff9f9cecda0d94b2acfcd847aeeba7ca36ac291c4ff6a3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5a0dc977696434dce7529115ee1d6cce69cbd4c501bbfb77ed0d240d6340d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf6d6f90f1a5acb192ed932732bab2c2a2bebe9f03fddc79c5dfe6423fab012c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b6649f7710d4c1435bfd703120b70e4514f95b4750b247b8ad0a69ed6952a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b6649f7710d4c1435bfd703120b70e4514f95b4750b247b8ad0a69ed6952a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:54Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:54 crc kubenswrapper[4700]: I0227 17:02:54.974739 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:54Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:54 crc kubenswrapper[4700]: I0227 17:02:54.993620 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c501ee5970cba15e45e1f1c8f0e33768019511f94dbbe509a7bbfea2c5b4a064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f4cd0745c61d33ac717a4a655f00713e2bd935c3cbba23c4ca0913859f6016f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:54Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:55 crc kubenswrapper[4700]: I0227 17:02:55.013897 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:55Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:55 crc kubenswrapper[4700]: I0227 17:02:55.030254 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dnkln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e772eecd-127f-4b26-8021-010fb507ce0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6df6de510d7e55695db4c89cbe9db84f91aca0e4bfbf2ed852407047c32822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v92s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dnkln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:55Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:55 crc kubenswrapper[4700]: I0227 17:02:55.053213 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47cf6a8e-df81-4ccb-992b-abb030862331\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857e486c0ff289c59f3de9933d1e0301be2f25b65e2b3f1fae6c4f149f900a4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8n2wb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:55Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:55 crc kubenswrapper[4700]: I0227 17:02:55.073498 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a336c2f16e12b411f111f6291366aa1b466f6b405896c152d1667b832a589b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:55Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:55 crc kubenswrapper[4700]: I0227 17:02:55.098353 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a34c0e3-2513-4e81-b6b1-80d1230475fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb3fcd96bb1fdabe12481cd07fcde6f19f153b6f52d0b48b3788ef4a0cd491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2eb3fcd96bb1fdabe12481cd07fcde6f19f153b6f52d0b48b3788ef4a0cd491\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"message\\\":\\\"or.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992312 6876 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992349 6876 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992495 6876 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 17:02:31.992823 6876 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992944 6876 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 17:02:31.993180 6876 ovnkube.go:599] Stopped ovnkube\\\\nI0227 17:02:31.993236 6876 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0227 17:02:31.993318 6876 ovnkube.go:137] failed to run ovnkube: [failed to start network contr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jtbqn_openshift-ovn-kubernetes(3a34c0e3-2513-4e81-b6b1-80d1230475fd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jtbqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:55Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:55 crc kubenswrapper[4700]: I0227 17:02:55.981024 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:55 crc kubenswrapper[4700]: I0227 17:02:55.981107 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:02:55 crc kubenswrapper[4700]: I0227 17:02:55.981198 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:02:55 crc kubenswrapper[4700]: E0227 17:02:55.981252 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:02:55 crc kubenswrapper[4700]: I0227 17:02:55.981296 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:55 crc kubenswrapper[4700]: E0227 17:02:55.981444 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:02:55 crc kubenswrapper[4700]: E0227 17:02:55.981640 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:02:55 crc kubenswrapper[4700]: E0227 17:02:55.981843 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:02:56 crc kubenswrapper[4700]: E0227 17:02:56.105771 4700 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 17:02:57 crc kubenswrapper[4700]: I0227 17:02:57.980123 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:02:57 crc kubenswrapper[4700]: I0227 17:02:57.980154 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:57 crc kubenswrapper[4700]: E0227 17:02:57.980825 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:02:57 crc kubenswrapper[4700]: I0227 17:02:57.980881 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:02:57 crc kubenswrapper[4700]: I0227 17:02:57.982766 4700 scope.go:117] "RemoveContainer" containerID="c2eb3fcd96bb1fdabe12481cd07fcde6f19f153b6f52d0b48b3788ef4a0cd491" Feb 27 17:02:57 crc kubenswrapper[4700]: I0227 17:02:57.983595 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:57 crc kubenswrapper[4700]: E0227 17:02:57.983868 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:02:57 crc kubenswrapper[4700]: E0227 17:02:57.984238 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:02:57 crc kubenswrapper[4700]: E0227 17:02:57.984716 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:02:58 crc kubenswrapper[4700]: I0227 17:02:58.747380 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jtbqn_3a34c0e3-2513-4e81-b6b1-80d1230475fd/ovnkube-controller/2.log" Feb 27 17:02:58 crc kubenswrapper[4700]: I0227 17:02:58.750986 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" event={"ID":"3a34c0e3-2513-4e81-b6b1-80d1230475fd","Type":"ContainerStarted","Data":"da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49"} Feb 27 17:02:58 crc kubenswrapper[4700]: I0227 17:02:58.751679 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:02:58 crc kubenswrapper[4700]: I0227 17:02:58.777579 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ee6013b-8a2d-4651-ae5f-98aaac016d22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af0992b3383953e562bf540bf7659d41e60aeee5d6103105eaab164797486dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a14230f224fb3b7b9b0ced93634313b66ac3abc376dca228bc5802890b0c393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f09ed7c87894bc0473c7015f0e0d55a28f893f5a2e8bf9085c86654c464371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa2391deb934395fe4743249887832731bf7c6ea457d66484bcb6b556e4ab33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64972d043accab95774784cc95dc1144e70c855eff777664c66cb121d519d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:58Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:58 crc kubenswrapper[4700]: I0227 17:02:58.797617 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a2d1b57-7fc1-4614-aac5-7144cdd812d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d2d2c9065b008ff0936c19bcce2ce6a574580911a8188084c5f978e08177b69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"27 17:01:51.956663 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 17:01:51.956671 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 17:01:51.956678 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 17:01:51.956715 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0227 17:01:51.960396 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1772211711\\\\\\\\\\\\\\\" (2026-02-27 17:01:50 +0000 UTC to 2026-03-29 17:01:51 +0000 UTC (now=2026-02-27 17:01:51.960134581 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960704 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1772211711\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1772211711\\\\\\\\\\\\\\\" (2026-02-27 16:01:51 +0000 UTC to 2027-02-27 16:01:51 +0000 UTC (now=2026-02-27 17:01:51.960667784 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960745 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0227 17:01:51.960790 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0227 17:01:51.960837 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960878 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960925 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\"\\\\nI0227 17:01:51.961071 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0227 17:01:51.958239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:01:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:58Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:58 crc kubenswrapper[4700]: I0227 17:02:58.814556 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:58Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:58 crc kubenswrapper[4700]: I0227 17:02:58.827138 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b45511-b94c-479f-98db-fd2c4eceec46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bddf1a7b7264d36174b90f1d0d4e49f029781b1acfce88582aa5b2575dcdd09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ef7c37b61de7a039756248f222bd1769b774703c406de5ad11abd9fb2f46dd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f9dq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:58Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:58 crc kubenswrapper[4700]: I0227 17:02:58.844445 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-s8k74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5288317-4810-4321-baa0-31a118a4dc36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-s8k74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:58Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:58 crc kubenswrapper[4700]: I0227 17:02:58.860920 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f425b79-78da-4f9e-a1ad-205a09466052\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3baf8927284b91f3edeec1dc906c2928955badb19062087790d720936b6f7624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd29e6d703945e0f366f2d109b7c35fdd6360c77091b165616e9832dddac1c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cxnv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:58Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:58 crc kubenswrapper[4700]: I0227 17:02:58.874318 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"032e16e8-09b3-456f-bebb-09d4b07bcebf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eabca65587272a0e2ce09d13f103fbbe5374fb6d5ed49bfa0e26de8336d9e683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:58Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:58 crc kubenswrapper[4700]: I0227 17:02:58.888623 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fcde75bb1f9dd52a2ac77ffaa7d60d8c785c399bbe45a7cbff39d1b926d5541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:58Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:58 crc kubenswrapper[4700]: I0227 17:02:58.900421 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-g66q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d4f1eeba71eeb576c1df386f5acaf2bc354e7ca08aa9aac0971f61db72f244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b6cx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-g66q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:58Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:58 crc kubenswrapper[4700]: I0227 17:02:58.919033 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rbptz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66eb22fb-e593-40ff-9ff3-5bb32912972d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c675574fe85ad3f55e0431b8d36fd1b83523ee502e675165aafad8551ba4560\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f58f4814976a1e25bef53555d7f5d14fa1cc800c8c5220801960758d5262b552\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T17:02:53Z\\\",\\\"message\\\":\\\"2026-02-27T17:02:07+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7a0028f6-f07c-4b10-a5a8-60005439ad4f\\\\n2026-02-27T17:02:07+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7a0028f6-f07c-4b10-a5a8-60005439ad4f to /host/opt/cni/bin/\\\\n2026-02-27T17:02:08Z [verbose] multus-daemon started\\\\n2026-02-27T17:02:08Z [verbose] Readiness Indicator file check\\\\n2026-02-27T17:02:53Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2dtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rbptz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:58Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:58 crc kubenswrapper[4700]: I0227 17:02:58.933191 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"309d44d0-45a4-4483-b588-7937902ee1fb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9da83072905af757ff9f9cecda0d94b2acfcd847aeeba7ca36ac291c4ff6a3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5a0dc977696434dce7529115ee1d6cce69cbd4c501bbfb77ed0d240d6340d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf6d6f90f1a5acb192ed932732bab2c2a2bebe9f03fddc79c5dfe6423fab012c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b6649f7710d4c1435bfd703120b70e4514f95b4750b247b8ad0a69ed6952a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b6649f7710d4c1435bfd703120b70e4514f95b4750b247b8ad0a69ed6952a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:58Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:58 crc kubenswrapper[4700]: I0227 17:02:58.955360 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:58Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:58 crc kubenswrapper[4700]: I0227 17:02:58.976043 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c501ee5970cba15e45e1f1c8f0e33768019511f94dbbe509a7bbfea2c5b4a064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f4cd0745c61d33ac717a4a655f00713e2bd935c3cbba23c4ca0913859f6016f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:58Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:58 crc kubenswrapper[4700]: I0227 17:02:58.991922 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:58Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:59 crc kubenswrapper[4700]: I0227 17:02:59.005162 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dnkln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e772eecd-127f-4b26-8021-010fb507ce0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6df6de510d7e55695db4c89cbe9db84f91aca0e4bfbf2ed852407047c32822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v92s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dnkln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:59Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:59 crc kubenswrapper[4700]: I0227 17:02:59.019756 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47cf6a8e-df81-4ccb-992b-abb030862331\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857e486c0ff289c59f3de9933d1e0301be2f25b65e2b3f1fae6c4f149f900a4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8n2wb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:59Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:59 crc kubenswrapper[4700]: I0227 17:02:59.033186 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a336c2f16e12b411f111f6291366aa1b466f6b405896c152d1667b832a589b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:59Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:59 crc kubenswrapper[4700]: I0227 17:02:59.051440 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a34c0e3-2513-4e81-b6b1-80d1230475fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2eb3fcd96bb1fdabe12481cd07fcde6f19f153b6f52d0b48b3788ef4a0cd491\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"message\\\":\\\"or.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992312 6876 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992349 6876 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992495 6876 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 17:02:31.992823 6876 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992944 6876 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 17:02:31.993180 6876 ovnkube.go:599] Stopped ovnkube\\\\nI0227 17:02:31.993236 6876 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0227 17:02:31.993318 6876 ovnkube.go:137] failed to run ovnkube: [failed to start network contr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jtbqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:59Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:59 crc kubenswrapper[4700]: I0227 17:02:59.757417 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jtbqn_3a34c0e3-2513-4e81-b6b1-80d1230475fd/ovnkube-controller/3.log" Feb 27 17:02:59 crc kubenswrapper[4700]: I0227 17:02:59.758373 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jtbqn_3a34c0e3-2513-4e81-b6b1-80d1230475fd/ovnkube-controller/2.log" Feb 27 17:02:59 crc kubenswrapper[4700]: I0227 17:02:59.762301 4700 generic.go:334] "Generic (PLEG): container finished" podID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerID="da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49" exitCode=1 Feb 27 17:02:59 crc kubenswrapper[4700]: I0227 17:02:59.762358 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" event={"ID":"3a34c0e3-2513-4e81-b6b1-80d1230475fd","Type":"ContainerDied","Data":"da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49"} Feb 27 17:02:59 crc kubenswrapper[4700]: I0227 17:02:59.762415 4700 scope.go:117] "RemoveContainer" containerID="c2eb3fcd96bb1fdabe12481cd07fcde6f19f153b6f52d0b48b3788ef4a0cd491" Feb 27 17:02:59 crc kubenswrapper[4700]: I0227 17:02:59.763128 4700 scope.go:117] "RemoveContainer" containerID="da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49" Feb 27 17:02:59 crc kubenswrapper[4700]: E0227 17:02:59.763322 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jtbqn_openshift-ovn-kubernetes(3a34c0e3-2513-4e81-b6b1-80d1230475fd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" Feb 27 17:02:59 crc kubenswrapper[4700]: I0227 17:02:59.781386 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-g66q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d4f1eeba71eeb576c1df386f5acaf2bc354e7ca08aa9aac0971f61db72f244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b6cx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-g66q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:59Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:59 crc kubenswrapper[4700]: I0227 17:02:59.804500 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rbptz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66eb22fb-e593-40ff-9ff3-5bb32912972d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c675574fe85ad3f55e0431b8d36fd1b83523ee502e675165aafad8551ba4560\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f58f4814976a1e25bef53555d7f5d14fa1cc800c8c5220801960758d5262b552\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T17:02:53Z\\\",\\\"message\\\":\\\"2026-02-27T17:02:07+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7a0028f6-f07c-4b10-a5a8-60005439ad4f\\\\n2026-02-27T17:02:07+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7a0028f6-f07c-4b10-a5a8-60005439ad4f to /host/opt/cni/bin/\\\\n2026-02-27T17:02:08Z [verbose] multus-daemon started\\\\n2026-02-27T17:02:08Z [verbose] Readiness Indicator file check\\\\n2026-02-27T17:02:53Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2dtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rbptz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:59Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:59 crc kubenswrapper[4700]: I0227 17:02:59.821373 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"032e16e8-09b3-456f-bebb-09d4b07bcebf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eabca65587272a0e2ce09d13f103fbbe5374fb6d5ed49bfa0e26de8336d9e683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:59Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:59 crc kubenswrapper[4700]: I0227 17:02:59.840723 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fcde75bb1f9dd52a2ac77ffaa7d60d8c785c399bbe45a7cbff39d1b926d5541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:59Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:59 crc kubenswrapper[4700]: I0227 17:02:59.861573 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c501ee5970cba15e45e1f1c8f0e33768019511f94dbbe509a7bbfea2c5b4a064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f4cd0745c61d33ac717a4a655f00713e2bd935c3cbba23c4ca0913859f6016f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:59Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:59 crc kubenswrapper[4700]: I0227 17:02:59.880971 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:59Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:59 crc kubenswrapper[4700]: I0227 17:02:59.898780 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dnkln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e772eecd-127f-4b26-8021-010fb507ce0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6df6de510d7e55695db4c89cbe9db84f91aca0e4bfbf2ed852407047c32822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v92s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dnkln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:59Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:59 crc kubenswrapper[4700]: I0227 17:02:59.924006 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47cf6a8e-df81-4ccb-992b-abb030862331\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857e486c0ff289c59f3de9933d1e0301be2f25b65e2b3f1fae6c4f149f900a4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8n2wb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:59Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:59 crc kubenswrapper[4700]: I0227 17:02:59.944587 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"309d44d0-45a4-4483-b588-7937902ee1fb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9da83072905af757ff9f9cecda0d94b2acfcd847aeeba7ca36ac291c4ff6a3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5a0dc977696434dce7529115ee1d6cce69cbd4c501bbfb77ed0d240d6340d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf6d6f90f1a5acb192ed932732bab2c2a2bebe9f03fddc79c5dfe6423fab012c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b6649f7710d4c1435bfd703120b70e4514f95b4750b247b8ad0a69ed6952a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b6649f7710d4c1435bfd703120b70e4514f95b4750b247b8ad0a69ed6952a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:59Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:59 crc kubenswrapper[4700]: I0227 17:02:59.961769 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:59Z is after 2025-08-24T17:21:41Z" Feb 27 17:02:59 crc kubenswrapper[4700]: I0227 17:02:59.980143 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:02:59 crc kubenswrapper[4700]: I0227 17:02:59.980198 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:02:59 crc kubenswrapper[4700]: I0227 17:02:59.980306 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:02:59 crc kubenswrapper[4700]: I0227 17:02:59.980606 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:02:59 crc kubenswrapper[4700]: E0227 17:02:59.980574 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:02:59 crc kubenswrapper[4700]: E0227 17:02:59.980778 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:02:59 crc kubenswrapper[4700]: E0227 17:02:59.980959 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:02:59 crc kubenswrapper[4700]: E0227 17:02:59.981101 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:03:00 crc kubenswrapper[4700]: I0227 17:03:00.000251 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a34c0e3-2513-4e81-b6b1-80d1230475fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2eb3fcd96bb1fdabe12481cd07fcde6f19f153b6f52d0b48b3788ef4a0cd491\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T17:02:32Z\\\",\\\"message\\\":\\\"or.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992312 6876 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992349 6876 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992495 6876 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 17:02:31.992823 6876 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0227 17:02:31.992944 6876 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 17:02:31.993180 6876 ovnkube.go:599] Stopped ovnkube\\\\nI0227 17:02:31.993236 6876 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0227 17:02:31.993318 6876 ovnkube.go:137] failed to run ovnkube: [failed to start network contr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T17:02:58Z\\\",\\\"message\\\":\\\"d (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 17:02:58.943875 7219 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0227 17:02:58.943947 7219 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0227 17:02:58.943980 7219 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0227 17:02:58.944014 7219 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0227 17:02:58.944017 7219 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0227 17:02:58.944044 7219 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0227 17:02:58.944060 7219 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0227 17:02:58.944071 7219 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0227 17:02:58.944079 7219 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0227 17:02:58.944097 7219 handler.go:208] Removed *v1.Node event handler 2\\\\nI0227 17:02:58.944115 7219 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0227 17:02:58.944177 7219 factory.go:656] Stopping watch factory\\\\nI0227 17:02:58.944208 7219 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0227 17:02:58.944209 7219 handler.go:208] Removed *v1.Node event handler 7\\\\nI0227 17:02:58.944107 7219 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jtbqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:02:59Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:00 crc kubenswrapper[4700]: I0227 17:03:00.023958 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a336c2f16e12b411f111f6291366aa1b466f6b405896c152d1667b832a589b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:00Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:00 crc kubenswrapper[4700]: I0227 17:03:00.043950 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:00Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:00 crc kubenswrapper[4700]: I0227 17:03:00.063585 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b45511-b94c-479f-98db-fd2c4eceec46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bddf1a7b7264d36174b90f1d0d4e49f029781b1acfce88582aa5b2575dcdd09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ef7c37b61de7a039756248f222bd1769b774703c406de5ad11abd9fb2f46dd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f9dq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:00Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:00 crc kubenswrapper[4700]: I0227 17:03:00.081883 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-s8k74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5288317-4810-4321-baa0-31a118a4dc36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-s8k74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:00Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:00 crc kubenswrapper[4700]: I0227 17:03:00.100330 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f425b79-78da-4f9e-a1ad-205a09466052\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3baf8927284b91f3edeec1dc906c2928955badb19062087790d720936b6f7624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd29e6d703945e0f366f2d109b7c35fdd6360c77091b165616e9832dddac1c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cxnv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:00Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:00 crc kubenswrapper[4700]: I0227 17:03:00.133194 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ee6013b-8a2d-4651-ae5f-98aaac016d22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af0992b3383953e562bf540bf7659d41e60aeee5d6103105eaab164797486dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a14230f224fb3b7b9b0ced93634313b66ac3abc376dca228bc5802890b0c393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f09ed7c87894bc0473c7015f0e0d55a28f893f5a2e8bf9085c86654c464371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa2391deb934395fe4743249887832731bf7c6ea457d66484bcb6b556e4ab33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64972d043accab95774784cc95dc1144e70c855eff777664c66cb121d519d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:00Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:00 crc kubenswrapper[4700]: I0227 17:03:00.156916 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a2d1b57-7fc1-4614-aac5-7144cdd812d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d2d2c9065b008ff0936c19bcce2ce6a574580911a8188084c5f978e08177b69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"27 17:01:51.956663 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 17:01:51.956671 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 17:01:51.956678 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 17:01:51.956715 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0227 17:01:51.960396 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1772211711\\\\\\\\\\\\\\\" (2026-02-27 17:01:50 +0000 UTC to 2026-03-29 17:01:51 +0000 UTC (now=2026-02-27 17:01:51.960134581 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960704 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1772211711\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1772211711\\\\\\\\\\\\\\\" (2026-02-27 16:01:51 +0000 UTC to 2027-02-27 16:01:51 +0000 UTC (now=2026-02-27 17:01:51.960667784 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960745 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0227 17:01:51.960790 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0227 17:01:51.960837 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960878 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960925 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\"\\\\nI0227 17:01:51.961071 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0227 17:01:51.958239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:01:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:00Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:00 crc kubenswrapper[4700]: I0227 17:03:00.768622 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jtbqn_3a34c0e3-2513-4e81-b6b1-80d1230475fd/ovnkube-controller/3.log" Feb 27 17:03:00 crc kubenswrapper[4700]: I0227 17:03:00.773420 4700 scope.go:117] "RemoveContainer" containerID="da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49" Feb 27 17:03:00 crc kubenswrapper[4700]: E0227 17:03:00.773756 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jtbqn_openshift-ovn-kubernetes(3a34c0e3-2513-4e81-b6b1-80d1230475fd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" Feb 27 17:03:00 crc kubenswrapper[4700]: I0227 17:03:00.792204 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"032e16e8-09b3-456f-bebb-09d4b07bcebf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eabca65587272a0e2ce09d13f103fbbe5374fb6d5ed49bfa0e26de8336d9e683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:00Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:00 crc kubenswrapper[4700]: I0227 17:03:00.813195 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fcde75bb1f9dd52a2ac77ffaa7d60d8c785c399bbe45a7cbff39d1b926d5541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:00Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:00 crc kubenswrapper[4700]: I0227 17:03:00.829925 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-g66q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d4f1eeba71eeb576c1df386f5acaf2bc354e7ca08aa9aac0971f61db72f244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b6cx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-g66q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:00Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:00 crc kubenswrapper[4700]: I0227 17:03:00.852596 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rbptz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66eb22fb-e593-40ff-9ff3-5bb32912972d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c675574fe85ad3f55e0431b8d36fd1b83523ee502e675165aafad8551ba4560\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f58f4814976a1e25bef53555d7f5d14fa1cc800c8c5220801960758d5262b552\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T17:02:53Z\\\",\\\"message\\\":\\\"2026-02-27T17:02:07+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7a0028f6-f07c-4b10-a5a8-60005439ad4f\\\\n2026-02-27T17:02:07+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7a0028f6-f07c-4b10-a5a8-60005439ad4f to /host/opt/cni/bin/\\\\n2026-02-27T17:02:08Z [verbose] multus-daemon started\\\\n2026-02-27T17:02:08Z [verbose] Readiness Indicator file check\\\\n2026-02-27T17:02:53Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2dtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rbptz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:00Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:00 crc kubenswrapper[4700]: I0227 17:03:00.871363 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"309d44d0-45a4-4483-b588-7937902ee1fb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9da83072905af757ff9f9cecda0d94b2acfcd847aeeba7ca36ac291c4ff6a3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5a0dc977696434dce7529115ee1d6cce69cbd4c501bbfb77ed0d240d6340d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf6d6f90f1a5acb192ed932732bab2c2a2bebe9f03fddc79c5dfe6423fab012c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b6649f7710d4c1435bfd703120b70e4514f95b4750b247b8ad0a69ed6952a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b6649f7710d4c1435bfd703120b70e4514f95b4750b247b8ad0a69ed6952a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:00Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:00 crc kubenswrapper[4700]: I0227 17:03:00.911795 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:00Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:00 crc kubenswrapper[4700]: I0227 17:03:00.944105 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c501ee5970cba15e45e1f1c8f0e33768019511f94dbbe509a7bbfea2c5b4a064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f4cd0745c61d33ac717a4a655f00713e2bd935c3cbba23c4ca0913859f6016f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:00Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:00 crc kubenswrapper[4700]: I0227 17:03:00.958746 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:00Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:00 crc kubenswrapper[4700]: I0227 17:03:00.970755 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dnkln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e772eecd-127f-4b26-8021-010fb507ce0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6df6de510d7e55695db4c89cbe9db84f91aca0e4bfbf2ed852407047c32822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v92s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dnkln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:00Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:00 crc kubenswrapper[4700]: I0227 17:03:00.992028 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47cf6a8e-df81-4ccb-992b-abb030862331\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857e486c0ff289c59f3de9933d1e0301be2f25b65e2b3f1fae6c4f149f900a4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8n2wb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:00Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:01 crc kubenswrapper[4700]: I0227 17:03:01.011656 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a336c2f16e12b411f111f6291366aa1b466f6b405896c152d1667b832a589b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:01Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:01 crc kubenswrapper[4700]: I0227 17:03:01.039699 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a34c0e3-2513-4e81-b6b1-80d1230475fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T17:02:58Z\\\",\\\"message\\\":\\\"d (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 17:02:58.943875 7219 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0227 17:02:58.943947 7219 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0227 17:02:58.943980 7219 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0227 17:02:58.944014 7219 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0227 17:02:58.944017 7219 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0227 17:02:58.944044 7219 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0227 17:02:58.944060 7219 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0227 17:02:58.944071 7219 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0227 17:02:58.944079 7219 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0227 17:02:58.944097 7219 handler.go:208] Removed *v1.Node event handler 2\\\\nI0227 17:02:58.944115 7219 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0227 17:02:58.944177 7219 factory.go:656] Stopping watch factory\\\\nI0227 17:02:58.944208 7219 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0227 17:02:58.944209 7219 handler.go:208] Removed *v1.Node event handler 7\\\\nI0227 17:02:58.944107 7219 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jtbqn_openshift-ovn-kubernetes(3a34c0e3-2513-4e81-b6b1-80d1230475fd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jtbqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:01Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:01 crc kubenswrapper[4700]: I0227 17:03:01.070974 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ee6013b-8a2d-4651-ae5f-98aaac016d22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af0992b3383953e562bf540bf7659d41e60aeee5d6103105eaab164797486dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a14230f224fb3b7b9b0ced93634313b66ac3abc376dca228bc5802890b0c393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f09ed7c87894bc0473c7015f0e0d55a28f893f5a2e8bf9085c86654c464371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa2391deb934395fe4743249887832731bf7c6ea457d66484bcb6b556e4ab33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64972d043accab95774784cc95dc1144e70c855eff777664c66cb121d519d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:01Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:01 crc kubenswrapper[4700]: I0227 17:03:01.088563 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a2d1b57-7fc1-4614-aac5-7144cdd812d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d2d2c9065b008ff0936c19bcce2ce6a574580911a8188084c5f978e08177b69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"27 17:01:51.956663 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 17:01:51.956671 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 17:01:51.956678 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 17:01:51.956715 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0227 17:01:51.960396 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1772211711\\\\\\\\\\\\\\\" (2026-02-27 17:01:50 +0000 UTC to 2026-03-29 17:01:51 +0000 UTC (now=2026-02-27 17:01:51.960134581 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960704 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1772211711\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1772211711\\\\\\\\\\\\\\\" (2026-02-27 16:01:51 +0000 UTC to 2027-02-27 16:01:51 +0000 UTC (now=2026-02-27 17:01:51.960667784 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960745 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0227 17:01:51.960790 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0227 17:01:51.960837 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960878 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960925 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\"\\\\nI0227 17:01:51.961071 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0227 17:01:51.958239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:01:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:01Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:01 crc kubenswrapper[4700]: E0227 17:03:01.106687 4700 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 17:03:01 crc kubenswrapper[4700]: I0227 17:03:01.116929 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:01Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:01 crc kubenswrapper[4700]: I0227 17:03:01.133721 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b45511-b94c-479f-98db-fd2c4eceec46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bddf1a7b7264d36174b90f1d0d4e49f029781b1acfce88582aa5b2575dcdd09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ef7c37b61de7a039756248f222bd1769b774703c406de5ad11abd9fb2f46dd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f9dq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:01Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:01 crc kubenswrapper[4700]: I0227 17:03:01.149327 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-s8k74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5288317-4810-4321-baa0-31a118a4dc36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-s8k74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:01Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:01 crc kubenswrapper[4700]: I0227 17:03:01.162928 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f425b79-78da-4f9e-a1ad-205a09466052\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3baf8927284b91f3edeec1dc906c2928955badb19062087790d720936b6f7624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd29e6d703945e0f366f2d109b7c35fdd6360c77091b165616e9832dddac1c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cxnv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:01Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:01 crc kubenswrapper[4700]: I0227 17:03:01.178244 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-g66q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d4f1eeba71eeb576c1df386f5acaf2bc354e7ca08aa9aac0971f61db72f244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b6cx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-g66q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:01Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:01 crc kubenswrapper[4700]: I0227 17:03:01.190942 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rbptz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66eb22fb-e593-40ff-9ff3-5bb32912972d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c675574fe85ad3f55e0431b8d36fd1b83523ee502e675165aafad8551ba4560\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f58f4814976a1e25bef53555d7f5d14fa1cc800c8c5220801960758d5262b552\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T17:02:53Z\\\",\\\"message\\\":\\\"2026-02-27T17:02:07+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7a0028f6-f07c-4b10-a5a8-60005439ad4f\\\\n2026-02-27T17:02:07+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7a0028f6-f07c-4b10-a5a8-60005439ad4f to /host/opt/cni/bin/\\\\n2026-02-27T17:02:08Z [verbose] multus-daemon started\\\\n2026-02-27T17:02:08Z [verbose] Readiness Indicator file check\\\\n2026-02-27T17:02:53Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2dtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rbptz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:01Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:01 crc kubenswrapper[4700]: I0227 17:03:01.202651 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"032e16e8-09b3-456f-bebb-09d4b07bcebf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eabca65587272a0e2ce09d13f103fbbe5374fb6d5ed49bfa0e26de8336d9e683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:01Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:01 crc kubenswrapper[4700]: I0227 17:03:01.218542 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fcde75bb1f9dd52a2ac77ffaa7d60d8c785c399bbe45a7cbff39d1b926d5541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:01Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:01 crc kubenswrapper[4700]: I0227 17:03:01.232093 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c501ee5970cba15e45e1f1c8f0e33768019511f94dbbe509a7bbfea2c5b4a064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f4cd0745c61d33ac717a4a655f00713e2bd935c3cbba23c4ca0913859f6016f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:01Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:01 crc kubenswrapper[4700]: I0227 17:03:01.250052 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:01Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:01 crc kubenswrapper[4700]: I0227 17:03:01.264011 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dnkln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e772eecd-127f-4b26-8021-010fb507ce0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6df6de510d7e55695db4c89cbe9db84f91aca0e4bfbf2ed852407047c32822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v92s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dnkln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:01Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:01 crc kubenswrapper[4700]: I0227 17:03:01.281606 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47cf6a8e-df81-4ccb-992b-abb030862331\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857e486c0ff289c59f3de9933d1e0301be2f25b65e2b3f1fae6c4f149f900a4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8n2wb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:01Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:01 crc kubenswrapper[4700]: I0227 17:03:01.299649 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"309d44d0-45a4-4483-b588-7937902ee1fb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9da83072905af757ff9f9cecda0d94b2acfcd847aeeba7ca36ac291c4ff6a3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5a0dc977696434dce7529115ee1d6cce69cbd4c501bbfb77ed0d240d6340d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf6d6f90f1a5acb192ed932732bab2c2a2bebe9f03fddc79c5dfe6423fab012c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b6649f7710d4c1435bfd703120b70e4514f95b4750b247b8ad0a69ed6952a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b6649f7710d4c1435bfd703120b70e4514f95b4750b247b8ad0a69ed6952a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:01Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:01 crc kubenswrapper[4700]: I0227 17:03:01.315598 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:01Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:01 crc kubenswrapper[4700]: I0227 17:03:01.347139 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a34c0e3-2513-4e81-b6b1-80d1230475fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T17:02:58Z\\\",\\\"message\\\":\\\"d (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 17:02:58.943875 7219 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0227 17:02:58.943947 7219 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0227 17:02:58.943980 7219 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0227 17:02:58.944014 7219 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0227 17:02:58.944017 7219 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0227 17:02:58.944044 7219 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0227 17:02:58.944060 7219 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0227 17:02:58.944071 7219 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0227 17:02:58.944079 7219 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0227 17:02:58.944097 7219 handler.go:208] Removed *v1.Node event handler 2\\\\nI0227 17:02:58.944115 7219 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0227 17:02:58.944177 7219 factory.go:656] Stopping watch factory\\\\nI0227 17:02:58.944208 7219 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0227 17:02:58.944209 7219 handler.go:208] Removed *v1.Node event handler 7\\\\nI0227 17:02:58.944107 7219 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jtbqn_openshift-ovn-kubernetes(3a34c0e3-2513-4e81-b6b1-80d1230475fd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jtbqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:01Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:01 crc kubenswrapper[4700]: I0227 17:03:01.361665 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a336c2f16e12b411f111f6291366aa1b466f6b405896c152d1667b832a589b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:01Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:01 crc kubenswrapper[4700]: I0227 17:03:01.381023 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:01Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:01 crc kubenswrapper[4700]: I0227 17:03:01.395947 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b45511-b94c-479f-98db-fd2c4eceec46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bddf1a7b7264d36174b90f1d0d4e49f029781b1acfce88582aa5b2575dcdd09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ef7c37b61de7a039756248f222bd1769b774703c406de5ad11abd9fb2f46dd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f9dq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:01Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:01 crc kubenswrapper[4700]: I0227 17:03:01.411056 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-s8k74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5288317-4810-4321-baa0-31a118a4dc36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-s8k74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:01Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:01 crc kubenswrapper[4700]: I0227 17:03:01.426319 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f425b79-78da-4f9e-a1ad-205a09466052\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3baf8927284b91f3edeec1dc906c2928955badb19062087790d720936b6f7624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd29e6d703945e0f366f2d109b7c35fdd6360c77091b165616e9832dddac1c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cxnv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:01Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:01 crc kubenswrapper[4700]: I0227 17:03:01.463822 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ee6013b-8a2d-4651-ae5f-98aaac016d22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af0992b3383953e562bf540bf7659d41e60aeee5d6103105eaab164797486dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a14230f224fb3b7b9b0ced93634313b66ac3abc376dca228bc5802890b0c393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f09ed7c87894bc0473c7015f0e0d55a28f893f5a2e8bf9085c86654c464371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa2391deb934395fe4743249887832731bf7c6ea457d66484bcb6b556e4ab33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64972d043accab95774784cc95dc1144e70c855eff777664c66cb121d519d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:01Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:01 crc kubenswrapper[4700]: I0227 17:03:01.486234 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a2d1b57-7fc1-4614-aac5-7144cdd812d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d2d2c9065b008ff0936c19bcce2ce6a574580911a8188084c5f978e08177b69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"27 17:01:51.956663 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 17:01:51.956671 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 17:01:51.956678 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 17:01:51.956715 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0227 17:01:51.960396 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1772211711\\\\\\\\\\\\\\\" (2026-02-27 17:01:50 +0000 UTC to 2026-03-29 17:01:51 +0000 UTC (now=2026-02-27 17:01:51.960134581 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960704 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1772211711\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1772211711\\\\\\\\\\\\\\\" (2026-02-27 16:01:51 +0000 UTC to 2027-02-27 16:01:51 +0000 UTC (now=2026-02-27 17:01:51.960667784 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960745 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0227 17:01:51.960790 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0227 17:01:51.960837 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960878 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960925 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\"\\\\nI0227 17:01:51.961071 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0227 17:01:51.958239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:01:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:01Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:01 crc kubenswrapper[4700]: I0227 17:03:01.980863 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:03:01 crc kubenswrapper[4700]: I0227 17:03:01.980909 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:03:01 crc kubenswrapper[4700]: E0227 17:03:01.981031 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:03:01 crc kubenswrapper[4700]: I0227 17:03:01.981117 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:03:01 crc kubenswrapper[4700]: E0227 17:03:01.981138 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:03:01 crc kubenswrapper[4700]: E0227 17:03:01.981309 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:03:01 crc kubenswrapper[4700]: I0227 17:03:01.981650 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:03:01 crc kubenswrapper[4700]: E0227 17:03:01.981803 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:03:02 crc kubenswrapper[4700]: I0227 17:03:02.827083 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:03:02 crc kubenswrapper[4700]: I0227 17:03:02.827149 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:03:02 crc kubenswrapper[4700]: I0227 17:03:02.827171 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:03:02 crc kubenswrapper[4700]: I0227 17:03:02.827197 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:03:02 crc kubenswrapper[4700]: I0227 17:03:02.827216 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:03:02Z","lastTransitionTime":"2026-02-27T17:03:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:03:02 crc kubenswrapper[4700]: E0227 17:03:02.848500 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:02Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:02 crc kubenswrapper[4700]: I0227 17:03:02.854187 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:03:02 crc kubenswrapper[4700]: I0227 17:03:02.854246 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:03:02 crc kubenswrapper[4700]: I0227 17:03:02.854263 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:03:02 crc kubenswrapper[4700]: I0227 17:03:02.854287 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:03:02 crc kubenswrapper[4700]: I0227 17:03:02.854304 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:03:02Z","lastTransitionTime":"2026-02-27T17:03:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:03:02 crc kubenswrapper[4700]: E0227 17:03:02.877443 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:02Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:02 crc kubenswrapper[4700]: I0227 17:03:02.882711 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:03:02 crc kubenswrapper[4700]: I0227 17:03:02.882760 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:03:02 crc kubenswrapper[4700]: I0227 17:03:02.882776 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:03:02 crc kubenswrapper[4700]: I0227 17:03:02.882799 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:03:02 crc kubenswrapper[4700]: I0227 17:03:02.882816 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:03:02Z","lastTransitionTime":"2026-02-27T17:03:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:03:02 crc kubenswrapper[4700]: E0227 17:03:02.902314 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:02Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:02 crc kubenswrapper[4700]: I0227 17:03:02.908434 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:03:02 crc kubenswrapper[4700]: I0227 17:03:02.908514 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:03:02 crc kubenswrapper[4700]: I0227 17:03:02.908532 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:03:02 crc kubenswrapper[4700]: I0227 17:03:02.908556 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:03:02 crc kubenswrapper[4700]: I0227 17:03:02.908574 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:03:02Z","lastTransitionTime":"2026-02-27T17:03:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:03:02 crc kubenswrapper[4700]: E0227 17:03:02.927808 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:02Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:02 crc kubenswrapper[4700]: I0227 17:03:02.933626 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:03:02 crc kubenswrapper[4700]: I0227 17:03:02.933692 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:03:02 crc kubenswrapper[4700]: I0227 17:03:02.933712 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:03:02 crc kubenswrapper[4700]: I0227 17:03:02.933736 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:03:02 crc kubenswrapper[4700]: I0227 17:03:02.933757 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:03:02Z","lastTransitionTime":"2026-02-27T17:03:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:03:02 crc kubenswrapper[4700]: E0227 17:03:02.955165 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:02Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:02 crc kubenswrapper[4700]: E0227 17:03:02.955840 4700 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 27 17:03:03 crc kubenswrapper[4700]: I0227 17:03:03.980455 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:03:03 crc kubenswrapper[4700]: I0227 17:03:03.980523 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:03:03 crc kubenswrapper[4700]: I0227 17:03:03.980513 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:03:03 crc kubenswrapper[4700]: I0227 17:03:03.980575 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:03:03 crc kubenswrapper[4700]: E0227 17:03:03.980863 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:03:03 crc kubenswrapper[4700]: E0227 17:03:03.981018 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:03:03 crc kubenswrapper[4700]: E0227 17:03:03.981171 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:03:03 crc kubenswrapper[4700]: E0227 17:03:03.981285 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:03:03 crc kubenswrapper[4700]: I0227 17:03:03.994710 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Feb 27 17:03:05 crc kubenswrapper[4700]: I0227 17:03:05.980924 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:03:05 crc kubenswrapper[4700]: I0227 17:03:05.980955 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:03:05 crc kubenswrapper[4700]: E0227 17:03:05.981063 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:03:05 crc kubenswrapper[4700]: I0227 17:03:05.981102 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:03:05 crc kubenswrapper[4700]: I0227 17:03:05.981152 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:03:05 crc kubenswrapper[4700]: E0227 17:03:05.981236 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:03:05 crc kubenswrapper[4700]: E0227 17:03:05.981377 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:03:05 crc kubenswrapper[4700]: E0227 17:03:05.981615 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:03:06 crc kubenswrapper[4700]: E0227 17:03:06.108021 4700 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 17:03:07 crc kubenswrapper[4700]: I0227 17:03:07.980394 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:03:07 crc kubenswrapper[4700]: E0227 17:03:07.980863 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:03:07 crc kubenswrapper[4700]: I0227 17:03:07.980715 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:03:07 crc kubenswrapper[4700]: E0227 17:03:07.980925 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:03:07 crc kubenswrapper[4700]: I0227 17:03:07.980738 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:03:07 crc kubenswrapper[4700]: E0227 17:03:07.980968 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:03:07 crc kubenswrapper[4700]: I0227 17:03:07.980683 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:03:07 crc kubenswrapper[4700]: E0227 17:03:07.981021 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:03:09 crc kubenswrapper[4700]: I0227 17:03:09.890689 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:03:09 crc kubenswrapper[4700]: I0227 17:03:09.890856 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:03:09 crc kubenswrapper[4700]: E0227 17:03:09.890897 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:13.890857633 +0000 UTC m=+213.876170410 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:03:09 crc kubenswrapper[4700]: I0227 17:03:09.890963 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:03:09 crc kubenswrapper[4700]: E0227 17:03:09.891031 4700 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 17:03:09 crc kubenswrapper[4700]: E0227 17:03:09.891061 4700 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 17:03:09 crc kubenswrapper[4700]: E0227 17:03:09.891081 4700 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 17:03:09 crc kubenswrapper[4700]: I0227 17:03:09.891094 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:03:09 crc kubenswrapper[4700]: I0227 17:03:09.891155 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:03:09 crc kubenswrapper[4700]: E0227 17:03:09.891242 4700 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 17:03:09 crc kubenswrapper[4700]: E0227 17:03:09.891289 4700 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 17:03:09 crc kubenswrapper[4700]: E0227 17:03:09.891290 4700 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 17:03:09 crc kubenswrapper[4700]: E0227 17:03:09.891316 4700 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 17:03:09 crc kubenswrapper[4700]: E0227 17:03:09.891389 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-27 17:04:13.891270164 +0000 UTC m=+213.876582941 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 17:03:09 crc kubenswrapper[4700]: E0227 17:03:09.891430 4700 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 17:03:09 crc kubenswrapper[4700]: E0227 17:03:09.891517 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 17:04:13.891499579 +0000 UTC m=+213.876812366 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 17:03:09 crc kubenswrapper[4700]: E0227 17:03:09.891547 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-27 17:04:13.89153353 +0000 UTC m=+213.876846317 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 17:03:09 crc kubenswrapper[4700]: E0227 17:03:09.891583 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 17:04:13.891565791 +0000 UTC m=+213.876878568 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 17:03:09 crc kubenswrapper[4700]: I0227 17:03:09.981123 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:03:09 crc kubenswrapper[4700]: I0227 17:03:09.981258 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:03:09 crc kubenswrapper[4700]: I0227 17:03:09.981137 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:03:09 crc kubenswrapper[4700]: I0227 17:03:09.981158 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:03:09 crc kubenswrapper[4700]: E0227 17:03:09.981421 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:03:09 crc kubenswrapper[4700]: E0227 17:03:09.981643 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:03:09 crc kubenswrapper[4700]: E0227 17:03:09.981792 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:03:09 crc kubenswrapper[4700]: E0227 17:03:09.981945 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:03:09 crc kubenswrapper[4700]: I0227 17:03:09.992691 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f5288317-4810-4321-baa0-31a118a4dc36-metrics-certs\") pod \"network-metrics-daemon-s8k74\" (UID: \"f5288317-4810-4321-baa0-31a118a4dc36\") " pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:03:09 crc kubenswrapper[4700]: E0227 17:03:09.992904 4700 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 17:03:09 crc kubenswrapper[4700]: E0227 17:03:09.993041 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f5288317-4810-4321-baa0-31a118a4dc36-metrics-certs podName:f5288317-4810-4321-baa0-31a118a4dc36 nodeName:}" failed. No retries permitted until 2026-02-27 17:04:13.993005038 +0000 UTC m=+213.978317965 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f5288317-4810-4321-baa0-31a118a4dc36-metrics-certs") pod "network-metrics-daemon-s8k74" (UID: "f5288317-4810-4321-baa0-31a118a4dc36") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 17:03:10 crc kubenswrapper[4700]: I0227 17:03:10.981876 4700 scope.go:117] "RemoveContainer" containerID="da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49" Feb 27 17:03:10 crc kubenswrapper[4700]: E0227 17:03:10.982134 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jtbqn_openshift-ovn-kubernetes(3a34c0e3-2513-4e81-b6b1-80d1230475fd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" Feb 27 17:03:11 crc kubenswrapper[4700]: I0227 17:03:11.017268 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a34c0e3-2513-4e81-b6b1-80d1230475fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T17:02:58Z\\\",\\\"message\\\":\\\"d (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 17:02:58.943875 7219 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0227 17:02:58.943947 7219 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0227 17:02:58.943980 7219 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0227 17:02:58.944014 7219 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0227 17:02:58.944017 7219 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0227 17:02:58.944044 7219 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0227 17:02:58.944060 7219 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0227 17:02:58.944071 7219 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0227 17:02:58.944079 7219 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0227 17:02:58.944097 7219 handler.go:208] Removed *v1.Node event handler 2\\\\nI0227 17:02:58.944115 7219 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0227 17:02:58.944177 7219 factory.go:656] Stopping watch factory\\\\nI0227 17:02:58.944208 7219 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0227 17:02:58.944209 7219 handler.go:208] Removed *v1.Node event handler 7\\\\nI0227 17:02:58.944107 7219 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jtbqn_openshift-ovn-kubernetes(3a34c0e3-2513-4e81-b6b1-80d1230475fd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jtbqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:11 crc kubenswrapper[4700]: I0227 17:03:11.038043 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ac82a3-64cf-4288-8379-6c98defac6b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb6c59b5f96fb98a2defc2a18bab3d015046d2352a51c1f408bb020c43d5f9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://523f47c5e3ee759e5c1f7a782c10d08a2eb10fba3833c7fe40d6105867420b77\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T17:01:13Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0227 17:00:43.228716 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0227 17:00:43.231145 1 observer_polling.go:159] Starting file observer\\\\nI0227 17:00:43.263774 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0227 17:00:43.269395 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0227 17:01:13.557772 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://555793b646e823bff19401b489c4ab8de44ee3cd8f82ebcd66b5b08e67450be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f172e74f73a1db01fc2f510976b9d061a51b44eec618fbad2e2ac2e0f33f45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4222e543e1fff164630ef807cefe3110095d68d3b8c90fe8dffdcdb5708abceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:11 crc kubenswrapper[4700]: I0227 17:03:11.058634 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a336c2f16e12b411f111f6291366aa1b466f6b405896c152d1667b832a589b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:11 crc kubenswrapper[4700]: I0227 17:03:11.075704 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:11 crc kubenswrapper[4700]: I0227 17:03:11.090893 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b45511-b94c-479f-98db-fd2c4eceec46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bddf1a7b7264d36174b90f1d0d4e49f029781b1acfce88582aa5b2575dcdd09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ef7c37b61de7a039756248f222bd1769b774703c406de5ad11abd9fb2f46dd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f9dq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:11 crc kubenswrapper[4700]: E0227 17:03:11.109608 4700 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 17:03:11 crc kubenswrapper[4700]: I0227 17:03:11.109928 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-s8k74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5288317-4810-4321-baa0-31a118a4dc36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-s8k74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:11 crc kubenswrapper[4700]: I0227 17:03:11.129926 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f425b79-78da-4f9e-a1ad-205a09466052\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3baf8927284b91f3edeec1dc906c2928955badb19062087790d720936b6f7624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd29e6d703945e0f366f2d109b7c35fdd6360c77091b165616e9832dddac1c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cxnv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:11 crc kubenswrapper[4700]: I0227 17:03:11.160787 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ee6013b-8a2d-4651-ae5f-98aaac016d22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af0992b3383953e562bf540bf7659d41e60aeee5d6103105eaab164797486dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a14230f224fb3b7b9b0ced93634313b66ac3abc376dca228bc5802890b0c393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f09ed7c87894bc0473c7015f0e0d55a28f893f5a2e8bf9085c86654c464371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa2391deb934395fe4743249887832731bf7c6ea457d66484bcb6b556e4ab33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64972d043accab95774784cc95dc1144e70c855eff777664c66cb121d519d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:11 crc kubenswrapper[4700]: I0227 17:03:11.182154 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a2d1b57-7fc1-4614-aac5-7144cdd812d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d2d2c9065b008ff0936c19bcce2ce6a574580911a8188084c5f978e08177b69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"27 17:01:51.956663 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 17:01:51.956671 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 17:01:51.956678 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 17:01:51.956715 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0227 17:01:51.960396 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1772211711\\\\\\\\\\\\\\\" (2026-02-27 17:01:50 +0000 UTC to 2026-03-29 17:01:51 +0000 UTC (now=2026-02-27 17:01:51.960134581 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960704 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1772211711\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1772211711\\\\\\\\\\\\\\\" (2026-02-27 16:01:51 +0000 UTC to 2027-02-27 16:01:51 +0000 UTC (now=2026-02-27 17:01:51.960667784 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960745 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0227 17:01:51.960790 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0227 17:01:51.960837 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960878 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960925 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\"\\\\nI0227 17:01:51.961071 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0227 17:01:51.958239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:01:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:11 crc kubenswrapper[4700]: I0227 17:03:11.197677 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-g66q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d4f1eeba71eeb576c1df386f5acaf2bc354e7ca08aa9aac0971f61db72f244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b6cx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-g66q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:11 crc kubenswrapper[4700]: I0227 17:03:11.219147 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rbptz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66eb22fb-e593-40ff-9ff3-5bb32912972d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c675574fe85ad3f55e0431b8d36fd1b83523ee502e675165aafad8551ba4560\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f58f4814976a1e25bef53555d7f5d14fa1cc800c8c5220801960758d5262b552\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T17:02:53Z\\\",\\\"message\\\":\\\"2026-02-27T17:02:07+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7a0028f6-f07c-4b10-a5a8-60005439ad4f\\\\n2026-02-27T17:02:07+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7a0028f6-f07c-4b10-a5a8-60005439ad4f to /host/opt/cni/bin/\\\\n2026-02-27T17:02:08Z [verbose] multus-daemon started\\\\n2026-02-27T17:02:08Z [verbose] Readiness Indicator file check\\\\n2026-02-27T17:02:53Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2dtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rbptz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:11 crc kubenswrapper[4700]: I0227 17:03:11.234054 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"032e16e8-09b3-456f-bebb-09d4b07bcebf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eabca65587272a0e2ce09d13f103fbbe5374fb6d5ed49bfa0e26de8336d9e683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:11 crc kubenswrapper[4700]: I0227 17:03:11.252410 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fcde75bb1f9dd52a2ac77ffaa7d60d8c785c399bbe45a7cbff39d1b926d5541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:11 crc kubenswrapper[4700]: I0227 17:03:11.272483 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c501ee5970cba15e45e1f1c8f0e33768019511f94dbbe509a7bbfea2c5b4a064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f4cd0745c61d33ac717a4a655f00713e2bd935c3cbba23c4ca0913859f6016f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:11 crc kubenswrapper[4700]: I0227 17:03:11.290922 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:11 crc kubenswrapper[4700]: I0227 17:03:11.306595 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dnkln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e772eecd-127f-4b26-8021-010fb507ce0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6df6de510d7e55695db4c89cbe9db84f91aca0e4bfbf2ed852407047c32822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v92s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dnkln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:11 crc kubenswrapper[4700]: I0227 17:03:11.334129 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47cf6a8e-df81-4ccb-992b-abb030862331\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857e486c0ff289c59f3de9933d1e0301be2f25b65e2b3f1fae6c4f149f900a4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8n2wb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:11 crc kubenswrapper[4700]: I0227 17:03:11.353256 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"309d44d0-45a4-4483-b588-7937902ee1fb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9da83072905af757ff9f9cecda0d94b2acfcd847aeeba7ca36ac291c4ff6a3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5a0dc977696434dce7529115ee1d6cce69cbd4c501bbfb77ed0d240d6340d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf6d6f90f1a5acb192ed932732bab2c2a2bebe9f03fddc79c5dfe6423fab012c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b6649f7710d4c1435bfd703120b70e4514f95b4750b247b8ad0a69ed6952a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b6649f7710d4c1435bfd703120b70e4514f95b4750b247b8ad0a69ed6952a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:11 crc kubenswrapper[4700]: I0227 17:03:11.375244 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:11Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:11 crc kubenswrapper[4700]: I0227 17:03:11.980849 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:03:11 crc kubenswrapper[4700]: I0227 17:03:11.980902 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:03:11 crc kubenswrapper[4700]: I0227 17:03:11.980945 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:03:11 crc kubenswrapper[4700]: E0227 17:03:11.981049 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:03:11 crc kubenswrapper[4700]: E0227 17:03:11.981204 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:03:11 crc kubenswrapper[4700]: I0227 17:03:11.981269 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:03:11 crc kubenswrapper[4700]: E0227 17:03:11.981548 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:03:11 crc kubenswrapper[4700]: E0227 17:03:11.981718 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:03:13 crc kubenswrapper[4700]: I0227 17:03:13.011989 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:03:13 crc kubenswrapper[4700]: I0227 17:03:13.012052 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:03:13 crc kubenswrapper[4700]: I0227 17:03:13.012070 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:03:13 crc kubenswrapper[4700]: I0227 17:03:13.012094 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:03:13 crc kubenswrapper[4700]: I0227 17:03:13.012111 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:03:13Z","lastTransitionTime":"2026-02-27T17:03:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:03:13 crc kubenswrapper[4700]: E0227 17:03:13.033364 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:13Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:13 crc kubenswrapper[4700]: I0227 17:03:13.039204 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:03:13 crc kubenswrapper[4700]: I0227 17:03:13.039257 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:03:13 crc kubenswrapper[4700]: I0227 17:03:13.039277 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:03:13 crc kubenswrapper[4700]: I0227 17:03:13.039304 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:03:13 crc kubenswrapper[4700]: I0227 17:03:13.039321 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:03:13Z","lastTransitionTime":"2026-02-27T17:03:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:03:13 crc kubenswrapper[4700]: E0227 17:03:13.059771 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:13Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:13 crc kubenswrapper[4700]: I0227 17:03:13.065098 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:03:13 crc kubenswrapper[4700]: I0227 17:03:13.065150 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:03:13 crc kubenswrapper[4700]: I0227 17:03:13.065166 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:03:13 crc kubenswrapper[4700]: I0227 17:03:13.065190 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:03:13 crc kubenswrapper[4700]: I0227 17:03:13.065206 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:03:13Z","lastTransitionTime":"2026-02-27T17:03:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:03:13 crc kubenswrapper[4700]: E0227 17:03:13.084795 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:13Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:13 crc kubenswrapper[4700]: I0227 17:03:13.089824 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:03:13 crc kubenswrapper[4700]: I0227 17:03:13.089889 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:03:13 crc kubenswrapper[4700]: I0227 17:03:13.089905 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:03:13 crc kubenswrapper[4700]: I0227 17:03:13.089935 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:03:13 crc kubenswrapper[4700]: I0227 17:03:13.089953 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:03:13Z","lastTransitionTime":"2026-02-27T17:03:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:03:13 crc kubenswrapper[4700]: E0227 17:03:13.109788 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:13Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:13 crc kubenswrapper[4700]: I0227 17:03:13.114893 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:03:13 crc kubenswrapper[4700]: I0227 17:03:13.114953 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:03:13 crc kubenswrapper[4700]: I0227 17:03:13.114971 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:03:13 crc kubenswrapper[4700]: I0227 17:03:13.114994 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:03:13 crc kubenswrapper[4700]: I0227 17:03:13.115013 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:03:13Z","lastTransitionTime":"2026-02-27T17:03:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:03:13 crc kubenswrapper[4700]: E0227 17:03:13.134856 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:13Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:13 crc kubenswrapper[4700]: E0227 17:03:13.135076 4700 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 27 17:03:13 crc kubenswrapper[4700]: I0227 17:03:13.980321 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:03:13 crc kubenswrapper[4700]: I0227 17:03:13.980388 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:03:13 crc kubenswrapper[4700]: I0227 17:03:13.980743 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:03:13 crc kubenswrapper[4700]: E0227 17:03:13.980790 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:03:13 crc kubenswrapper[4700]: E0227 17:03:13.980875 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:03:13 crc kubenswrapper[4700]: E0227 17:03:13.980941 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:03:13 crc kubenswrapper[4700]: I0227 17:03:13.981160 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:03:13 crc kubenswrapper[4700]: E0227 17:03:13.981501 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:03:15 crc kubenswrapper[4700]: I0227 17:03:15.980109 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:03:15 crc kubenswrapper[4700]: I0227 17:03:15.980214 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:03:15 crc kubenswrapper[4700]: E0227 17:03:15.980283 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:03:15 crc kubenswrapper[4700]: E0227 17:03:15.980397 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:03:15 crc kubenswrapper[4700]: I0227 17:03:15.980515 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:03:15 crc kubenswrapper[4700]: E0227 17:03:15.980602 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:03:15 crc kubenswrapper[4700]: I0227 17:03:15.980658 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:03:15 crc kubenswrapper[4700]: E0227 17:03:15.980755 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:03:16 crc kubenswrapper[4700]: E0227 17:03:16.111430 4700 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 17:03:17 crc kubenswrapper[4700]: I0227 17:03:17.980025 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:03:17 crc kubenswrapper[4700]: I0227 17:03:17.980126 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:03:17 crc kubenswrapper[4700]: E0227 17:03:17.981502 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:03:17 crc kubenswrapper[4700]: I0227 17:03:17.980211 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:03:17 crc kubenswrapper[4700]: E0227 17:03:17.981685 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:03:17 crc kubenswrapper[4700]: I0227 17:03:17.980177 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:03:17 crc kubenswrapper[4700]: E0227 17:03:17.982019 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:03:17 crc kubenswrapper[4700]: E0227 17:03:17.982272 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:03:19 crc kubenswrapper[4700]: I0227 17:03:19.981017 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:03:19 crc kubenswrapper[4700]: I0227 17:03:19.981120 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:03:19 crc kubenswrapper[4700]: E0227 17:03:19.981224 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:03:19 crc kubenswrapper[4700]: E0227 17:03:19.981320 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:03:19 crc kubenswrapper[4700]: I0227 17:03:19.981663 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:03:19 crc kubenswrapper[4700]: E0227 17:03:19.981743 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:03:19 crc kubenswrapper[4700]: I0227 17:03:19.981786 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:03:19 crc kubenswrapper[4700]: E0227 17:03:19.981943 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:03:20 crc kubenswrapper[4700]: I0227 17:03:20.998077 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f425b79-78da-4f9e-a1ad-205a09466052\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3baf8927284b91f3edeec1dc906c2928955badb19062087790d720936b6f7624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd29e6d703945e0f366f2d109b7c35fdd6360c77091b165616e9832dddac1c05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dt85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cxnv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:20Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:21 crc kubenswrapper[4700]: I0227 17:03:21.030385 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ee6013b-8a2d-4651-ae5f-98aaac016d22\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9af0992b3383953e562bf540bf7659d41e60aeee5d6103105eaab164797486dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a14230f224fb3b7b9b0ced93634313b66ac3abc376dca228bc5802890b0c393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66f09ed7c87894bc0473c7015f0e0d55a28f893f5a2e8bf9085c86654c464371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fa2391deb934395fe4743249887832731bf7c6ea457d66484bcb6b556e4ab33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e64972d043accab95774784cc95dc1144e70c855eff777664c66cb121d519d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70333399155d21d27ea6b690eea176ffcd36cbb6f4bc9ba81c58069665a83c30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5814b2035d53b8dc273f51633a1971f70e094ecbb65a0e0d8dd3d9e672211f8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bda0cdcda79639566e243da018e743db671960e11f570e8c0d0fa6c76b8edef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:21Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:21 crc kubenswrapper[4700]: I0227 17:03:21.053147 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a2d1b57-7fc1-4614-aac5-7144cdd812d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d2d2c9065b008ff0936c19bcce2ce6a574580911a8188084c5f978e08177b69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T17:01:51Z\\\",\\\"message\\\":\\\"27 17:01:51.956663 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0227 17:01:51.956671 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0227 17:01:51.956678 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0227 17:01:51.956715 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI0227 17:01:51.960396 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1772211711\\\\\\\\\\\\\\\" (2026-02-27 17:01:50 +0000 UTC to 2026-03-29 17:01:51 +0000 UTC (now=2026-02-27 17:01:51.960134581 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960704 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1772211711\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1772211711\\\\\\\\\\\\\\\" (2026-02-27 16:01:51 +0000 UTC to 2027-02-27 16:01:51 +0000 UTC (now=2026-02-27 17:01:51.960667784 +0000 UTC))\\\\\\\"\\\\nI0227 17:01:51.960745 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0227 17:01:51.960790 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0227 17:01:51.960837 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960878 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0227 17:01:51.960925 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3925473695/tls.crt::/tmp/serving-cert-3925473695/tls.key\\\\\\\"\\\\nI0227 17:01:51.961071 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0227 17:01:51.958239 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:01:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:21Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:21 crc kubenswrapper[4700]: I0227 17:03:21.071989 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:21Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:21 crc kubenswrapper[4700]: I0227 17:03:21.088656 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b45511-b94c-479f-98db-fd2c4eceec46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bddf1a7b7264d36174b90f1d0d4e49f029781b1acfce88582aa5b2575dcdd09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ef7c37b61de7a039756248f222bd1769b774703c406de5ad11abd9fb2f46dd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqdf7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-f9dq7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:21Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:21 crc kubenswrapper[4700]: I0227 17:03:21.104564 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-s8k74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5288317-4810-4321-baa0-31a118a4dc36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5wt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-s8k74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:21Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:21 crc kubenswrapper[4700]: E0227 17:03:21.112513 4700 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 17:03:21 crc kubenswrapper[4700]: I0227 17:03:21.124766 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"032e16e8-09b3-456f-bebb-09d4b07bcebf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eabca65587272a0e2ce09d13f103fbbe5374fb6d5ed49bfa0e26de8336d9e683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3d8c8f84ab445906d3d11891732b92334f80b6380ae98895060b689729a09f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:21Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:21 crc kubenswrapper[4700]: I0227 17:03:21.141655 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fcde75bb1f9dd52a2ac77ffaa7d60d8c785c399bbe45a7cbff39d1b926d5541\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:21Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:21 crc kubenswrapper[4700]: I0227 17:03:21.155658 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-g66q8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2472dab1-91b8-4be5-bd0c-dbe6522f5b94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55d4f1eeba71eeb576c1df386f5acaf2bc354e7ca08aa9aac0971f61db72f244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b6cx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-g66q8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:21Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:21 crc kubenswrapper[4700]: I0227 17:03:21.175240 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rbptz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66eb22fb-e593-40ff-9ff3-5bb32912972d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c675574fe85ad3f55e0431b8d36fd1b83523ee502e675165aafad8551ba4560\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f58f4814976a1e25bef53555d7f5d14fa1cc800c8c5220801960758d5262b552\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T17:02:53Z\\\",\\\"message\\\":\\\"2026-02-27T17:02:07+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7a0028f6-f07c-4b10-a5a8-60005439ad4f\\\\n2026-02-27T17:02:07+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7a0028f6-f07c-4b10-a5a8-60005439ad4f to /host/opt/cni/bin/\\\\n2026-02-27T17:02:08Z [verbose] multus-daemon started\\\\n2026-02-27T17:02:08Z [verbose] Readiness Indicator file check\\\\n2026-02-27T17:02:53Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v2dtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rbptz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:21Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:21 crc kubenswrapper[4700]: I0227 17:03:21.196242 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47cf6a8e-df81-4ccb-992b-abb030862331\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857e486c0ff289c59f3de9933d1e0301be2f25b65e2b3f1fae6c4f149f900a4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d07c3e8f729d22b188ae56a4829487ed21e5c5c5df03570c4346792a4a349919\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45f7eab7c943abc2f0341937715ddbad8795ad55d9dd343de49724c53e4217b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edb82079b8a125ae116110a15bb46367ae0a4f61a176a4b01b67a06e2973bba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c30f73ddf3775a7cdc6fa88bcdc16a4ccd03105bc577c156560b6025aa1545af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe43f09ba1c69e24b1fd6bf42a698eef011a7f8a9f2d5f0078f73d3defcb1942\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99f28282ddb7c47b53f83c9d769b53a662fc2302f42987aed58452ab7e758209\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8d8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8n2wb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:21Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:21 crc kubenswrapper[4700]: I0227 17:03:21.213677 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"309d44d0-45a4-4483-b588-7937902ee1fb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9da83072905af757ff9f9cecda0d94b2acfcd847aeeba7ca36ac291c4ff6a3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5a0dc977696434dce7529115ee1d6cce69cbd4c501bbfb77ed0d240d6340d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf6d6f90f1a5acb192ed932732bab2c2a2bebe9f03fddc79c5dfe6423fab012c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b6649f7710d4c1435bfd703120b70e4514f95b4750b247b8ad0a69ed6952a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b6649f7710d4c1435bfd703120b70e4514f95b4750b247b8ad0a69ed6952a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:00:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:21Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:21 crc kubenswrapper[4700]: I0227 17:03:21.233791 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:21Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:21 crc kubenswrapper[4700]: I0227 17:03:21.254297 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c501ee5970cba15e45e1f1c8f0e33768019511f94dbbe509a7bbfea2c5b4a064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f4cd0745c61d33ac717a4a655f00713e2bd935c3cbba23c4ca0913859f6016f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:21Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:21 crc kubenswrapper[4700]: I0227 17:03:21.273614 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:21Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:21 crc kubenswrapper[4700]: I0227 17:03:21.291898 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dnkln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e772eecd-127f-4b26-8021-010fb507ce0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6df6de510d7e55695db4c89cbe9db84f91aca0e4bfbf2ed852407047c32822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v92s4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dnkln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:21Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:21 crc kubenswrapper[4700]: I0227 17:03:21.311414 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ac82a3-64cf-4288-8379-6c98defac6b7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:01:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:00:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb6c59b5f96fb98a2defc2a18bab3d015046d2352a51c1f408bb020c43d5f9f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://523f47c5e3ee759e5c1f7a782c10d08a2eb10fba3833c7fe40d6105867420b77\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-27T17:01:13Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0227 17:00:43.228716 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0227 17:00:43.231145 1 observer_polling.go:159] Starting file observer\\\\nI0227 17:00:43.263774 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0227 17:00:43.269395 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0227 17:01:13.557772 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://555793b646e823bff19401b489c4ab8de44ee3cd8f82ebcd66b5b08e67450be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f172e74f73a1db01fc2f510976b9d061a51b44eec618fbad2e2ac2e0f33f45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4222e543e1fff164630ef807cefe3110095d68d3b8c90fe8dffdcdb5708abceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:00:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:00:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:21Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:21 crc kubenswrapper[4700]: I0227 17:03:21.348051 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a336c2f16e12b411f111f6291366aa1b466f6b405896c152d1667b832a589b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:21Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:21 crc kubenswrapper[4700]: I0227 17:03:21.381351 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a34c0e3-2513-4e81-b6b1-80d1230475fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-27T17:02:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-27T17:02:58Z\\\",\\\"message\\\":\\\"d (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0227 17:02:58.943875 7219 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0227 17:02:58.943947 7219 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0227 17:02:58.943980 7219 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0227 17:02:58.944014 7219 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0227 17:02:58.944017 7219 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0227 17:02:58.944044 7219 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0227 17:02:58.944060 7219 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0227 17:02:58.944071 7219 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0227 17:02:58.944079 7219 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0227 17:02:58.944097 7219 handler.go:208] Removed *v1.Node event handler 2\\\\nI0227 17:02:58.944115 7219 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0227 17:02:58.944177 7219 factory.go:656] Stopping watch factory\\\\nI0227 17:02:58.944208 7219 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0227 17:02:58.944209 7219 handler.go:208] Removed *v1.Node event handler 7\\\\nI0227 17:02:58.944107 7219 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jtbqn_openshift-ovn-kubernetes(3a34c0e3-2513-4e81-b6b1-80d1230475fd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:02:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:02:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-748j7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-27T17:02:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jtbqn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:21Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:21 crc kubenswrapper[4700]: I0227 17:03:21.980740 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:03:21 crc kubenswrapper[4700]: I0227 17:03:21.980747 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:03:21 crc kubenswrapper[4700]: E0227 17:03:21.981491 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:03:21 crc kubenswrapper[4700]: I0227 17:03:21.980889 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:03:21 crc kubenswrapper[4700]: E0227 17:03:21.981583 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:03:21 crc kubenswrapper[4700]: E0227 17:03:21.981680 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:03:21 crc kubenswrapper[4700]: I0227 17:03:21.980746 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:03:21 crc kubenswrapper[4700]: E0227 17:03:21.982381 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:03:23 crc kubenswrapper[4700]: I0227 17:03:23.331014 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:03:23 crc kubenswrapper[4700]: I0227 17:03:23.331106 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:03:23 crc kubenswrapper[4700]: I0227 17:03:23.331125 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:03:23 crc kubenswrapper[4700]: I0227 17:03:23.331149 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:03:23 crc kubenswrapper[4700]: I0227 17:03:23.331165 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:03:23Z","lastTransitionTime":"2026-02-27T17:03:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:03:23 crc kubenswrapper[4700]: E0227 17:03:23.352095 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:23Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:23 crc kubenswrapper[4700]: I0227 17:03:23.357686 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:03:23 crc kubenswrapper[4700]: I0227 17:03:23.357755 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:03:23 crc kubenswrapper[4700]: I0227 17:03:23.357774 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:03:23 crc kubenswrapper[4700]: I0227 17:03:23.358216 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:03:23 crc kubenswrapper[4700]: I0227 17:03:23.358274 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:03:23Z","lastTransitionTime":"2026-02-27T17:03:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:03:23 crc kubenswrapper[4700]: E0227 17:03:23.378762 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:23Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:23 crc kubenswrapper[4700]: I0227 17:03:23.384283 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:03:23 crc kubenswrapper[4700]: I0227 17:03:23.384415 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:03:23 crc kubenswrapper[4700]: I0227 17:03:23.384439 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:03:23 crc kubenswrapper[4700]: I0227 17:03:23.384486 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:03:23 crc kubenswrapper[4700]: I0227 17:03:23.384507 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:03:23Z","lastTransitionTime":"2026-02-27T17:03:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:03:23 crc kubenswrapper[4700]: E0227 17:03:23.405090 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:23Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:23 crc kubenswrapper[4700]: I0227 17:03:23.410265 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:03:23 crc kubenswrapper[4700]: I0227 17:03:23.410367 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:03:23 crc kubenswrapper[4700]: I0227 17:03:23.410382 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:03:23 crc kubenswrapper[4700]: I0227 17:03:23.410401 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:03:23 crc kubenswrapper[4700]: I0227 17:03:23.410413 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:03:23Z","lastTransitionTime":"2026-02-27T17:03:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:03:23 crc kubenswrapper[4700]: E0227 17:03:23.431618 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:23Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:23 crc kubenswrapper[4700]: I0227 17:03:23.436191 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:03:23 crc kubenswrapper[4700]: I0227 17:03:23.436237 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:03:23 crc kubenswrapper[4700]: I0227 17:03:23.436255 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:03:23 crc kubenswrapper[4700]: I0227 17:03:23.436281 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:03:23 crc kubenswrapper[4700]: I0227 17:03:23.436300 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:03:23Z","lastTransitionTime":"2026-02-27T17:03:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:03:23 crc kubenswrapper[4700]: E0227 17:03:23.457195 4700 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T17:03:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"aa2cf3e6-9e57-4694-81ae-b6e008a63c28\\\",\\\"systemUUID\\\":\\\"c03ccb1e-6d2c-4062-a2f2-4418103b1627\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T17:03:23Z is after 2025-08-24T17:21:41Z" Feb 27 17:03:23 crc kubenswrapper[4700]: E0227 17:03:23.457505 4700 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 27 17:03:23 crc kubenswrapper[4700]: I0227 17:03:23.980792 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:03:23 crc kubenswrapper[4700]: I0227 17:03:23.980818 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:03:23 crc kubenswrapper[4700]: I0227 17:03:23.980946 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:03:23 crc kubenswrapper[4700]: E0227 17:03:23.981154 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:03:23 crc kubenswrapper[4700]: I0227 17:03:23.981230 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:03:23 crc kubenswrapper[4700]: E0227 17:03:23.981373 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:03:23 crc kubenswrapper[4700]: E0227 17:03:23.981562 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:03:23 crc kubenswrapper[4700]: E0227 17:03:23.981769 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:03:24 crc kubenswrapper[4700]: I0227 17:03:24.981642 4700 scope.go:117] "RemoveContainer" containerID="da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49" Feb 27 17:03:24 crc kubenswrapper[4700]: E0227 17:03:24.981941 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jtbqn_openshift-ovn-kubernetes(3a34c0e3-2513-4e81-b6b1-80d1230475fd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" Feb 27 17:03:25 crc kubenswrapper[4700]: I0227 17:03:25.980312 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:03:25 crc kubenswrapper[4700]: I0227 17:03:25.980378 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:03:25 crc kubenswrapper[4700]: E0227 17:03:25.980713 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:03:25 crc kubenswrapper[4700]: I0227 17:03:25.980776 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:03:25 crc kubenswrapper[4700]: I0227 17:03:25.980788 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:03:25 crc kubenswrapper[4700]: E0227 17:03:25.980896 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:03:25 crc kubenswrapper[4700]: E0227 17:03:25.981031 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:03:25 crc kubenswrapper[4700]: E0227 17:03:25.981187 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:03:26 crc kubenswrapper[4700]: E0227 17:03:26.114096 4700 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 17:03:27 crc kubenswrapper[4700]: I0227 17:03:27.981167 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:03:27 crc kubenswrapper[4700]: I0227 17:03:27.981212 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:03:27 crc kubenswrapper[4700]: I0227 17:03:27.981217 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:03:27 crc kubenswrapper[4700]: I0227 17:03:27.981277 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:03:27 crc kubenswrapper[4700]: E0227 17:03:27.981442 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:03:27 crc kubenswrapper[4700]: E0227 17:03:27.981618 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:03:27 crc kubenswrapper[4700]: E0227 17:03:27.981713 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:03:27 crc kubenswrapper[4700]: E0227 17:03:27.981804 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:03:29 crc kubenswrapper[4700]: I0227 17:03:29.980062 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:03:29 crc kubenswrapper[4700]: I0227 17:03:29.980186 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:03:29 crc kubenswrapper[4700]: I0227 17:03:29.980230 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:03:29 crc kubenswrapper[4700]: I0227 17:03:29.980089 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:03:29 crc kubenswrapper[4700]: E0227 17:03:29.980511 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:03:29 crc kubenswrapper[4700]: E0227 17:03:29.980692 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:03:29 crc kubenswrapper[4700]: E0227 17:03:29.980830 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:03:29 crc kubenswrapper[4700]: E0227 17:03:29.981014 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:03:31 crc kubenswrapper[4700]: I0227 17:03:31.089824 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-dnkln" podStartSLOduration=117.089790611 podStartE2EDuration="1m57.089790611s" podCreationTimestamp="2026-02-27 17:01:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:03:31.060510453 +0000 UTC m=+171.045823240" watchObservedRunningTime="2026-02-27 17:03:31.089790611 +0000 UTC m=+171.075103388" Feb 27 17:03:31 crc kubenswrapper[4700]: E0227 17:03:31.114893 4700 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 17:03:31 crc kubenswrapper[4700]: I0227 17:03:31.115431 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=47.115408665 podStartE2EDuration="47.115408665s" podCreationTimestamp="2026-02-27 17:02:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:03:31.111968616 +0000 UTC m=+171.097281443" watchObservedRunningTime="2026-02-27 17:03:31.115408665 +0000 UTC m=+171.100721422" Feb 27 17:03:31 crc kubenswrapper[4700]: I0227 17:03:31.115601 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-8n2wb" podStartSLOduration=117.11559194 podStartE2EDuration="1m57.11559194s" podCreationTimestamp="2026-02-27 17:01:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:03:31.09012633 +0000 UTC m=+171.075439117" watchObservedRunningTime="2026-02-27 17:03:31.11559194 +0000 UTC m=+171.100904717" Feb 27 17:03:31 crc kubenswrapper[4700]: I0227 17:03:31.204197 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=28.204172334 podStartE2EDuration="28.204172334s" podCreationTimestamp="2026-02-27 17:03:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:03:31.202729036 +0000 UTC m=+171.188041873" watchObservedRunningTime="2026-02-27 17:03:31.204172334 +0000 UTC m=+171.189485121" Feb 27 17:03:31 crc kubenswrapper[4700]: I0227 17:03:31.269721 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podStartSLOduration=117.269692551 podStartE2EDuration="1m57.269692551s" podCreationTimestamp="2026-02-27 17:01:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:03:31.268002577 +0000 UTC m=+171.253315404" watchObservedRunningTime="2026-02-27 17:03:31.269692551 +0000 UTC m=+171.255005348" Feb 27 17:03:31 crc kubenswrapper[4700]: I0227 17:03:31.300232 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cxnv6" podStartSLOduration=116.300204431 podStartE2EDuration="1m56.300204431s" podCreationTimestamp="2026-02-27 17:01:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:03:31.299585365 +0000 UTC m=+171.284898192" watchObservedRunningTime="2026-02-27 17:03:31.300204431 +0000 UTC m=+171.285517218" Feb 27 17:03:31 crc kubenswrapper[4700]: I0227 17:03:31.328286 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=81.328253618 podStartE2EDuration="1m21.328253618s" podCreationTimestamp="2026-02-27 17:02:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:03:31.326520773 +0000 UTC m=+171.311833530" watchObservedRunningTime="2026-02-27 17:03:31.328253618 +0000 UTC m=+171.313566405" Feb 27 17:03:31 crc kubenswrapper[4700]: I0227 17:03:31.354237 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=84.35421168 podStartE2EDuration="1m24.35421168s" podCreationTimestamp="2026-02-27 17:02:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:03:31.354016855 +0000 UTC m=+171.339329612" watchObservedRunningTime="2026-02-27 17:03:31.35421168 +0000 UTC m=+171.339524467" Feb 27 17:03:31 crc kubenswrapper[4700]: I0227 17:03:31.366943 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-g66q8" podStartSLOduration=117.366919629 podStartE2EDuration="1m57.366919629s" podCreationTimestamp="2026-02-27 17:01:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:03:31.366722834 +0000 UTC m=+171.352035601" watchObservedRunningTime="2026-02-27 17:03:31.366919629 +0000 UTC m=+171.352232416" Feb 27 17:03:31 crc kubenswrapper[4700]: I0227 17:03:31.400675 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-rbptz" podStartSLOduration=117.400647332 podStartE2EDuration="1m57.400647332s" podCreationTimestamp="2026-02-27 17:01:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:03:31.399895433 +0000 UTC m=+171.385208240" watchObservedRunningTime="2026-02-27 17:03:31.400647332 +0000 UTC m=+171.385960109" Feb 27 17:03:31 crc kubenswrapper[4700]: I0227 17:03:31.414984 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=79.414957613 podStartE2EDuration="1m19.414957613s" podCreationTimestamp="2026-02-27 17:02:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:03:31.4132889 +0000 UTC m=+171.398601677" watchObservedRunningTime="2026-02-27 17:03:31.414957613 +0000 UTC m=+171.400270420" Feb 27 17:03:31 crc kubenswrapper[4700]: I0227 17:03:31.980723 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:03:31 crc kubenswrapper[4700]: I0227 17:03:31.980788 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:03:31 crc kubenswrapper[4700]: I0227 17:03:31.980754 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:03:31 crc kubenswrapper[4700]: I0227 17:03:31.980847 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:03:31 crc kubenswrapper[4700]: E0227 17:03:31.981265 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:03:31 crc kubenswrapper[4700]: E0227 17:03:31.981397 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:03:31 crc kubenswrapper[4700]: E0227 17:03:31.981529 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:03:31 crc kubenswrapper[4700]: E0227 17:03:31.981622 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:03:33 crc kubenswrapper[4700]: I0227 17:03:33.681881 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 17:03:33 crc kubenswrapper[4700]: I0227 17:03:33.681954 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 17:03:33 crc kubenswrapper[4700]: I0227 17:03:33.681974 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 17:03:33 crc kubenswrapper[4700]: I0227 17:03:33.682004 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 17:03:33 crc kubenswrapper[4700]: I0227 17:03:33.682024 4700 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T17:03:33Z","lastTransitionTime":"2026-02-27T17:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 17:03:33 crc kubenswrapper[4700]: I0227 17:03:33.751563 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9b5p"] Feb 27 17:03:33 crc kubenswrapper[4700]: I0227 17:03:33.752521 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9b5p" Feb 27 17:03:33 crc kubenswrapper[4700]: I0227 17:03:33.759149 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 27 17:03:33 crc kubenswrapper[4700]: I0227 17:03:33.760812 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 27 17:03:33 crc kubenswrapper[4700]: I0227 17:03:33.760936 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 27 17:03:33 crc kubenswrapper[4700]: I0227 17:03:33.766106 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 27 17:03:33 crc kubenswrapper[4700]: I0227 17:03:33.868509 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c93f3302-6af1-46e7-9405-e9a376636a47-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-b9b5p\" (UID: \"c93f3302-6af1-46e7-9405-e9a376636a47\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9b5p" Feb 27 17:03:33 crc kubenswrapper[4700]: I0227 17:03:33.868655 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c93f3302-6af1-46e7-9405-e9a376636a47-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-b9b5p\" (UID: \"c93f3302-6af1-46e7-9405-e9a376636a47\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9b5p" Feb 27 17:03:33 crc kubenswrapper[4700]: I0227 17:03:33.868733 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c93f3302-6af1-46e7-9405-e9a376636a47-service-ca\") pod \"cluster-version-operator-5c965bbfc6-b9b5p\" (UID: \"c93f3302-6af1-46e7-9405-e9a376636a47\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9b5p" Feb 27 17:03:33 crc kubenswrapper[4700]: I0227 17:03:33.868845 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c93f3302-6af1-46e7-9405-e9a376636a47-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-b9b5p\" (UID: \"c93f3302-6af1-46e7-9405-e9a376636a47\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9b5p" Feb 27 17:03:33 crc kubenswrapper[4700]: I0227 17:03:33.868887 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c93f3302-6af1-46e7-9405-e9a376636a47-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-b9b5p\" (UID: \"c93f3302-6af1-46e7-9405-e9a376636a47\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9b5p" Feb 27 17:03:33 crc kubenswrapper[4700]: I0227 17:03:33.970126 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c93f3302-6af1-46e7-9405-e9a376636a47-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-b9b5p\" (UID: \"c93f3302-6af1-46e7-9405-e9a376636a47\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9b5p" Feb 27 17:03:33 crc kubenswrapper[4700]: I0227 17:03:33.970004 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c93f3302-6af1-46e7-9405-e9a376636a47-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-b9b5p\" (UID: \"c93f3302-6af1-46e7-9405-e9a376636a47\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9b5p" Feb 27 17:03:33 crc kubenswrapper[4700]: I0227 17:03:33.970320 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c93f3302-6af1-46e7-9405-e9a376636a47-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-b9b5p\" (UID: \"c93f3302-6af1-46e7-9405-e9a376636a47\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9b5p" Feb 27 17:03:33 crc kubenswrapper[4700]: I0227 17:03:33.971724 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c93f3302-6af1-46e7-9405-e9a376636a47-service-ca\") pod \"cluster-version-operator-5c965bbfc6-b9b5p\" (UID: \"c93f3302-6af1-46e7-9405-e9a376636a47\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9b5p" Feb 27 17:03:33 crc kubenswrapper[4700]: I0227 17:03:33.971868 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c93f3302-6af1-46e7-9405-e9a376636a47-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-b9b5p\" (UID: \"c93f3302-6af1-46e7-9405-e9a376636a47\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9b5p" Feb 27 17:03:33 crc kubenswrapper[4700]: I0227 17:03:33.971934 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c93f3302-6af1-46e7-9405-e9a376636a47-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-b9b5p\" (UID: \"c93f3302-6af1-46e7-9405-e9a376636a47\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9b5p" Feb 27 17:03:33 crc kubenswrapper[4700]: I0227 17:03:33.972065 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c93f3302-6af1-46e7-9405-e9a376636a47-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-b9b5p\" (UID: \"c93f3302-6af1-46e7-9405-e9a376636a47\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9b5p" Feb 27 17:03:33 crc kubenswrapper[4700]: I0227 17:03:33.973557 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c93f3302-6af1-46e7-9405-e9a376636a47-service-ca\") pod \"cluster-version-operator-5c965bbfc6-b9b5p\" (UID: \"c93f3302-6af1-46e7-9405-e9a376636a47\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9b5p" Feb 27 17:03:33 crc kubenswrapper[4700]: I0227 17:03:33.980958 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:03:33 crc kubenswrapper[4700]: I0227 17:03:33.980974 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:03:33 crc kubenswrapper[4700]: I0227 17:03:33.981084 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:03:33 crc kubenswrapper[4700]: E0227 17:03:33.981593 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:03:33 crc kubenswrapper[4700]: I0227 17:03:33.981085 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:03:33 crc kubenswrapper[4700]: E0227 17:03:33.981741 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:03:33 crc kubenswrapper[4700]: E0227 17:03:33.981850 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:03:33 crc kubenswrapper[4700]: E0227 17:03:33.981362 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:03:33 crc kubenswrapper[4700]: I0227 17:03:33.982846 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c93f3302-6af1-46e7-9405-e9a376636a47-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-b9b5p\" (UID: \"c93f3302-6af1-46e7-9405-e9a376636a47\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9b5p" Feb 27 17:03:34 crc kubenswrapper[4700]: I0227 17:03:34.003757 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c93f3302-6af1-46e7-9405-e9a376636a47-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-b9b5p\" (UID: \"c93f3302-6af1-46e7-9405-e9a376636a47\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9b5p" Feb 27 17:03:34 crc kubenswrapper[4700]: I0227 17:03:34.012880 4700 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Feb 27 17:03:34 crc kubenswrapper[4700]: I0227 17:03:34.027622 4700 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 27 17:03:34 crc kubenswrapper[4700]: I0227 17:03:34.082455 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9b5p" Feb 27 17:03:34 crc kubenswrapper[4700]: I0227 17:03:34.904823 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9b5p" event={"ID":"c93f3302-6af1-46e7-9405-e9a376636a47","Type":"ContainerStarted","Data":"82fb3590d1581da2beca27cb5b0b2b09c0f8c98546bb74b1995f71de2a592e6a"} Feb 27 17:03:34 crc kubenswrapper[4700]: I0227 17:03:34.904883 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9b5p" event={"ID":"c93f3302-6af1-46e7-9405-e9a376636a47","Type":"ContainerStarted","Data":"4b3b418d025e93e1600ab6521a6b0505d92ba76ae01e1c9a447ee84d08bd8266"} Feb 27 17:03:34 crc kubenswrapper[4700]: I0227 17:03:34.919045 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9b5p" podStartSLOduration=120.919029876 podStartE2EDuration="2m0.919029876s" podCreationTimestamp="2026-02-27 17:01:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:03:34.918812281 +0000 UTC m=+174.904125038" watchObservedRunningTime="2026-02-27 17:03:34.919029876 +0000 UTC m=+174.904342623" Feb 27 17:03:35 crc kubenswrapper[4700]: I0227 17:03:35.980789 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:03:35 crc kubenswrapper[4700]: I0227 17:03:35.980819 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:03:35 crc kubenswrapper[4700]: E0227 17:03:35.981440 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:03:35 crc kubenswrapper[4700]: I0227 17:03:35.980915 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:03:35 crc kubenswrapper[4700]: E0227 17:03:35.981602 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:03:35 crc kubenswrapper[4700]: E0227 17:03:35.981250 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:03:35 crc kubenswrapper[4700]: I0227 17:03:35.980858 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:03:35 crc kubenswrapper[4700]: E0227 17:03:35.981729 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:03:36 crc kubenswrapper[4700]: E0227 17:03:36.115926 4700 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 17:03:37 crc kubenswrapper[4700]: I0227 17:03:37.980966 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:03:37 crc kubenswrapper[4700]: I0227 17:03:37.981026 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:03:37 crc kubenswrapper[4700]: E0227 17:03:37.981139 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:03:37 crc kubenswrapper[4700]: I0227 17:03:37.980991 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:03:37 crc kubenswrapper[4700]: I0227 17:03:37.981203 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:03:37 crc kubenswrapper[4700]: E0227 17:03:37.981329 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:03:37 crc kubenswrapper[4700]: E0227 17:03:37.981618 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:03:37 crc kubenswrapper[4700]: E0227 17:03:37.981740 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:03:39 crc kubenswrapper[4700]: I0227 17:03:39.922775 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-rbptz_66eb22fb-e593-40ff-9ff3-5bb32912972d/kube-multus/1.log" Feb 27 17:03:39 crc kubenswrapper[4700]: I0227 17:03:39.924496 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-rbptz_66eb22fb-e593-40ff-9ff3-5bb32912972d/kube-multus/0.log" Feb 27 17:03:39 crc kubenswrapper[4700]: I0227 17:03:39.924571 4700 generic.go:334] "Generic (PLEG): container finished" podID="66eb22fb-e593-40ff-9ff3-5bb32912972d" containerID="9c675574fe85ad3f55e0431b8d36fd1b83523ee502e675165aafad8551ba4560" exitCode=1 Feb 27 17:03:39 crc kubenswrapper[4700]: I0227 17:03:39.924613 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-rbptz" event={"ID":"66eb22fb-e593-40ff-9ff3-5bb32912972d","Type":"ContainerDied","Data":"9c675574fe85ad3f55e0431b8d36fd1b83523ee502e675165aafad8551ba4560"} Feb 27 17:03:39 crc kubenswrapper[4700]: I0227 17:03:39.924662 4700 scope.go:117] "RemoveContainer" containerID="f58f4814976a1e25bef53555d7f5d14fa1cc800c8c5220801960758d5262b552" Feb 27 17:03:39 crc kubenswrapper[4700]: I0227 17:03:39.925389 4700 scope.go:117] "RemoveContainer" containerID="9c675574fe85ad3f55e0431b8d36fd1b83523ee502e675165aafad8551ba4560" Feb 27 17:03:39 crc kubenswrapper[4700]: E0227 17:03:39.925744 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-rbptz_openshift-multus(66eb22fb-e593-40ff-9ff3-5bb32912972d)\"" pod="openshift-multus/multus-rbptz" podUID="66eb22fb-e593-40ff-9ff3-5bb32912972d" Feb 27 17:03:39 crc kubenswrapper[4700]: I0227 17:03:39.980348 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:03:39 crc kubenswrapper[4700]: I0227 17:03:39.980378 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:03:39 crc kubenswrapper[4700]: I0227 17:03:39.980528 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:03:39 crc kubenswrapper[4700]: I0227 17:03:39.980348 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:03:39 crc kubenswrapper[4700]: E0227 17:03:39.980558 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:03:39 crc kubenswrapper[4700]: E0227 17:03:39.980689 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:03:39 crc kubenswrapper[4700]: E0227 17:03:39.980791 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:03:39 crc kubenswrapper[4700]: E0227 17:03:39.980953 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:03:39 crc kubenswrapper[4700]: I0227 17:03:39.982279 4700 scope.go:117] "RemoveContainer" containerID="da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49" Feb 27 17:03:40 crc kubenswrapper[4700]: I0227 17:03:40.932345 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jtbqn_3a34c0e3-2513-4e81-b6b1-80d1230475fd/ovnkube-controller/3.log" Feb 27 17:03:40 crc kubenswrapper[4700]: I0227 17:03:40.936208 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" event={"ID":"3a34c0e3-2513-4e81-b6b1-80d1230475fd","Type":"ContainerStarted","Data":"e2e83f8a1f5d3909754673dab13333df53c0524f2c5c6b33831c04129dc65302"} Feb 27 17:03:40 crc kubenswrapper[4700]: I0227 17:03:40.937672 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-rbptz_66eb22fb-e593-40ff-9ff3-5bb32912972d/kube-multus/1.log" Feb 27 17:03:40 crc kubenswrapper[4700]: I0227 17:03:40.973438 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" podStartSLOduration=125.973406409 podStartE2EDuration="2m5.973406409s" podCreationTimestamp="2026-02-27 17:01:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:03:40.972582488 +0000 UTC m=+180.957895265" watchObservedRunningTime="2026-02-27 17:03:40.973406409 +0000 UTC m=+180.958719196" Feb 27 17:03:41 crc kubenswrapper[4700]: I0227 17:03:41.040264 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-s8k74"] Feb 27 17:03:41 crc kubenswrapper[4700]: I0227 17:03:41.040374 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:03:41 crc kubenswrapper[4700]: E0227 17:03:41.040496 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:03:41 crc kubenswrapper[4700]: E0227 17:03:41.117082 4700 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 17:03:41 crc kubenswrapper[4700]: I0227 17:03:41.980967 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:03:41 crc kubenswrapper[4700]: I0227 17:03:41.981077 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:03:41 crc kubenswrapper[4700]: E0227 17:03:41.981589 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:03:41 crc kubenswrapper[4700]: I0227 17:03:41.981240 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:03:41 crc kubenswrapper[4700]: E0227 17:03:41.981854 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:03:41 crc kubenswrapper[4700]: E0227 17:03:41.982006 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:03:42 crc kubenswrapper[4700]: I0227 17:03:42.980240 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:03:42 crc kubenswrapper[4700]: E0227 17:03:42.980487 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:03:43 crc kubenswrapper[4700]: I0227 17:03:43.981148 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:03:43 crc kubenswrapper[4700]: I0227 17:03:43.981226 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:03:43 crc kubenswrapper[4700]: I0227 17:03:43.981146 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:03:43 crc kubenswrapper[4700]: E0227 17:03:43.981326 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:03:43 crc kubenswrapper[4700]: E0227 17:03:43.981564 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:03:43 crc kubenswrapper[4700]: E0227 17:03:43.981646 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:03:44 crc kubenswrapper[4700]: I0227 17:03:44.981094 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:03:44 crc kubenswrapper[4700]: E0227 17:03:44.981334 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:03:45 crc kubenswrapper[4700]: I0227 17:03:45.980968 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:03:45 crc kubenswrapper[4700]: I0227 17:03:45.981182 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:03:45 crc kubenswrapper[4700]: E0227 17:03:45.981451 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:03:45 crc kubenswrapper[4700]: E0227 17:03:45.981639 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:03:45 crc kubenswrapper[4700]: I0227 17:03:45.981752 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:03:45 crc kubenswrapper[4700]: E0227 17:03:45.982085 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:03:46 crc kubenswrapper[4700]: E0227 17:03:46.118780 4700 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 17:03:46 crc kubenswrapper[4700]: I0227 17:03:46.980710 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:03:46 crc kubenswrapper[4700]: E0227 17:03:46.980989 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:03:47 crc kubenswrapper[4700]: I0227 17:03:47.980872 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:03:47 crc kubenswrapper[4700]: I0227 17:03:47.980982 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:03:47 crc kubenswrapper[4700]: E0227 17:03:47.981061 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:03:47 crc kubenswrapper[4700]: I0227 17:03:47.980997 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:03:47 crc kubenswrapper[4700]: E0227 17:03:47.981241 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:03:47 crc kubenswrapper[4700]: E0227 17:03:47.981364 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:03:48 crc kubenswrapper[4700]: I0227 17:03:48.980538 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:03:48 crc kubenswrapper[4700]: E0227 17:03:48.980747 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:03:49 crc kubenswrapper[4700]: I0227 17:03:49.980598 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:03:49 crc kubenswrapper[4700]: I0227 17:03:49.980636 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:03:49 crc kubenswrapper[4700]: I0227 17:03:49.980598 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:03:49 crc kubenswrapper[4700]: E0227 17:03:49.980834 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:03:49 crc kubenswrapper[4700]: E0227 17:03:49.980936 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:03:49 crc kubenswrapper[4700]: E0227 17:03:49.981077 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:03:50 crc kubenswrapper[4700]: I0227 17:03:50.980130 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:03:50 crc kubenswrapper[4700]: E0227 17:03:50.982503 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:03:50 crc kubenswrapper[4700]: I0227 17:03:50.982982 4700 scope.go:117] "RemoveContainer" containerID="9c675574fe85ad3f55e0431b8d36fd1b83523ee502e675165aafad8551ba4560" Feb 27 17:03:51 crc kubenswrapper[4700]: E0227 17:03:51.119534 4700 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 27 17:03:51 crc kubenswrapper[4700]: I0227 17:03:51.980038 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:03:51 crc kubenswrapper[4700]: I0227 17:03:51.980180 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:03:51 crc kubenswrapper[4700]: E0227 17:03:51.980253 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:03:51 crc kubenswrapper[4700]: I0227 17:03:51.980038 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:03:51 crc kubenswrapper[4700]: E0227 17:03:51.980424 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:03:51 crc kubenswrapper[4700]: E0227 17:03:51.980600 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:03:51 crc kubenswrapper[4700]: I0227 17:03:51.985806 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-rbptz_66eb22fb-e593-40ff-9ff3-5bb32912972d/kube-multus/1.log" Feb 27 17:03:51 crc kubenswrapper[4700]: I0227 17:03:51.985900 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-rbptz" event={"ID":"66eb22fb-e593-40ff-9ff3-5bb32912972d","Type":"ContainerStarted","Data":"ebdea7f95d1c1fbe317f3b2abbfaf311938c90e4554c498e7dd55e758928933f"} Feb 27 17:03:52 crc kubenswrapper[4700]: I0227 17:03:52.981217 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:03:52 crc kubenswrapper[4700]: E0227 17:03:52.981603 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:03:53 crc kubenswrapper[4700]: I0227 17:03:53.980592 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:03:53 crc kubenswrapper[4700]: I0227 17:03:53.980644 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:03:53 crc kubenswrapper[4700]: E0227 17:03:53.980767 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:03:53 crc kubenswrapper[4700]: I0227 17:03:53.980884 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:03:53 crc kubenswrapper[4700]: E0227 17:03:53.981127 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:03:53 crc kubenswrapper[4700]: E0227 17:03:53.981300 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:03:54 crc kubenswrapper[4700]: I0227 17:03:54.980323 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:03:54 crc kubenswrapper[4700]: E0227 17:03:54.980547 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-s8k74" podUID="f5288317-4810-4321-baa0-31a118a4dc36" Feb 27 17:03:55 crc kubenswrapper[4700]: I0227 17:03:55.980393 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:03:55 crc kubenswrapper[4700]: I0227 17:03:55.980589 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:03:55 crc kubenswrapper[4700]: I0227 17:03:55.980552 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:03:55 crc kubenswrapper[4700]: E0227 17:03:55.980701 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 17:03:55 crc kubenswrapper[4700]: E0227 17:03:55.980842 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 17:03:55 crc kubenswrapper[4700]: E0227 17:03:55.980941 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 17:03:56 crc kubenswrapper[4700]: I0227 17:03:56.980979 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:03:56 crc kubenswrapper[4700]: I0227 17:03:56.984141 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 27 17:03:56 crc kubenswrapper[4700]: I0227 17:03:56.984442 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 27 17:03:57 crc kubenswrapper[4700]: I0227 17:03:57.980441 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:03:57 crc kubenswrapper[4700]: I0227 17:03:57.980760 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:03:57 crc kubenswrapper[4700]: I0227 17:03:57.981255 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:03:57 crc kubenswrapper[4700]: I0227 17:03:57.983348 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 27 17:03:57 crc kubenswrapper[4700]: I0227 17:03:57.983519 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 27 17:03:57 crc kubenswrapper[4700]: I0227 17:03:57.984178 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 27 17:03:57 crc kubenswrapper[4700]: I0227 17:03:57.984352 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.731692 4700 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.792526 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-95wjf"] Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.793413 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-95wjf" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.802902 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.803008 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.803209 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.803950 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.804699 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-6wfhz"] Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.805650 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.807158 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.807372 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-frd4g"] Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.807810 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.808513 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-frd4g" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.809929 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-zxt8n"] Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.810970 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.821202 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-vj29g"] Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.822250 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-vj29g" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.822447 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-w298r"] Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.823346 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-w298r" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.823982 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-66kw4"] Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.825335 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-66kw4" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.832432 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.834818 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-94pws"] Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.835571 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-cc8ct"] Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.836051 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-cc8ct" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.836268 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-b9772"] Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.837511 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-b9772" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.837554 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b6vld"] Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.838867 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b6vld" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.836646 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-94pws" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.845505 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.846322 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.846577 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-b5tz8"] Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.847059 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.847266 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-b5tz8" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.848069 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-btxgl"] Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.848791 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-btxgl" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.849627 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-v4nr9"] Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.849919 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.850066 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.850176 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.850283 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-v4nr9" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.850312 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.850548 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.853481 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.853567 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.854005 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.854242 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.854347 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.854577 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.854287 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.855145 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.855334 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.860560 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-kq9sb"] Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.861187 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-2jbbg"] Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.861931 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2jbbg" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.862432 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-kq9sb" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.861933 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.884247 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-k97gv"] Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.885385 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.888981 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.889665 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.889865 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.890107 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.890105 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.890326 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.903347 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.903414 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.903653 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.903958 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppmpb\" (UniqueName: \"kubernetes.io/projected/4ea98230-93e8-4ef6-86d6-e25b8448c632-kube-api-access-ppmpb\") pod \"machine-api-operator-5694c8668f-b9772\" (UID: \"4ea98230-93e8-4ef6-86d6-e25b8448c632\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-b9772" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904006 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a833ce54-1145-4b0e-8136-d1747e52673b-service-ca-bundle\") pod \"authentication-operator-69f744f599-cc8ct\" (UID: \"a833ce54-1145-4b0e-8136-d1747e52673b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cc8ct" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904023 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904039 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904059 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39f2673f-09d0-4815-babb-a8f5961eee45-serving-cert\") pod \"console-operator-58897d9998-vj29g\" (UID: \"39f2673f-09d0-4815-babb-a8f5961eee45\") " pod="openshift-console-operator/console-operator-58897d9998-vj29g" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904073 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904103 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/af75f830-5407-4eb7-ae1b-b09e6ef72737-machine-approver-tls\") pod \"machine-approver-56656f9798-94pws\" (UID: \"af75f830-5407-4eb7-ae1b-b09e6ef72737\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-94pws" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904119 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af75f830-5407-4eb7-ae1b-b09e6ef72737-config\") pod \"machine-approver-56656f9798-94pws\" (UID: \"af75f830-5407-4eb7-ae1b-b09e6ef72737\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-94pws" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904134 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52cl4\" (UniqueName: \"kubernetes.io/projected/91eb41f7-eb34-496d-b65d-20dfc7abf7f0-kube-api-access-52cl4\") pod \"openshift-config-operator-7777fb866f-btxgl\" (UID: \"91eb41f7-eb34-496d-b65d-20dfc7abf7f0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-btxgl" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904153 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3fdb7434-0752-4484-b253-948bbde41112-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-2jbbg\" (UID: \"3fdb7434-0752-4484-b253-948bbde41112\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2jbbg" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904167 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39f2673f-09d0-4815-babb-a8f5961eee45-config\") pod \"console-operator-58897d9998-vj29g\" (UID: \"39f2673f-09d0-4815-babb-a8f5961eee45\") " pod="openshift-console-operator/console-operator-58897d9998-vj29g" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904181 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f9d75460-0c2b-4c82-89d0-e41f8b8c275c-etcd-serving-ca\") pod \"apiserver-76f77b778f-6wfhz\" (UID: \"f9d75460-0c2b-4c82-89d0-e41f8b8c275c\") " pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904197 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a833ce54-1145-4b0e-8136-d1747e52673b-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-cc8ct\" (UID: \"a833ce54-1145-4b0e-8136-d1747e52673b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cc8ct" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904212 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/91eb41f7-eb34-496d-b65d-20dfc7abf7f0-available-featuregates\") pod \"openshift-config-operator-7777fb866f-btxgl\" (UID: \"91eb41f7-eb34-496d-b65d-20dfc7abf7f0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-btxgl" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904227 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3fdb7434-0752-4484-b253-948bbde41112-etcd-client\") pod \"apiserver-7bbb656c7d-2jbbg\" (UID: \"3fdb7434-0752-4484-b253-948bbde41112\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2jbbg" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904241 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/91eb41f7-eb34-496d-b65d-20dfc7abf7f0-serving-cert\") pod \"openshift-config-operator-7777fb866f-btxgl\" (UID: \"91eb41f7-eb34-496d-b65d-20dfc7abf7f0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-btxgl" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904255 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9d75460-0c2b-4c82-89d0-e41f8b8c275c-serving-cert\") pod \"apiserver-76f77b778f-6wfhz\" (UID: \"f9d75460-0c2b-4c82-89d0-e41f8b8c275c\") " pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904270 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/f9d75460-0c2b-4c82-89d0-e41f8b8c275c-image-import-ca\") pod \"apiserver-76f77b778f-6wfhz\" (UID: \"f9d75460-0c2b-4c82-89d0-e41f8b8c275c\") " pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904285 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9d75460-0c2b-4c82-89d0-e41f8b8c275c-config\") pod \"apiserver-76f77b778f-6wfhz\" (UID: \"f9d75460-0c2b-4c82-89d0-e41f8b8c275c\") " pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904302 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0adc70e1-038b-45c1-80a8-9e1f938fc161-config\") pod \"controller-manager-879f6c89f-95wjf\" (UID: \"0adc70e1-038b-45c1-80a8-9e1f938fc161\") " pod="openshift-controller-manager/controller-manager-879f6c89f-95wjf" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904315 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/f9d75460-0c2b-4c82-89d0-e41f8b8c275c-audit\") pod \"apiserver-76f77b778f-6wfhz\" (UID: \"f9d75460-0c2b-4c82-89d0-e41f8b8c275c\") " pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904332 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904349 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dd3b00b4-f42e-406c-9b88-b7e2602a2493-client-ca\") pod \"route-controller-manager-6576b87f9c-frd4g\" (UID: \"dd3b00b4-f42e-406c-9b88-b7e2602a2493\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-frd4g" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904362 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3fdb7434-0752-4484-b253-948bbde41112-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-2jbbg\" (UID: \"3fdb7434-0752-4484-b253-948bbde41112\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2jbbg" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904380 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5przj\" (UniqueName: \"kubernetes.io/projected/0adc70e1-038b-45c1-80a8-9e1f938fc161-kube-api-access-5przj\") pod \"controller-manager-879f6c89f-95wjf\" (UID: \"0adc70e1-038b-45c1-80a8-9e1f938fc161\") " pod="openshift-controller-manager/controller-manager-879f6c89f-95wjf" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904394 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a833ce54-1145-4b0e-8136-d1747e52673b-config\") pod \"authentication-operator-69f744f599-cc8ct\" (UID: \"a833ce54-1145-4b0e-8136-d1747e52673b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cc8ct" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904408 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xt2c\" (UniqueName: \"kubernetes.io/projected/af75f830-5407-4eb7-ae1b-b09e6ef72737-kube-api-access-9xt2c\") pod \"machine-approver-56656f9798-94pws\" (UID: \"af75f830-5407-4eb7-ae1b-b09e6ef72737\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-94pws" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904428 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szvzs\" (UniqueName: \"kubernetes.io/projected/39f2673f-09d0-4815-babb-a8f5961eee45-kube-api-access-szvzs\") pod \"console-operator-58897d9998-vj29g\" (UID: \"39f2673f-09d0-4815-babb-a8f5961eee45\") " pod="openshift-console-operator/console-operator-58897d9998-vj29g" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904443 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0adc70e1-038b-45c1-80a8-9e1f938fc161-serving-cert\") pod \"controller-manager-879f6c89f-95wjf\" (UID: \"0adc70e1-038b-45c1-80a8-9e1f938fc161\") " pod="openshift-controller-manager/controller-manager-879f6c89f-95wjf" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904481 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/f9d75460-0c2b-4c82-89d0-e41f8b8c275c-node-pullsecrets\") pod \"apiserver-76f77b778f-6wfhz\" (UID: \"f9d75460-0c2b-4c82-89d0-e41f8b8c275c\") " pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904497 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f9d75460-0c2b-4c82-89d0-e41f8b8c275c-audit-dir\") pod \"apiserver-76f77b778f-6wfhz\" (UID: \"f9d75460-0c2b-4c82-89d0-e41f8b8c275c\") " pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904512 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f9d75460-0c2b-4c82-89d0-e41f8b8c275c-encryption-config\") pod \"apiserver-76f77b778f-6wfhz\" (UID: \"f9d75460-0c2b-4c82-89d0-e41f8b8c275c\") " pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904527 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rffxp\" (UniqueName: \"kubernetes.io/projected/af74d952-25fb-4e1a-ab91-9c606b4b00ab-kube-api-access-rffxp\") pod \"downloads-7954f5f757-v4nr9\" (UID: \"af74d952-25fb-4e1a-ab91-9c606b4b00ab\") " pod="openshift-console/downloads-7954f5f757-v4nr9" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904544 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904561 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3fdb7434-0752-4484-b253-948bbde41112-serving-cert\") pod \"apiserver-7bbb656c7d-2jbbg\" (UID: \"3fdb7434-0752-4484-b253-948bbde41112\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2jbbg" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904575 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3fdb7434-0752-4484-b253-948bbde41112-encryption-config\") pod \"apiserver-7bbb656c7d-2jbbg\" (UID: \"3fdb7434-0752-4484-b253-948bbde41112\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2jbbg" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904590 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd3b00b4-f42e-406c-9b88-b7e2602a2493-serving-cert\") pod \"route-controller-manager-6576b87f9c-frd4g\" (UID: \"dd3b00b4-f42e-406c-9b88-b7e2602a2493\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-frd4g" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904612 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5m8t\" (UniqueName: \"kubernetes.io/projected/f9d75460-0c2b-4c82-89d0-e41f8b8c275c-kube-api-access-z5m8t\") pod \"apiserver-76f77b778f-6wfhz\" (UID: \"f9d75460-0c2b-4c82-89d0-e41f8b8c275c\") " pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904629 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd3b00b4-f42e-406c-9b88-b7e2602a2493-config\") pod \"route-controller-manager-6576b87f9c-frd4g\" (UID: \"dd3b00b4-f42e-406c-9b88-b7e2602a2493\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-frd4g" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904653 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904670 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3fdb7434-0752-4484-b253-948bbde41112-audit-policies\") pod \"apiserver-7bbb656c7d-2jbbg\" (UID: \"3fdb7434-0752-4484-b253-948bbde41112\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2jbbg" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904684 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a833ce54-1145-4b0e-8136-d1747e52673b-serving-cert\") pod \"authentication-operator-69f744f599-cc8ct\" (UID: \"a833ce54-1145-4b0e-8136-d1747e52673b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cc8ct" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904697 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f9d75460-0c2b-4c82-89d0-e41f8b8c275c-etcd-client\") pod \"apiserver-76f77b778f-6wfhz\" (UID: \"f9d75460-0c2b-4c82-89d0-e41f8b8c275c\") " pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904716 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4n8qd\" (UniqueName: \"kubernetes.io/projected/dd3b00b4-f42e-406c-9b88-b7e2602a2493-kube-api-access-4n8qd\") pod \"route-controller-manager-6576b87f9c-frd4g\" (UID: \"dd3b00b4-f42e-406c-9b88-b7e2602a2493\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-frd4g" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904730 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ns985\" (UniqueName: \"kubernetes.io/projected/3fdb7434-0752-4484-b253-948bbde41112-kube-api-access-ns985\") pod \"apiserver-7bbb656c7d-2jbbg\" (UID: \"3fdb7434-0752-4484-b253-948bbde41112\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2jbbg" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904744 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/af75f830-5407-4eb7-ae1b-b09e6ef72737-auth-proxy-config\") pod \"machine-approver-56656f9798-94pws\" (UID: \"af75f830-5407-4eb7-ae1b-b09e6ef72737\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-94pws" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904759 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904775 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904793 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ea98230-93e8-4ef6-86d6-e25b8448c632-config\") pod \"machine-api-operator-5694c8668f-b9772\" (UID: \"4ea98230-93e8-4ef6-86d6-e25b8448c632\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-b9772" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904809 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0adc70e1-038b-45c1-80a8-9e1f938fc161-client-ca\") pod \"controller-manager-879f6c89f-95wjf\" (UID: \"0adc70e1-038b-45c1-80a8-9e1f938fc161\") " pod="openshift-controller-manager/controller-manager-879f6c89f-95wjf" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904824 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0adc70e1-038b-45c1-80a8-9e1f938fc161-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-95wjf\" (UID: \"0adc70e1-038b-45c1-80a8-9e1f938fc161\") " pod="openshift-controller-manager/controller-manager-879f6c89f-95wjf" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904840 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/649e4fdc-b0b6-4231-8038-d4fa805db298-audit-policies\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904858 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8k6kl\" (UniqueName: \"kubernetes.io/projected/a833ce54-1145-4b0e-8136-d1747e52673b-kube-api-access-8k6kl\") pod \"authentication-operator-69f744f599-cc8ct\" (UID: \"a833ce54-1145-4b0e-8136-d1747e52673b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cc8ct" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904893 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-db9mr\" (UniqueName: \"kubernetes.io/projected/649e4fdc-b0b6-4231-8038-d4fa805db298-kube-api-access-db9mr\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904909 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f9d75460-0c2b-4c82-89d0-e41f8b8c275c-trusted-ca-bundle\") pod \"apiserver-76f77b778f-6wfhz\" (UID: \"f9d75460-0c2b-4c82-89d0-e41f8b8c275c\") " pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904924 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904940 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904954 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4ea98230-93e8-4ef6-86d6-e25b8448c632-images\") pod \"machine-api-operator-5694c8668f-b9772\" (UID: \"4ea98230-93e8-4ef6-86d6-e25b8448c632\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-b9772" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.904969 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/39f2673f-09d0-4815-babb-a8f5961eee45-trusted-ca\") pod \"console-operator-58897d9998-vj29g\" (UID: \"39f2673f-09d0-4815-babb-a8f5961eee45\") " pod="openshift-console-operator/console-operator-58897d9998-vj29g" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.905006 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/4ea98230-93e8-4ef6-86d6-e25b8448c632-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-b9772\" (UID: \"4ea98230-93e8-4ef6-86d6-e25b8448c632\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-b9772" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.905021 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3fdb7434-0752-4484-b253-948bbde41112-audit-dir\") pod \"apiserver-7bbb656c7d-2jbbg\" (UID: \"3fdb7434-0752-4484-b253-948bbde41112\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2jbbg" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.905035 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/649e4fdc-b0b6-4231-8038-d4fa805db298-audit-dir\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.905050 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.905785 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.906209 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.906441 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.907243 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.907515 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.907929 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.908111 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.908234 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.908336 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.908505 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.908681 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.908830 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.908979 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.909641 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.909797 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.910076 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.910304 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.916575 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.916782 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.916931 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.917095 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.917247 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.917507 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.917736 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.918181 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.918380 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-l2nwx"] Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.918773 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9mh7s"] Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.919097 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9mh7s" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.919339 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.919491 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-l2nwx" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.919734 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.919871 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.919892 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.920076 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.920279 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.920381 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.920449 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.920492 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.920570 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.920709 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.921062 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.923091 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.922795 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-nw2mh"] Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.921574 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.923739 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-nw2mh" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.930157 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c7zb5"] Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.930696 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c7zb5" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.931026 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.931305 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.931632 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.931787 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.931892 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.932001 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.932114 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.932292 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.937997 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.938196 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.938280 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.938355 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.938482 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.938555 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.939934 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.940155 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.940261 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.942144 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.942246 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.942489 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.944940 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.946711 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.946985 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.947551 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.949337 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.949799 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.951184 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.953139 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.959404 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.961655 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-jzpsv"] Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.964105 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sbqxk"] Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.964437 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-jzpsv" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.969979 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sbqxk" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.970675 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2blr9"] Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.971672 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2blr9" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.976245 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.978815 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.979313 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.980346 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.980591 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.987217 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-95wjf"] Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.998363 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-58xmr"] Feb 27 17:04:04 crc kubenswrapper[4700]: I0227 17:04:04.998938 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-58xmr" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.005297 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fvzmk"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.005790 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dd3b00b4-f42e-406c-9b88-b7e2602a2493-client-ca\") pod \"route-controller-manager-6576b87f9c-frd4g\" (UID: \"dd3b00b4-f42e-406c-9b88-b7e2602a2493\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-frd4g" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.005824 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3fdb7434-0752-4484-b253-948bbde41112-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-2jbbg\" (UID: \"3fdb7434-0752-4484-b253-948bbde41112\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2jbbg" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.005851 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a833ce54-1145-4b0e-8136-d1747e52673b-config\") pod \"authentication-operator-69f744f599-cc8ct\" (UID: \"a833ce54-1145-4b0e-8136-d1747e52673b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cc8ct" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.005875 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5przj\" (UniqueName: \"kubernetes.io/projected/0adc70e1-038b-45c1-80a8-9e1f938fc161-kube-api-access-5przj\") pod \"controller-manager-879f6c89f-95wjf\" (UID: \"0adc70e1-038b-45c1-80a8-9e1f938fc161\") " pod="openshift-controller-manager/controller-manager-879f6c89f-95wjf" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.005907 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ac062b00-1a72-44b4-90c9-de1573e3e217-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-66kw4\" (UID: \"ac062b00-1a72-44b4-90c9-de1573e3e217\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-66kw4" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.005931 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xt2c\" (UniqueName: \"kubernetes.io/projected/af75f830-5407-4eb7-ae1b-b09e6ef72737-kube-api-access-9xt2c\") pod \"machine-approver-56656f9798-94pws\" (UID: \"af75f830-5407-4eb7-ae1b-b09e6ef72737\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-94pws" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.005937 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fvzmk" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.005955 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lp2mq\" (UniqueName: \"kubernetes.io/projected/f2257cba-c045-4865-aecf-03642416f2cb-kube-api-access-lp2mq\") pod \"openshift-apiserver-operator-796bbdcf4f-w298r\" (UID: \"f2257cba-c045-4865-aecf-03642416f2cb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-w298r" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.005982 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/f9d75460-0c2b-4c82-89d0-e41f8b8c275c-node-pullsecrets\") pod \"apiserver-76f77b778f-6wfhz\" (UID: \"f9d75460-0c2b-4c82-89d0-e41f8b8c275c\") " pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006006 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szvzs\" (UniqueName: \"kubernetes.io/projected/39f2673f-09d0-4815-babb-a8f5961eee45-kube-api-access-szvzs\") pod \"console-operator-58897d9998-vj29g\" (UID: \"39f2673f-09d0-4815-babb-a8f5961eee45\") " pod="openshift-console-operator/console-operator-58897d9998-vj29g" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006027 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0adc70e1-038b-45c1-80a8-9e1f938fc161-serving-cert\") pod \"controller-manager-879f6c89f-95wjf\" (UID: \"0adc70e1-038b-45c1-80a8-9e1f938fc161\") " pod="openshift-controller-manager/controller-manager-879f6c89f-95wjf" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006050 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f9d75460-0c2b-4c82-89d0-e41f8b8c275c-audit-dir\") pod \"apiserver-76f77b778f-6wfhz\" (UID: \"f9d75460-0c2b-4c82-89d0-e41f8b8c275c\") " pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006074 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zp8mz\" (UniqueName: \"kubernetes.io/projected/26f6041e-592f-48c6-93e5-35517c775186-kube-api-access-zp8mz\") pod \"cluster-samples-operator-665b6dd947-b6vld\" (UID: \"26f6041e-592f-48c6-93e5-35517c775186\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b6vld" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006098 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f9d75460-0c2b-4c82-89d0-e41f8b8c275c-encryption-config\") pod \"apiserver-76f77b778f-6wfhz\" (UID: \"f9d75460-0c2b-4c82-89d0-e41f8b8c275c\") " pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006123 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rffxp\" (UniqueName: \"kubernetes.io/projected/af74d952-25fb-4e1a-ab91-9c606b4b00ab-kube-api-access-rffxp\") pod \"downloads-7954f5f757-v4nr9\" (UID: \"af74d952-25fb-4e1a-ab91-9c606b4b00ab\") " pod="openshift-console/downloads-7954f5f757-v4nr9" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006150 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006173 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/facc8439-0b4d-4228-95b2-d01d48ee1c53-trusted-ca-bundle\") pod \"console-f9d7485db-kq9sb\" (UID: \"facc8439-0b4d-4228-95b2-d01d48ee1c53\") " pod="openshift-console/console-f9d7485db-kq9sb" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006195 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd3b00b4-f42e-406c-9b88-b7e2602a2493-serving-cert\") pod \"route-controller-manager-6576b87f9c-frd4g\" (UID: \"dd3b00b4-f42e-406c-9b88-b7e2602a2493\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-frd4g" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006218 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3fdb7434-0752-4484-b253-948bbde41112-serving-cert\") pod \"apiserver-7bbb656c7d-2jbbg\" (UID: \"3fdb7434-0752-4484-b253-948bbde41112\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2jbbg" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006238 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3fdb7434-0752-4484-b253-948bbde41112-encryption-config\") pod \"apiserver-7bbb656c7d-2jbbg\" (UID: \"3fdb7434-0752-4484-b253-948bbde41112\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2jbbg" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006275 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5m8t\" (UniqueName: \"kubernetes.io/projected/f9d75460-0c2b-4c82-89d0-e41f8b8c275c-kube-api-access-z5m8t\") pod \"apiserver-76f77b778f-6wfhz\" (UID: \"f9d75460-0c2b-4c82-89d0-e41f8b8c275c\") " pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006298 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/ac062b00-1a72-44b4-90c9-de1573e3e217-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-66kw4\" (UID: \"ac062b00-1a72-44b4-90c9-de1573e3e217\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-66kw4" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006325 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/facc8439-0b4d-4228-95b2-d01d48ee1c53-console-oauth-config\") pod \"console-f9d7485db-kq9sb\" (UID: \"facc8439-0b4d-4228-95b2-d01d48ee1c53\") " pod="openshift-console/console-f9d7485db-kq9sb" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006352 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd3b00b4-f42e-406c-9b88-b7e2602a2493-config\") pod \"route-controller-manager-6576b87f9c-frd4g\" (UID: \"dd3b00b4-f42e-406c-9b88-b7e2602a2493\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-frd4g" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006373 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006396 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/facc8439-0b4d-4228-95b2-d01d48ee1c53-service-ca\") pod \"console-f9d7485db-kq9sb\" (UID: \"facc8439-0b4d-4228-95b2-d01d48ee1c53\") " pod="openshift-console/console-f9d7485db-kq9sb" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006419 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3fdb7434-0752-4484-b253-948bbde41112-audit-policies\") pod \"apiserver-7bbb656c7d-2jbbg\" (UID: \"3fdb7434-0752-4484-b253-948bbde41112\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2jbbg" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006440 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a833ce54-1145-4b0e-8136-d1747e52673b-serving-cert\") pod \"authentication-operator-69f744f599-cc8ct\" (UID: \"a833ce54-1145-4b0e-8136-d1747e52673b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cc8ct" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006484 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f9d75460-0c2b-4c82-89d0-e41f8b8c275c-etcd-client\") pod \"apiserver-76f77b778f-6wfhz\" (UID: \"f9d75460-0c2b-4c82-89d0-e41f8b8c275c\") " pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006511 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006537 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4n8qd\" (UniqueName: \"kubernetes.io/projected/dd3b00b4-f42e-406c-9b88-b7e2602a2493-kube-api-access-4n8qd\") pod \"route-controller-manager-6576b87f9c-frd4g\" (UID: \"dd3b00b4-f42e-406c-9b88-b7e2602a2493\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-frd4g" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006559 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ns985\" (UniqueName: \"kubernetes.io/projected/3fdb7434-0752-4484-b253-948bbde41112-kube-api-access-ns985\") pod \"apiserver-7bbb656c7d-2jbbg\" (UID: \"3fdb7434-0752-4484-b253-948bbde41112\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2jbbg" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006580 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/af75f830-5407-4eb7-ae1b-b09e6ef72737-auth-proxy-config\") pod \"machine-approver-56656f9798-94pws\" (UID: \"af75f830-5407-4eb7-ae1b-b09e6ef72737\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-94pws" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006604 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006668 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ea98230-93e8-4ef6-86d6-e25b8448c632-config\") pod \"machine-api-operator-5694c8668f-b9772\" (UID: \"4ea98230-93e8-4ef6-86d6-e25b8448c632\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-b9772" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006698 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0adc70e1-038b-45c1-80a8-9e1f938fc161-client-ca\") pod \"controller-manager-879f6c89f-95wjf\" (UID: \"0adc70e1-038b-45c1-80a8-9e1f938fc161\") " pod="openshift-controller-manager/controller-manager-879f6c89f-95wjf" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006742 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0adc70e1-038b-45c1-80a8-9e1f938fc161-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-95wjf\" (UID: \"0adc70e1-038b-45c1-80a8-9e1f938fc161\") " pod="openshift-controller-manager/controller-manager-879f6c89f-95wjf" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006763 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/649e4fdc-b0b6-4231-8038-d4fa805db298-audit-policies\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006785 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2257cba-c045-4865-aecf-03642416f2cb-config\") pod \"openshift-apiserver-operator-796bbdcf4f-w298r\" (UID: \"f2257cba-c045-4865-aecf-03642416f2cb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-w298r" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006808 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/26f6041e-592f-48c6-93e5-35517c775186-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-b6vld\" (UID: \"26f6041e-592f-48c6-93e5-35517c775186\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b6vld" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006833 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8k6kl\" (UniqueName: \"kubernetes.io/projected/a833ce54-1145-4b0e-8136-d1747e52673b-kube-api-access-8k6kl\") pod \"authentication-operator-69f744f599-cc8ct\" (UID: \"a833ce54-1145-4b0e-8136-d1747e52673b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cc8ct" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006855 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f2257cba-c045-4865-aecf-03642416f2cb-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-w298r\" (UID: \"f2257cba-c045-4865-aecf-03642416f2cb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-w298r" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006878 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swq2d\" (UniqueName: \"kubernetes.io/projected/ac062b00-1a72-44b4-90c9-de1573e3e217-kube-api-access-swq2d\") pod \"cluster-image-registry-operator-dc59b4c8b-66kw4\" (UID: \"ac062b00-1a72-44b4-90c9-de1573e3e217\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-66kw4" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006902 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-db9mr\" (UniqueName: \"kubernetes.io/projected/649e4fdc-b0b6-4231-8038-d4fa805db298-kube-api-access-db9mr\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006924 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f9d75460-0c2b-4c82-89d0-e41f8b8c275c-trusted-ca-bundle\") pod \"apiserver-76f77b778f-6wfhz\" (UID: \"f9d75460-0c2b-4c82-89d0-e41f8b8c275c\") " pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006947 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006959 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dd3b00b4-f42e-406c-9b88-b7e2602a2493-client-ca\") pod \"route-controller-manager-6576b87f9c-frd4g\" (UID: \"dd3b00b4-f42e-406c-9b88-b7e2602a2493\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-frd4g" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.006974 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.007044 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4ea98230-93e8-4ef6-86d6-e25b8448c632-images\") pod \"machine-api-operator-5694c8668f-b9772\" (UID: \"4ea98230-93e8-4ef6-86d6-e25b8448c632\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-b9772" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.007074 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/39f2673f-09d0-4815-babb-a8f5961eee45-trusted-ca\") pod \"console-operator-58897d9998-vj29g\" (UID: \"39f2673f-09d0-4815-babb-a8f5961eee45\") " pod="openshift-console-operator/console-operator-58897d9998-vj29g" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.007099 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3fdb7434-0752-4484-b253-948bbde41112-audit-dir\") pod \"apiserver-7bbb656c7d-2jbbg\" (UID: \"3fdb7434-0752-4484-b253-948bbde41112\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2jbbg" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.007130 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a833ce54-1145-4b0e-8136-d1747e52673b-config\") pod \"authentication-operator-69f744f599-cc8ct\" (UID: \"a833ce54-1145-4b0e-8136-d1747e52673b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cc8ct" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.007162 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/4ea98230-93e8-4ef6-86d6-e25b8448c632-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-b9772\" (UID: \"4ea98230-93e8-4ef6-86d6-e25b8448c632\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-b9772" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.007194 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/649e4fdc-b0b6-4231-8038-d4fa805db298-audit-dir\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.007222 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.007251 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppmpb\" (UniqueName: \"kubernetes.io/projected/4ea98230-93e8-4ef6-86d6-e25b8448c632-kube-api-access-ppmpb\") pod \"machine-api-operator-5694c8668f-b9772\" (UID: \"4ea98230-93e8-4ef6-86d6-e25b8448c632\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-b9772" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.007279 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a833ce54-1145-4b0e-8136-d1747e52673b-service-ca-bundle\") pod \"authentication-operator-69f744f599-cc8ct\" (UID: \"a833ce54-1145-4b0e-8136-d1747e52673b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cc8ct" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.007305 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.007329 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.007335 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/f9d75460-0c2b-4c82-89d0-e41f8b8c275c-node-pullsecrets\") pod \"apiserver-76f77b778f-6wfhz\" (UID: \"f9d75460-0c2b-4c82-89d0-e41f8b8c275c\") " pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.007357 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/facc8439-0b4d-4228-95b2-d01d48ee1c53-oauth-serving-cert\") pod \"console-f9d7485db-kq9sb\" (UID: \"facc8439-0b4d-4228-95b2-d01d48ee1c53\") " pod="openshift-console/console-f9d7485db-kq9sb" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.007397 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39f2673f-09d0-4815-babb-a8f5961eee45-serving-cert\") pod \"console-operator-58897d9998-vj29g\" (UID: \"39f2673f-09d0-4815-babb-a8f5961eee45\") " pod="openshift-console-operator/console-operator-58897d9998-vj29g" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.007420 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.007444 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6f4c\" (UniqueName: \"kubernetes.io/projected/facc8439-0b4d-4228-95b2-d01d48ee1c53-kube-api-access-d6f4c\") pod \"console-f9d7485db-kq9sb\" (UID: \"facc8439-0b4d-4228-95b2-d01d48ee1c53\") " pod="openshift-console/console-f9d7485db-kq9sb" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.007495 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/af75f830-5407-4eb7-ae1b-b09e6ef72737-machine-approver-tls\") pod \"machine-approver-56656f9798-94pws\" (UID: \"af75f830-5407-4eb7-ae1b-b09e6ef72737\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-94pws" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.007516 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af75f830-5407-4eb7-ae1b-b09e6ef72737-config\") pod \"machine-approver-56656f9798-94pws\" (UID: \"af75f830-5407-4eb7-ae1b-b09e6ef72737\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-94pws" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.007540 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52cl4\" (UniqueName: \"kubernetes.io/projected/91eb41f7-eb34-496d-b65d-20dfc7abf7f0-kube-api-access-52cl4\") pod \"openshift-config-operator-7777fb866f-btxgl\" (UID: \"91eb41f7-eb34-496d-b65d-20dfc7abf7f0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-btxgl" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.007564 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/facc8439-0b4d-4228-95b2-d01d48ee1c53-console-config\") pod \"console-f9d7485db-kq9sb\" (UID: \"facc8439-0b4d-4228-95b2-d01d48ee1c53\") " pod="openshift-console/console-f9d7485db-kq9sb" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.007591 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3fdb7434-0752-4484-b253-948bbde41112-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-2jbbg\" (UID: \"3fdb7434-0752-4484-b253-948bbde41112\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2jbbg" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.007613 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39f2673f-09d0-4815-babb-a8f5961eee45-config\") pod \"console-operator-58897d9998-vj29g\" (UID: \"39f2673f-09d0-4815-babb-a8f5961eee45\") " pod="openshift-console-operator/console-operator-58897d9998-vj29g" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.007638 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f9d75460-0c2b-4c82-89d0-e41f8b8c275c-etcd-serving-ca\") pod \"apiserver-76f77b778f-6wfhz\" (UID: \"f9d75460-0c2b-4c82-89d0-e41f8b8c275c\") " pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.007662 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3fdb7434-0752-4484-b253-948bbde41112-etcd-client\") pod \"apiserver-7bbb656c7d-2jbbg\" (UID: \"3fdb7434-0752-4484-b253-948bbde41112\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2jbbg" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.007686 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a833ce54-1145-4b0e-8136-d1747e52673b-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-cc8ct\" (UID: \"a833ce54-1145-4b0e-8136-d1747e52673b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cc8ct" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.007711 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/91eb41f7-eb34-496d-b65d-20dfc7abf7f0-available-featuregates\") pod \"openshift-config-operator-7777fb866f-btxgl\" (UID: \"91eb41f7-eb34-496d-b65d-20dfc7abf7f0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-btxgl" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.007744 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/91eb41f7-eb34-496d-b65d-20dfc7abf7f0-serving-cert\") pod \"openshift-config-operator-7777fb866f-btxgl\" (UID: \"91eb41f7-eb34-496d-b65d-20dfc7abf7f0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-btxgl" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.007770 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9d75460-0c2b-4c82-89d0-e41f8b8c275c-serving-cert\") pod \"apiserver-76f77b778f-6wfhz\" (UID: \"f9d75460-0c2b-4c82-89d0-e41f8b8c275c\") " pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.007796 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ac062b00-1a72-44b4-90c9-de1573e3e217-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-66kw4\" (UID: \"ac062b00-1a72-44b4-90c9-de1573e3e217\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-66kw4" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.007865 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/f9d75460-0c2b-4c82-89d0-e41f8b8c275c-image-import-ca\") pod \"apiserver-76f77b778f-6wfhz\" (UID: \"f9d75460-0c2b-4c82-89d0-e41f8b8c275c\") " pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.007893 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9d75460-0c2b-4c82-89d0-e41f8b8c275c-config\") pod \"apiserver-76f77b778f-6wfhz\" (UID: \"f9d75460-0c2b-4c82-89d0-e41f8b8c275c\") " pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.007918 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/facc8439-0b4d-4228-95b2-d01d48ee1c53-console-serving-cert\") pod \"console-f9d7485db-kq9sb\" (UID: \"facc8439-0b4d-4228-95b2-d01d48ee1c53\") " pod="openshift-console/console-f9d7485db-kq9sb" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.007947 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0adc70e1-038b-45c1-80a8-9e1f938fc161-config\") pod \"controller-manager-879f6c89f-95wjf\" (UID: \"0adc70e1-038b-45c1-80a8-9e1f938fc161\") " pod="openshift-controller-manager/controller-manager-879f6c89f-95wjf" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.007968 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/f9d75460-0c2b-4c82-89d0-e41f8b8c275c-audit\") pod \"apiserver-76f77b778f-6wfhz\" (UID: \"f9d75460-0c2b-4c82-89d0-e41f8b8c275c\") " pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.007993 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.008274 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.008327 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3fdb7434-0752-4484-b253-948bbde41112-audit-dir\") pod \"apiserver-7bbb656c7d-2jbbg\" (UID: \"3fdb7434-0752-4484-b253-948bbde41112\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2jbbg" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.008788 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f9d75460-0c2b-4c82-89d0-e41f8b8c275c-audit-dir\") pod \"apiserver-76f77b778f-6wfhz\" (UID: \"f9d75460-0c2b-4c82-89d0-e41f8b8c275c\") " pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.008895 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4ea98230-93e8-4ef6-86d6-e25b8448c632-images\") pod \"machine-api-operator-5694c8668f-b9772\" (UID: \"4ea98230-93e8-4ef6-86d6-e25b8448c632\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-b9772" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.009251 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3fdb7434-0752-4484-b253-948bbde41112-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-2jbbg\" (UID: \"3fdb7434-0752-4484-b253-948bbde41112\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2jbbg" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.009371 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a833ce54-1145-4b0e-8136-d1747e52673b-service-ca-bundle\") pod \"authentication-operator-69f744f599-cc8ct\" (UID: \"a833ce54-1145-4b0e-8136-d1747e52673b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cc8ct" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.009517 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/649e4fdc-b0b6-4231-8038-d4fa805db298-audit-dir\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.010027 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/af75f830-5407-4eb7-ae1b-b09e6ef72737-auth-proxy-config\") pod \"machine-approver-56656f9798-94pws\" (UID: \"af75f830-5407-4eb7-ae1b-b09e6ef72737\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-94pws" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.010573 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd3b00b4-f42e-406c-9b88-b7e2602a2493-config\") pod \"route-controller-manager-6576b87f9c-frd4g\" (UID: \"dd3b00b4-f42e-406c-9b88-b7e2602a2493\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-frd4g" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.010952 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/39f2673f-09d0-4815-babb-a8f5961eee45-trusted-ca\") pod \"console-operator-58897d9998-vj29g\" (UID: \"39f2673f-09d0-4815-babb-a8f5961eee45\") " pod="openshift-console-operator/console-operator-58897d9998-vj29g" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.010973 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ea98230-93e8-4ef6-86d6-e25b8448c632-config\") pod \"machine-api-operator-5694c8668f-b9772\" (UID: \"4ea98230-93e8-4ef6-86d6-e25b8448c632\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-b9772" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.011751 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0adc70e1-038b-45c1-80a8-9e1f938fc161-client-ca\") pod \"controller-manager-879f6c89f-95wjf\" (UID: \"0adc70e1-038b-45c1-80a8-9e1f938fc161\") " pod="openshift-controller-manager/controller-manager-879f6c89f-95wjf" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.013151 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0adc70e1-038b-45c1-80a8-9e1f938fc161-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-95wjf\" (UID: \"0adc70e1-038b-45c1-80a8-9e1f938fc161\") " pod="openshift-controller-manager/controller-manager-879f6c89f-95wjf" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.013187 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-kq896"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.013700 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.013920 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kq896" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.013922 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3fdb7434-0752-4484-b253-948bbde41112-audit-policies\") pod \"apiserver-7bbb656c7d-2jbbg\" (UID: \"3fdb7434-0752-4484-b253-948bbde41112\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2jbbg" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.014096 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f9d75460-0c2b-4c82-89d0-e41f8b8c275c-trusted-ca-bundle\") pod \"apiserver-76f77b778f-6wfhz\" (UID: \"f9d75460-0c2b-4c82-89d0-e41f8b8c275c\") " pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.013923 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f9d75460-0c2b-4c82-89d0-e41f8b8c275c-encryption-config\") pod \"apiserver-76f77b778f-6wfhz\" (UID: \"f9d75460-0c2b-4c82-89d0-e41f8b8c275c\") " pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.014590 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.015170 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f9d75460-0c2b-4c82-89d0-e41f8b8c275c-etcd-serving-ca\") pod \"apiserver-76f77b778f-6wfhz\" (UID: \"f9d75460-0c2b-4c82-89d0-e41f8b8c275c\") " pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.015473 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.015618 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.015668 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3fdb7434-0752-4484-b253-948bbde41112-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-2jbbg\" (UID: \"3fdb7434-0752-4484-b253-948bbde41112\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2jbbg" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.015948 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/91eb41f7-eb34-496d-b65d-20dfc7abf7f0-available-featuregates\") pod \"openshift-config-operator-7777fb866f-btxgl\" (UID: \"91eb41f7-eb34-496d-b65d-20dfc7abf7f0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-btxgl" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.015955 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af75f830-5407-4eb7-ae1b-b09e6ef72737-config\") pod \"machine-approver-56656f9798-94pws\" (UID: \"af75f830-5407-4eb7-ae1b-b09e6ef72737\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-94pws" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.016230 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0adc70e1-038b-45c1-80a8-9e1f938fc161-serving-cert\") pod \"controller-manager-879f6c89f-95wjf\" (UID: \"0adc70e1-038b-45c1-80a8-9e1f938fc161\") " pod="openshift-controller-manager/controller-manager-879f6c89f-95wjf" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.016514 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f9d75460-0c2b-4c82-89d0-e41f8b8c275c-etcd-client\") pod \"apiserver-76f77b778f-6wfhz\" (UID: \"f9d75460-0c2b-4c82-89d0-e41f8b8c275c\") " pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.016866 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.017087 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.017080 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/f9d75460-0c2b-4c82-89d0-e41f8b8c275c-image-import-ca\") pod \"apiserver-76f77b778f-6wfhz\" (UID: \"f9d75460-0c2b-4c82-89d0-e41f8b8c275c\") " pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.009382 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/649e4fdc-b0b6-4231-8038-d4fa805db298-audit-policies\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.017615 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0adc70e1-038b-45c1-80a8-9e1f938fc161-config\") pod \"controller-manager-879f6c89f-95wjf\" (UID: \"0adc70e1-038b-45c1-80a8-9e1f938fc161\") " pod="openshift-controller-manager/controller-manager-879f6c89f-95wjf" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.017672 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-frd4g"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.017691 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hc86c"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.018025 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-hc86c" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.018152 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9d75460-0c2b-4c82-89d0-e41f8b8c275c-config\") pod \"apiserver-76f77b778f-6wfhz\" (UID: \"f9d75460-0c2b-4c82-89d0-e41f8b8c275c\") " pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.018252 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-5svpj"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.018880 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5svpj" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.019211 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/f9d75460-0c2b-4c82-89d0-e41f8b8c275c-audit\") pod \"apiserver-76f77b778f-6wfhz\" (UID: \"f9d75460-0c2b-4c82-89d0-e41f8b8c275c\") " pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.019431 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a833ce54-1145-4b0e-8136-d1747e52673b-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-cc8ct\" (UID: \"a833ce54-1145-4b0e-8136-d1747e52673b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cc8ct" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.020102 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39f2673f-09d0-4815-babb-a8f5961eee45-config\") pod \"console-operator-58897d9998-vj29g\" (UID: \"39f2673f-09d0-4815-babb-a8f5961eee45\") " pod="openshift-console-operator/console-operator-58897d9998-vj29g" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.020173 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-vj29g"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.020899 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.021328 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3fdb7434-0752-4484-b253-948bbde41112-serving-cert\") pod \"apiserver-7bbb656c7d-2jbbg\" (UID: \"3fdb7434-0752-4484-b253-948bbde41112\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2jbbg" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.021661 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3fdb7434-0752-4484-b253-948bbde41112-encryption-config\") pod \"apiserver-7bbb656c7d-2jbbg\" (UID: \"3fdb7434-0752-4484-b253-948bbde41112\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2jbbg" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.022046 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.022197 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/4ea98230-93e8-4ef6-86d6-e25b8448c632-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-b9772\" (UID: \"4ea98230-93e8-4ef6-86d6-e25b8448c632\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-b9772" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.024390 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.029836 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.030646 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd3b00b4-f42e-406c-9b88-b7e2602a2493-serving-cert\") pod \"route-controller-manager-6576b87f9c-frd4g\" (UID: \"dd3b00b4-f42e-406c-9b88-b7e2602a2493\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-frd4g" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.031685 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9d75460-0c2b-4c82-89d0-e41f8b8c275c-serving-cert\") pod \"apiserver-76f77b778f-6wfhz\" (UID: \"f9d75460-0c2b-4c82-89d0-e41f8b8c275c\") " pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.033615 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a833ce54-1145-4b0e-8136-d1747e52673b-serving-cert\") pod \"authentication-operator-69f744f599-cc8ct\" (UID: \"a833ce54-1145-4b0e-8136-d1747e52673b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cc8ct" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.036585 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3fdb7434-0752-4484-b253-948bbde41112-etcd-client\") pod \"apiserver-7bbb656c7d-2jbbg\" (UID: \"3fdb7434-0752-4484-b253-948bbde41112\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2jbbg" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.036676 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.037799 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-sqthq"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.038245 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/91eb41f7-eb34-496d-b65d-20dfc7abf7f0-serving-cert\") pod \"openshift-config-operator-7777fb866f-btxgl\" (UID: \"91eb41f7-eb34-496d-b65d-20dfc7abf7f0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-btxgl" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.038431 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/af75f830-5407-4eb7-ae1b-b09e6ef72737-machine-approver-tls\") pod \"machine-approver-56656f9798-94pws\" (UID: \"af75f830-5407-4eb7-ae1b-b09e6ef72737\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-94pws" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.047298 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-djxgw"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.047436 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sqthq" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.047623 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcngj"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.047698 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-djxgw" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.047892 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-72b2b"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.047959 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcngj" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.048308 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-s8rlm"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.048628 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-72b2b" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.048814 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p6lt7"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.049089 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-66kw4"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.049111 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-v4nr9"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.049123 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-b5tz8"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.049191 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p6lt7" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.049423 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-s8rlm" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.049772 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-cc8ct"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.051086 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-cm7vp"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.051482 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-cm7vp" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.051621 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-btxgl"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.052737 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-zxt8n"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.053802 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-w298r"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.054960 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rmtjp"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.055445 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-rmtjp" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.056503 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-gqjqp"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.056531 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.058204 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536864-kmfkv"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.058556 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-gqjqp" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.058965 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536860-dw5fg"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.059083 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536864-kmfkv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.059302 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536860-dw5fg" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.060552 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-l2nwx"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.062493 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-6wfhz"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.063579 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-jzpsv"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.065196 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sbqxk"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.066149 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39f2673f-09d0-4815-babb-a8f5961eee45-serving-cert\") pod \"console-operator-58897d9998-vj29g\" (UID: \"39f2673f-09d0-4815-babb-a8f5961eee45\") " pod="openshift-console-operator/console-operator-58897d9998-vj29g" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.067863 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b6vld"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.069065 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-nw2mh"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.071220 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hc86c"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.072436 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-kq9sb"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.073609 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-b9772"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.075406 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c7zb5"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.075586 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.076720 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-kq896"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.077785 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-djxgw"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.078785 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-n6ls6"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.079704 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-n6ls6" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.079863 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-k97gv"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.080838 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-2jbbg"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.081938 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-blwkv"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.082969 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-5svpj"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.083050 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-blwkv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.084020 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9mh7s"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.085013 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fvzmk"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.085981 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-sqthq"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.086980 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2blr9"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.088059 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rmtjp"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.089047 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-cm7vp"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.090155 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-s8rlm"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.091123 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536864-kmfkv"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.092215 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcngj"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.093554 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-72b2b"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.094656 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p6lt7"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.095885 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-blwkv"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.095902 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.096932 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-gqjqp"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.097944 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536860-dw5fg"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.098938 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-w6ccg"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.099527 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-w6ccg" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.100178 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-w6ccg"] Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.108412 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ac062b00-1a72-44b4-90c9-de1573e3e217-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-66kw4\" (UID: \"ac062b00-1a72-44b4-90c9-de1573e3e217\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-66kw4" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.108442 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/facc8439-0b4d-4228-95b2-d01d48ee1c53-console-serving-cert\") pod \"console-f9d7485db-kq9sb\" (UID: \"facc8439-0b4d-4228-95b2-d01d48ee1c53\") " pod="openshift-console/console-f9d7485db-kq9sb" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.108484 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ac062b00-1a72-44b4-90c9-de1573e3e217-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-66kw4\" (UID: \"ac062b00-1a72-44b4-90c9-de1573e3e217\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-66kw4" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.108508 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lp2mq\" (UniqueName: \"kubernetes.io/projected/f2257cba-c045-4865-aecf-03642416f2cb-kube-api-access-lp2mq\") pod \"openshift-apiserver-operator-796bbdcf4f-w298r\" (UID: \"f2257cba-c045-4865-aecf-03642416f2cb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-w298r" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.108532 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zp8mz\" (UniqueName: \"kubernetes.io/projected/26f6041e-592f-48c6-93e5-35517c775186-kube-api-access-zp8mz\") pod \"cluster-samples-operator-665b6dd947-b6vld\" (UID: \"26f6041e-592f-48c6-93e5-35517c775186\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b6vld" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.108560 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/facc8439-0b4d-4228-95b2-d01d48ee1c53-trusted-ca-bundle\") pod \"console-f9d7485db-kq9sb\" (UID: \"facc8439-0b4d-4228-95b2-d01d48ee1c53\") " pod="openshift-console/console-f9d7485db-kq9sb" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.108591 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/ac062b00-1a72-44b4-90c9-de1573e3e217-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-66kw4\" (UID: \"ac062b00-1a72-44b4-90c9-de1573e3e217\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-66kw4" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.108609 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/facc8439-0b4d-4228-95b2-d01d48ee1c53-console-oauth-config\") pod \"console-f9d7485db-kq9sb\" (UID: \"facc8439-0b4d-4228-95b2-d01d48ee1c53\") " pod="openshift-console/console-f9d7485db-kq9sb" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.108626 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/facc8439-0b4d-4228-95b2-d01d48ee1c53-service-ca\") pod \"console-f9d7485db-kq9sb\" (UID: \"facc8439-0b4d-4228-95b2-d01d48ee1c53\") " pod="openshift-console/console-f9d7485db-kq9sb" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.108659 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2257cba-c045-4865-aecf-03642416f2cb-config\") pod \"openshift-apiserver-operator-796bbdcf4f-w298r\" (UID: \"f2257cba-c045-4865-aecf-03642416f2cb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-w298r" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.108682 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f2257cba-c045-4865-aecf-03642416f2cb-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-w298r\" (UID: \"f2257cba-c045-4865-aecf-03642416f2cb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-w298r" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.108698 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/26f6041e-592f-48c6-93e5-35517c775186-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-b6vld\" (UID: \"26f6041e-592f-48c6-93e5-35517c775186\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b6vld" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.108722 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swq2d\" (UniqueName: \"kubernetes.io/projected/ac062b00-1a72-44b4-90c9-de1573e3e217-kube-api-access-swq2d\") pod \"cluster-image-registry-operator-dc59b4c8b-66kw4\" (UID: \"ac062b00-1a72-44b4-90c9-de1573e3e217\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-66kw4" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.108776 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/facc8439-0b4d-4228-95b2-d01d48ee1c53-oauth-serving-cert\") pod \"console-f9d7485db-kq9sb\" (UID: \"facc8439-0b4d-4228-95b2-d01d48ee1c53\") " pod="openshift-console/console-f9d7485db-kq9sb" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.109165 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6f4c\" (UniqueName: \"kubernetes.io/projected/facc8439-0b4d-4228-95b2-d01d48ee1c53-kube-api-access-d6f4c\") pod \"console-f9d7485db-kq9sb\" (UID: \"facc8439-0b4d-4228-95b2-d01d48ee1c53\") " pod="openshift-console/console-f9d7485db-kq9sb" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.109215 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/facc8439-0b4d-4228-95b2-d01d48ee1c53-console-config\") pod \"console-f9d7485db-kq9sb\" (UID: \"facc8439-0b4d-4228-95b2-d01d48ee1c53\") " pod="openshift-console/console-f9d7485db-kq9sb" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.109887 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2257cba-c045-4865-aecf-03642416f2cb-config\") pod \"openshift-apiserver-operator-796bbdcf4f-w298r\" (UID: \"f2257cba-c045-4865-aecf-03642416f2cb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-w298r" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.109919 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ac062b00-1a72-44b4-90c9-de1573e3e217-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-66kw4\" (UID: \"ac062b00-1a72-44b4-90c9-de1573e3e217\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-66kw4" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.110867 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/facc8439-0b4d-4228-95b2-d01d48ee1c53-oauth-serving-cert\") pod \"console-f9d7485db-kq9sb\" (UID: \"facc8439-0b4d-4228-95b2-d01d48ee1c53\") " pod="openshift-console/console-f9d7485db-kq9sb" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.110939 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/facc8439-0b4d-4228-95b2-d01d48ee1c53-console-config\") pod \"console-f9d7485db-kq9sb\" (UID: \"facc8439-0b4d-4228-95b2-d01d48ee1c53\") " pod="openshift-console/console-f9d7485db-kq9sb" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.111052 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/facc8439-0b4d-4228-95b2-d01d48ee1c53-service-ca\") pod \"console-f9d7485db-kq9sb\" (UID: \"facc8439-0b4d-4228-95b2-d01d48ee1c53\") " pod="openshift-console/console-f9d7485db-kq9sb" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.111177 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/facc8439-0b4d-4228-95b2-d01d48ee1c53-trusted-ca-bundle\") pod \"console-f9d7485db-kq9sb\" (UID: \"facc8439-0b4d-4228-95b2-d01d48ee1c53\") " pod="openshift-console/console-f9d7485db-kq9sb" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.112291 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/facc8439-0b4d-4228-95b2-d01d48ee1c53-console-oauth-config\") pod \"console-f9d7485db-kq9sb\" (UID: \"facc8439-0b4d-4228-95b2-d01d48ee1c53\") " pod="openshift-console/console-f9d7485db-kq9sb" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.112487 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/26f6041e-592f-48c6-93e5-35517c775186-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-b6vld\" (UID: \"26f6041e-592f-48c6-93e5-35517c775186\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b6vld" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.113267 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/facc8439-0b4d-4228-95b2-d01d48ee1c53-console-serving-cert\") pod \"console-f9d7485db-kq9sb\" (UID: \"facc8439-0b4d-4228-95b2-d01d48ee1c53\") " pod="openshift-console/console-f9d7485db-kq9sb" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.113722 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/ac062b00-1a72-44b4-90c9-de1573e3e217-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-66kw4\" (UID: \"ac062b00-1a72-44b4-90c9-de1573e3e217\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-66kw4" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.114932 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f2257cba-c045-4865-aecf-03642416f2cb-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-w298r\" (UID: \"f2257cba-c045-4865-aecf-03642416f2cb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-w298r" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.115613 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.135786 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.157264 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.176226 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.196259 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.216127 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.236955 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.255624 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.276381 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.296464 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.317289 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.336551 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.356845 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.377865 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.396085 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.418824 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.439043 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.458240 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.477117 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.496536 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.517404 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.536352 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.557769 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.577173 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.615673 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ee57d6e4-3550-46d0-947e-fe6db5cf3291-ca-trust-extracted\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.616054 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ee57d6e4-3550-46d0-947e-fe6db5cf3291-registry-certificates\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.616099 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ee57d6e4-3550-46d0-947e-fe6db5cf3291-trusted-ca\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.616167 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f84fe6b6-37ca-406b-a8dd-d2f40638c832-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-b5tz8\" (UID: \"f84fe6b6-37ca-406b-a8dd-d2f40638c832\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-b5tz8" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.616199 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skzrc\" (UniqueName: \"kubernetes.io/projected/f84fe6b6-37ca-406b-a8dd-d2f40638c832-kube-api-access-skzrc\") pod \"openshift-controller-manager-operator-756b6f6bc6-b5tz8\" (UID: \"f84fe6b6-37ca-406b-a8dd-d2f40638c832\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-b5tz8" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.616257 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ee57d6e4-3550-46d0-947e-fe6db5cf3291-installation-pull-secrets\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.616289 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6pzj\" (UniqueName: \"kubernetes.io/projected/ee57d6e4-3550-46d0-947e-fe6db5cf3291-kube-api-access-z6pzj\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.616381 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.616561 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.616985 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ee57d6e4-3550-46d0-947e-fe6db5cf3291-registry-tls\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.617029 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ee57d6e4-3550-46d0-947e-fe6db5cf3291-bound-sa-token\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:05 crc kubenswrapper[4700]: E0227 17:04:05.617082 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:06.11706284 +0000 UTC m=+206.102375627 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.617156 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f84fe6b6-37ca-406b-a8dd-d2f40638c832-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-b5tz8\" (UID: \"f84fe6b6-37ca-406b-a8dd-d2f40638c832\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-b5tz8" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.637024 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.656288 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.676037 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.697114 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.716094 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.718312 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:05 crc kubenswrapper[4700]: E0227 17:04:05.718599 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:06.218566285 +0000 UTC m=+206.203879072 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.718710 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ee57d6e4-3550-46d0-947e-fe6db5cf3291-bound-sa-token\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.718816 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/8eae2362-d971-4645-a71e-7bf6634a536d-etcd-ca\") pod \"etcd-operator-b45778765-l2nwx\" (UID: \"8eae2362-d971-4645-a71e-7bf6634a536d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-l2nwx" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.718859 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a2a9e61-d9b7-4891-8771-76137ccd27e1-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-fvzmk\" (UID: \"0a2a9e61-d9b7-4891-8771-76137ccd27e1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fvzmk" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.718951 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/cce840a6-7988-43de-a6d7-07187abb65c3-default-certificate\") pod \"router-default-5444994796-58xmr\" (UID: \"cce840a6-7988-43de-a6d7-07187abb65c3\") " pod="openshift-ingress/router-default-5444994796-58xmr" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.719775 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsszp\" (UniqueName: \"kubernetes.io/projected/7fb5109d-a02b-453b-abac-9f0637d9e8fc-kube-api-access-fsszp\") pod \"dns-operator-744455d44c-jzpsv\" (UID: \"7fb5109d-a02b-453b-abac-9f0637d9e8fc\") " pod="openshift-dns-operator/dns-operator-744455d44c-jzpsv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.719860 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c80e05b3-583f-4960-a0cc-351ff9ed3084-profile-collector-cert\") pod \"catalog-operator-68c6474976-djxgw\" (UID: \"c80e05b3-583f-4960-a0cc-351ff9ed3084\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-djxgw" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.719903 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2a528ef-fbc8-4429-a4dd-f330c594fa4a-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-9mh7s\" (UID: \"c2a528ef-fbc8-4429-a4dd-f330c594fa4a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9mh7s" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.719937 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/543c318a-e2c5-4e19-8746-333817f1e361-srv-cert\") pod \"olm-operator-6b444d44fb-kcngj\" (UID: \"543c318a-e2c5-4e19-8746-333817f1e361\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcngj" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.720079 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/483ac8b1-c17c-456d-9749-2ab085230e7a-signing-key\") pod \"service-ca-9c57cc56f-rmtjp\" (UID: \"483ac8b1-c17c-456d-9749-2ab085230e7a\") " pod="openshift-service-ca/service-ca-9c57cc56f-rmtjp" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.720119 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/173cb70c-93b2-4d78-a3df-a524bf604c37-webhook-cert\") pod \"packageserver-d55dfcdfc-p6lt7\" (UID: \"173cb70c-93b2-4d78-a3df-a524bf604c37\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p6lt7" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.720184 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/09f36818-91cb-4ce3-bd5f-ccaf37766551-metrics-tls\") pod \"dns-default-gqjqp\" (UID: \"09f36818-91cb-4ce3-bd5f-ccaf37766551\") " pod="openshift-dns/dns-default-gqjqp" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.720281 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/f3526e22-b6e6-4a3a-95ec-67a831e5678a-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-sbqxk\" (UID: \"f3526e22-b6e6-4a3a-95ec-67a831e5678a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sbqxk" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.720319 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f3c625f5-89ca-4691-ab34-3b31516aecf6-metrics-tls\") pod \"ingress-operator-5b745b69d9-5svpj\" (UID: \"f3c625f5-89ca-4691-ab34-3b31516aecf6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5svpj" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.720373 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3fbc309b-7d6b-4598-8854-c07a038d15bd-serving-cert\") pod \"service-ca-operator-777779d784-cm7vp\" (UID: \"3fbc309b-7d6b-4598-8854-c07a038d15bd\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cm7vp" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.720415 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ba36da60-1e6a-4010-a017-802d1cbbae71-csi-data-dir\") pod \"csi-hostpathplugin-blwkv\" (UID: \"ba36da60-1e6a-4010-a017-802d1cbbae71\") " pod="hostpath-provisioner/csi-hostpathplugin-blwkv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.720502 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ee57d6e4-3550-46d0-947e-fe6db5cf3291-registry-certificates\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.720565 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ee57d6e4-3550-46d0-947e-fe6db5cf3291-trusted-ca\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.722180 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e17a6bea-8ec8-40db-a566-44feacd2c584-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2blr9\" (UID: \"e17a6bea-8ec8-40db-a566-44feacd2c584\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2blr9" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.722407 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f84fe6b6-37ca-406b-a8dd-d2f40638c832-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-b5tz8\" (UID: \"f84fe6b6-37ca-406b-a8dd-d2f40638c832\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-b5tz8" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.722991 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skzrc\" (UniqueName: \"kubernetes.io/projected/f84fe6b6-37ca-406b-a8dd-d2f40638c832-kube-api-access-skzrc\") pod \"openshift-controller-manager-operator-756b6f6bc6-b5tz8\" (UID: \"f84fe6b6-37ca-406b-a8dd-d2f40638c832\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-b5tz8" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.723096 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ba36da60-1e6a-4010-a017-802d1cbbae71-plugins-dir\") pod \"csi-hostpathplugin-blwkv\" (UID: \"ba36da60-1e6a-4010-a017-802d1cbbae71\") " pod="hostpath-provisioner/csi-hostpathplugin-blwkv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.723251 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/173cb70c-93b2-4d78-a3df-a524bf604c37-apiservice-cert\") pod \"packageserver-d55dfcdfc-p6lt7\" (UID: \"173cb70c-93b2-4d78-a3df-a524bf604c37\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p6lt7" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.723358 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98wkm\" (UniqueName: \"kubernetes.io/projected/39da2f34-cd0c-460e-9b62-0e6662251261-kube-api-access-98wkm\") pod \"machine-config-server-n6ls6\" (UID: \"39da2f34-cd0c-460e-9b62-0e6662251261\") " pod="openshift-machine-config-operator/machine-config-server-n6ls6" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.723952 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8eae2362-d971-4645-a71e-7bf6634a536d-etcd-client\") pod \"etcd-operator-b45778765-l2nwx\" (UID: \"8eae2362-d971-4645-a71e-7bf6634a536d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-l2nwx" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.724442 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ee57d6e4-3550-46d0-947e-fe6db5cf3291-trusted-ca\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.724523 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cce840a6-7988-43de-a6d7-07187abb65c3-metrics-certs\") pod \"router-default-5444994796-58xmr\" (UID: \"cce840a6-7988-43de-a6d7-07187abb65c3\") " pod="openshift-ingress/router-default-5444994796-58xmr" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.724640 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/39da2f34-cd0c-460e-9b62-0e6662251261-certs\") pod \"machine-config-server-n6ls6\" (UID: \"39da2f34-cd0c-460e-9b62-0e6662251261\") " pod="openshift-machine-config-operator/machine-config-server-n6ls6" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.724714 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/8eae2362-d971-4645-a71e-7bf6634a536d-etcd-service-ca\") pod \"etcd-operator-b45778765-l2nwx\" (UID: \"8eae2362-d971-4645-a71e-7bf6634a536d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-l2nwx" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.725405 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/66e8ed9b-99d5-4643-aa2a-4720ea4195b9-config-volume\") pod \"collect-profiles-29536860-dw5fg\" (UID: \"66e8ed9b-99d5-4643-aa2a-4720ea4195b9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536860-dw5fg" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.725503 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/483ac8b1-c17c-456d-9749-2ab085230e7a-signing-cabundle\") pod \"service-ca-9c57cc56f-rmtjp\" (UID: \"483ac8b1-c17c-456d-9749-2ab085230e7a\") " pod="openshift-service-ca/service-ca-9c57cc56f-rmtjp" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.725589 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3fbc309b-7d6b-4598-8854-c07a038d15bd-config\") pod \"service-ca-operator-777779d784-cm7vp\" (UID: \"3fbc309b-7d6b-4598-8854-c07a038d15bd\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cm7vp" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.725629 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/39da2f34-cd0c-460e-9b62-0e6662251261-node-bootstrap-token\") pod \"machine-config-server-n6ls6\" (UID: \"39da2f34-cd0c-460e-9b62-0e6662251261\") " pod="openshift-machine-config-operator/machine-config-server-n6ls6" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.725664 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ba36da60-1e6a-4010-a017-802d1cbbae71-socket-dir\") pod \"csi-hostpathplugin-blwkv\" (UID: \"ba36da60-1e6a-4010-a017-802d1cbbae71\") " pod="hostpath-provisioner/csi-hostpathplugin-blwkv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.725702 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnh2k\" (UniqueName: \"kubernetes.io/projected/173cb70c-93b2-4d78-a3df-a524bf604c37-kube-api-access-hnh2k\") pod \"packageserver-d55dfcdfc-p6lt7\" (UID: \"173cb70c-93b2-4d78-a3df-a524bf604c37\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p6lt7" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.725750 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.725792 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0008dd34-6337-445e-9db2-2b662d682a68-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-c7zb5\" (UID: \"0008dd34-6337-445e-9db2-2b662d682a68\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c7zb5" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.725846 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0cc33e9e-5984-454d-880c-5187145a2c36-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hc86c\" (UID: \"0cc33e9e-5984-454d-880c-5187145a2c36\") " pod="openshift-marketplace/marketplace-operator-79b997595-hc86c" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.725915 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hc4kp\" (UniqueName: \"kubernetes.io/projected/0a2a9e61-d9b7-4891-8771-76137ccd27e1-kube-api-access-hc4kp\") pod \"kube-storage-version-migrator-operator-b67b599dd-fvzmk\" (UID: \"0a2a9e61-d9b7-4891-8771-76137ccd27e1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fvzmk" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.725961 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/749f12f7-ef9a-449d-a276-85ee4afe830e-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-nw2mh\" (UID: \"749f12f7-ef9a-449d-a276-85ee4afe830e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-nw2mh" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.726025 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c80e05b3-583f-4960-a0cc-351ff9ed3084-srv-cert\") pod \"catalog-operator-68c6474976-djxgw\" (UID: \"c80e05b3-583f-4960-a0cc-351ff9ed3084\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-djxgw" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.726161 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6skm\" (UniqueName: \"kubernetes.io/projected/f3526e22-b6e6-4a3a-95ec-67a831e5678a-kube-api-access-m6skm\") pod \"control-plane-machine-set-operator-78cbb6b69f-sbqxk\" (UID: \"f3526e22-b6e6-4a3a-95ec-67a831e5678a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sbqxk" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.726230 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09f36818-91cb-4ce3-bd5f-ccaf37766551-config-volume\") pod \"dns-default-gqjqp\" (UID: \"09f36818-91cb-4ce3-bd5f-ccaf37766551\") " pod="openshift-dns/dns-default-gqjqp" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.726273 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ba36da60-1e6a-4010-a017-802d1cbbae71-mountpoint-dir\") pod \"csi-hostpathplugin-blwkv\" (UID: \"ba36da60-1e6a-4010-a017-802d1cbbae71\") " pod="hostpath-provisioner/csi-hostpathplugin-blwkv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.726309 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8eae2362-d971-4645-a71e-7bf6634a536d-serving-cert\") pod \"etcd-operator-b45778765-l2nwx\" (UID: \"8eae2362-d971-4645-a71e-7bf6634a536d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-l2nwx" Feb 27 17:04:05 crc kubenswrapper[4700]: E0227 17:04:05.726548 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:06.226521384 +0000 UTC m=+206.211834171 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.726631 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7jtz\" (UniqueName: \"kubernetes.io/projected/e17a6bea-8ec8-40db-a566-44feacd2c584-kube-api-access-g7jtz\") pod \"machine-config-controller-84d6567774-2blr9\" (UID: \"e17a6bea-8ec8-40db-a566-44feacd2c584\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2blr9" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.726721 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/25827d09-2429-43d7-95a0-82279fd15a17-cert\") pod \"ingress-canary-w6ccg\" (UID: \"25827d09-2429-43d7-95a0-82279fd15a17\") " pod="openshift-ingress-canary/ingress-canary-w6ccg" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.726784 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0008dd34-6337-445e-9db2-2b662d682a68-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-c7zb5\" (UID: \"0008dd34-6337-445e-9db2-2b662d682a68\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c7zb5" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.726874 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lltk4\" (UniqueName: \"kubernetes.io/projected/8eae2362-d971-4645-a71e-7bf6634a536d-kube-api-access-lltk4\") pod \"etcd-operator-b45778765-l2nwx\" (UID: \"8eae2362-d971-4645-a71e-7bf6634a536d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-l2nwx" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.726911 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7fb5109d-a02b-453b-abac-9f0637d9e8fc-metrics-tls\") pod \"dns-operator-744455d44c-jzpsv\" (UID: \"7fb5109d-a02b-453b-abac-9f0637d9e8fc\") " pod="openshift-dns-operator/dns-operator-744455d44c-jzpsv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.726951 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ee57d6e4-3550-46d0-947e-fe6db5cf3291-registry-tls\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.726989 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0a2a9e61-d9b7-4891-8771-76137ccd27e1-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-fvzmk\" (UID: \"0a2a9e61-d9b7-4891-8771-76137ccd27e1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fvzmk" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.727052 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pw7l\" (UniqueName: \"kubernetes.io/projected/1c2bf6c9-eba6-499e-b214-c2336431714a-kube-api-access-8pw7l\") pod \"migrator-59844c95c7-sqthq\" (UID: \"1c2bf6c9-eba6-499e-b214-c2336431714a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sqthq" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.725917 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ee57d6e4-3550-46d0-947e-fe6db5cf3291-registry-certificates\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.727170 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/cce840a6-7988-43de-a6d7-07187abb65c3-stats-auth\") pod \"router-default-5444994796-58xmr\" (UID: \"cce840a6-7988-43de-a6d7-07187abb65c3\") " pod="openshift-ingress/router-default-5444994796-58xmr" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.727261 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/eceb2456-36c4-4f96-83d8-495b3eb834a8-images\") pod \"machine-config-operator-74547568cd-kq896\" (UID: \"eceb2456-36c4-4f96-83d8-495b3eb834a8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kq896" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.727297 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ba36da60-1e6a-4010-a017-802d1cbbae71-registration-dir\") pod \"csi-hostpathplugin-blwkv\" (UID: \"ba36da60-1e6a-4010-a017-802d1cbbae71\") " pod="hostpath-provisioner/csi-hostpathplugin-blwkv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.727366 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f84fe6b6-37ca-406b-a8dd-d2f40638c832-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-b5tz8\" (UID: \"f84fe6b6-37ca-406b-a8dd-d2f40638c832\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-b5tz8" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.728443 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f84fe6b6-37ca-406b-a8dd-d2f40638c832-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-b5tz8\" (UID: \"f84fe6b6-37ca-406b-a8dd-d2f40638c832\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-b5tz8" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.728706 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n99md\" (UniqueName: \"kubernetes.io/projected/f3c625f5-89ca-4691-ab34-3b31516aecf6-kube-api-access-n99md\") pod \"ingress-operator-5b745b69d9-5svpj\" (UID: \"f3c625f5-89ca-4691-ab34-3b31516aecf6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5svpj" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.728913 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ee57d6e4-3550-46d0-947e-fe6db5cf3291-ca-trust-extracted\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.729160 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/749f12f7-ef9a-449d-a276-85ee4afe830e-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-nw2mh\" (UID: \"749f12f7-ef9a-449d-a276-85ee4afe830e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-nw2mh" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.729250 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c2a528ef-fbc8-4429-a4dd-f330c594fa4a-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-9mh7s\" (UID: \"c2a528ef-fbc8-4429-a4dd-f330c594fa4a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9mh7s" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.729304 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8dc9b0d3-c9e6-4532-bbea-0a29077b7eab-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-72b2b\" (UID: \"8dc9b0d3-c9e6-4532-bbea-0a29077b7eab\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-72b2b" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.729488 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8eae2362-d971-4645-a71e-7bf6634a536d-config\") pod \"etcd-operator-b45778765-l2nwx\" (UID: \"8eae2362-d971-4645-a71e-7bf6634a536d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-l2nwx" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.729624 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pknc\" (UniqueName: \"kubernetes.io/projected/ba36da60-1e6a-4010-a017-802d1cbbae71-kube-api-access-4pknc\") pod \"csi-hostpathplugin-blwkv\" (UID: \"ba36da60-1e6a-4010-a017-802d1cbbae71\") " pod="hostpath-provisioner/csi-hostpathplugin-blwkv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.729687 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbvwg\" (UniqueName: \"kubernetes.io/projected/09f36818-91cb-4ce3-bd5f-ccaf37766551-kube-api-access-nbvwg\") pod \"dns-default-gqjqp\" (UID: \"09f36818-91cb-4ce3-bd5f-ccaf37766551\") " pod="openshift-dns/dns-default-gqjqp" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.729728 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f3c625f5-89ca-4691-ab34-3b31516aecf6-bound-sa-token\") pod \"ingress-operator-5b745b69d9-5svpj\" (UID: \"f3c625f5-89ca-4691-ab34-3b31516aecf6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5svpj" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.729804 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ee57d6e4-3550-46d0-947e-fe6db5cf3291-ca-trust-extracted\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.729809 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0cc33e9e-5984-454d-880c-5187145a2c36-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hc86c\" (UID: \"0cc33e9e-5984-454d-880c-5187145a2c36\") " pod="openshift-marketplace/marketplace-operator-79b997595-hc86c" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.729946 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9zcl\" (UniqueName: \"kubernetes.io/projected/66e8ed9b-99d5-4643-aa2a-4720ea4195b9-kube-api-access-h9zcl\") pod \"collect-profiles-29536860-dw5fg\" (UID: \"66e8ed9b-99d5-4643-aa2a-4720ea4195b9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536860-dw5fg" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.729986 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/eceb2456-36c4-4f96-83d8-495b3eb834a8-proxy-tls\") pod \"machine-config-operator-74547568cd-kq896\" (UID: \"eceb2456-36c4-4f96-83d8-495b3eb834a8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kq896" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.730063 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n962r\" (UniqueName: \"kubernetes.io/projected/3166d7d3-7842-4655-b10b-24e5731d77e0-kube-api-access-n962r\") pod \"auto-csr-approver-29536864-kmfkv\" (UID: \"3166d7d3-7842-4655-b10b-24e5731d77e0\") " pod="openshift-infra/auto-csr-approver-29536864-kmfkv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.730119 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lgjw\" (UniqueName: \"kubernetes.io/projected/eceb2456-36c4-4f96-83d8-495b3eb834a8-kube-api-access-9lgjw\") pod \"machine-config-operator-74547568cd-kq896\" (UID: \"eceb2456-36c4-4f96-83d8-495b3eb834a8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kq896" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.730216 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbdkx\" (UniqueName: \"kubernetes.io/projected/cce840a6-7988-43de-a6d7-07187abb65c3-kube-api-access-kbdkx\") pod \"router-default-5444994796-58xmr\" (UID: \"cce840a6-7988-43de-a6d7-07187abb65c3\") " pod="openshift-ingress/router-default-5444994796-58xmr" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.730386 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2a528ef-fbc8-4429-a4dd-f330c594fa4a-config\") pod \"kube-apiserver-operator-766d6c64bb-9mh7s\" (UID: \"c2a528ef-fbc8-4429-a4dd-f330c594fa4a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9mh7s" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.730438 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qphmr\" (UniqueName: \"kubernetes.io/projected/0cc33e9e-5984-454d-880c-5187145a2c36-kube-api-access-qphmr\") pod \"marketplace-operator-79b997595-hc86c\" (UID: \"0cc33e9e-5984-454d-880c-5187145a2c36\") " pod="openshift-marketplace/marketplace-operator-79b997595-hc86c" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.730537 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtzjd\" (UniqueName: \"kubernetes.io/projected/87dc0d93-e849-4b96-a1dd-1abaf03dd7cb-kube-api-access-jtzjd\") pod \"multus-admission-controller-857f4d67dd-s8rlm\" (UID: \"87dc0d93-e849-4b96-a1dd-1abaf03dd7cb\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-s8rlm" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.730596 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6pzj\" (UniqueName: \"kubernetes.io/projected/ee57d6e4-3550-46d0-947e-fe6db5cf3291-kube-api-access-z6pzj\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.730637 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zh67h\" (UniqueName: \"kubernetes.io/projected/25827d09-2429-43d7-95a0-82279fd15a17-kube-api-access-zh67h\") pod \"ingress-canary-w6ccg\" (UID: \"25827d09-2429-43d7-95a0-82279fd15a17\") " pod="openshift-ingress-canary/ingress-canary-w6ccg" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.730803 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ee57d6e4-3550-46d0-947e-fe6db5cf3291-installation-pull-secrets\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.730865 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lch6n\" (UniqueName: \"kubernetes.io/projected/3fbc309b-7d6b-4598-8854-c07a038d15bd-kube-api-access-lch6n\") pod \"service-ca-operator-777779d784-cm7vp\" (UID: \"3fbc309b-7d6b-4598-8854-c07a038d15bd\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cm7vp" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.730999 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f3c625f5-89ca-4691-ab34-3b31516aecf6-trusted-ca\") pod \"ingress-operator-5b745b69d9-5svpj\" (UID: \"f3c625f5-89ca-4691-ab34-3b31516aecf6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5svpj" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.731041 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/87dc0d93-e849-4b96-a1dd-1abaf03dd7cb-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-s8rlm\" (UID: \"87dc0d93-e849-4b96-a1dd-1abaf03dd7cb\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-s8rlm" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.731078 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/66e8ed9b-99d5-4643-aa2a-4720ea4195b9-secret-volume\") pod \"collect-profiles-29536860-dw5fg\" (UID: \"66e8ed9b-99d5-4643-aa2a-4720ea4195b9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536860-dw5fg" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.731114 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dn8cv\" (UniqueName: \"kubernetes.io/projected/543c318a-e2c5-4e19-8746-333817f1e361-kube-api-access-dn8cv\") pod \"olm-operator-6b444d44fb-kcngj\" (UID: \"543c318a-e2c5-4e19-8746-333817f1e361\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcngj" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.731148 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e17a6bea-8ec8-40db-a566-44feacd2c584-proxy-tls\") pod \"machine-config-controller-84d6567774-2blr9\" (UID: \"e17a6bea-8ec8-40db-a566-44feacd2c584\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2blr9" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.731186 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/749f12f7-ef9a-449d-a276-85ee4afe830e-config\") pod \"kube-controller-manager-operator-78b949d7b-nw2mh\" (UID: \"749f12f7-ef9a-449d-a276-85ee4afe830e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-nw2mh" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.731274 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f84fe6b6-37ca-406b-a8dd-d2f40638c832-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-b5tz8\" (UID: \"f84fe6b6-37ca-406b-a8dd-d2f40638c832\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-b5tz8" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.731326 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/543c318a-e2c5-4e19-8746-333817f1e361-profile-collector-cert\") pod \"olm-operator-6b444d44fb-kcngj\" (UID: \"543c318a-e2c5-4e19-8746-333817f1e361\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcngj" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.731437 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cce840a6-7988-43de-a6d7-07187abb65c3-service-ca-bundle\") pod \"router-default-5444994796-58xmr\" (UID: \"cce840a6-7988-43de-a6d7-07187abb65c3\") " pod="openshift-ingress/router-default-5444994796-58xmr" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.731689 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0008dd34-6337-445e-9db2-2b662d682a68-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-c7zb5\" (UID: \"0008dd34-6337-445e-9db2-2b662d682a68\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c7zb5" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.731733 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4786h\" (UniqueName: \"kubernetes.io/projected/8dc9b0d3-c9e6-4532-bbea-0a29077b7eab-kube-api-access-4786h\") pod \"package-server-manager-789f6589d5-72b2b\" (UID: \"8dc9b0d3-c9e6-4532-bbea-0a29077b7eab\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-72b2b" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.731774 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/eceb2456-36c4-4f96-83d8-495b3eb834a8-auth-proxy-config\") pod \"machine-config-operator-74547568cd-kq896\" (UID: \"eceb2456-36c4-4f96-83d8-495b3eb834a8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kq896" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.731932 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/173cb70c-93b2-4d78-a3df-a524bf604c37-tmpfs\") pod \"packageserver-d55dfcdfc-p6lt7\" (UID: \"173cb70c-93b2-4d78-a3df-a524bf604c37\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p6lt7" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.731980 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtn8v\" (UniqueName: \"kubernetes.io/projected/c80e05b3-583f-4960-a0cc-351ff9ed3084-kube-api-access-vtn8v\") pod \"catalog-operator-68c6474976-djxgw\" (UID: \"c80e05b3-583f-4960-a0cc-351ff9ed3084\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-djxgw" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.732330 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ee57d6e4-3550-46d0-947e-fe6db5cf3291-registry-tls\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.732600 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgwl5\" (UniqueName: \"kubernetes.io/projected/483ac8b1-c17c-456d-9749-2ab085230e7a-kube-api-access-zgwl5\") pod \"service-ca-9c57cc56f-rmtjp\" (UID: \"483ac8b1-c17c-456d-9749-2ab085230e7a\") " pod="openshift-service-ca/service-ca-9c57cc56f-rmtjp" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.734993 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ee57d6e4-3550-46d0-947e-fe6db5cf3291-installation-pull-secrets\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.737488 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.756550 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.776727 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.797314 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.816746 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.834510 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:05 crc kubenswrapper[4700]: E0227 17:04:05.834905 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:06.334813596 +0000 UTC m=+206.320126373 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.835140 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/8eae2362-d971-4645-a71e-7bf6634a536d-etcd-ca\") pod \"etcd-operator-b45778765-l2nwx\" (UID: \"8eae2362-d971-4645-a71e-7bf6634a536d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-l2nwx" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.835186 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a2a9e61-d9b7-4891-8771-76137ccd27e1-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-fvzmk\" (UID: \"0a2a9e61-d9b7-4891-8771-76137ccd27e1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fvzmk" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.835229 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/cce840a6-7988-43de-a6d7-07187abb65c3-default-certificate\") pod \"router-default-5444994796-58xmr\" (UID: \"cce840a6-7988-43de-a6d7-07187abb65c3\") " pod="openshift-ingress/router-default-5444994796-58xmr" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.835269 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsszp\" (UniqueName: \"kubernetes.io/projected/7fb5109d-a02b-453b-abac-9f0637d9e8fc-kube-api-access-fsszp\") pod \"dns-operator-744455d44c-jzpsv\" (UID: \"7fb5109d-a02b-453b-abac-9f0637d9e8fc\") " pod="openshift-dns-operator/dns-operator-744455d44c-jzpsv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.835300 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c80e05b3-583f-4960-a0cc-351ff9ed3084-profile-collector-cert\") pod \"catalog-operator-68c6474976-djxgw\" (UID: \"c80e05b3-583f-4960-a0cc-351ff9ed3084\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-djxgw" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.835333 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/543c318a-e2c5-4e19-8746-333817f1e361-srv-cert\") pod \"olm-operator-6b444d44fb-kcngj\" (UID: \"543c318a-e2c5-4e19-8746-333817f1e361\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcngj" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.835393 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2a528ef-fbc8-4429-a4dd-f330c594fa4a-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-9mh7s\" (UID: \"c2a528ef-fbc8-4429-a4dd-f330c594fa4a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9mh7s" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.835477 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/483ac8b1-c17c-456d-9749-2ab085230e7a-signing-key\") pod \"service-ca-9c57cc56f-rmtjp\" (UID: \"483ac8b1-c17c-456d-9749-2ab085230e7a\") " pod="openshift-service-ca/service-ca-9c57cc56f-rmtjp" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.835509 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/173cb70c-93b2-4d78-a3df-a524bf604c37-webhook-cert\") pod \"packageserver-d55dfcdfc-p6lt7\" (UID: \"173cb70c-93b2-4d78-a3df-a524bf604c37\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p6lt7" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.835584 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/09f36818-91cb-4ce3-bd5f-ccaf37766551-metrics-tls\") pod \"dns-default-gqjqp\" (UID: \"09f36818-91cb-4ce3-bd5f-ccaf37766551\") " pod="openshift-dns/dns-default-gqjqp" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.836579 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f3c625f5-89ca-4691-ab34-3b31516aecf6-metrics-tls\") pod \"ingress-operator-5b745b69d9-5svpj\" (UID: \"f3c625f5-89ca-4691-ab34-3b31516aecf6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5svpj" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.836678 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/f3526e22-b6e6-4a3a-95ec-67a831e5678a-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-sbqxk\" (UID: \"f3526e22-b6e6-4a3a-95ec-67a831e5678a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sbqxk" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.836759 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3fbc309b-7d6b-4598-8854-c07a038d15bd-serving-cert\") pod \"service-ca-operator-777779d784-cm7vp\" (UID: \"3fbc309b-7d6b-4598-8854-c07a038d15bd\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cm7vp" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.836864 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ba36da60-1e6a-4010-a017-802d1cbbae71-csi-data-dir\") pod \"csi-hostpathplugin-blwkv\" (UID: \"ba36da60-1e6a-4010-a017-802d1cbbae71\") " pod="hostpath-provisioner/csi-hostpathplugin-blwkv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.836934 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e17a6bea-8ec8-40db-a566-44feacd2c584-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2blr9\" (UID: \"e17a6bea-8ec8-40db-a566-44feacd2c584\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2blr9" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.837021 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ba36da60-1e6a-4010-a017-802d1cbbae71-plugins-dir\") pod \"csi-hostpathplugin-blwkv\" (UID: \"ba36da60-1e6a-4010-a017-802d1cbbae71\") " pod="hostpath-provisioner/csi-hostpathplugin-blwkv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.837090 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/173cb70c-93b2-4d78-a3df-a524bf604c37-apiservice-cert\") pod \"packageserver-d55dfcdfc-p6lt7\" (UID: \"173cb70c-93b2-4d78-a3df-a524bf604c37\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p6lt7" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.837140 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8eae2362-d971-4645-a71e-7bf6634a536d-etcd-client\") pod \"etcd-operator-b45778765-l2nwx\" (UID: \"8eae2362-d971-4645-a71e-7bf6634a536d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-l2nwx" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.837158 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.837177 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98wkm\" (UniqueName: \"kubernetes.io/projected/39da2f34-cd0c-460e-9b62-0e6662251261-kube-api-access-98wkm\") pod \"machine-config-server-n6ls6\" (UID: \"39da2f34-cd0c-460e-9b62-0e6662251261\") " pod="openshift-machine-config-operator/machine-config-server-n6ls6" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.837209 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cce840a6-7988-43de-a6d7-07187abb65c3-metrics-certs\") pod \"router-default-5444994796-58xmr\" (UID: \"cce840a6-7988-43de-a6d7-07187abb65c3\") " pod="openshift-ingress/router-default-5444994796-58xmr" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.837240 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/39da2f34-cd0c-460e-9b62-0e6662251261-certs\") pod \"machine-config-server-n6ls6\" (UID: \"39da2f34-cd0c-460e-9b62-0e6662251261\") " pod="openshift-machine-config-operator/machine-config-server-n6ls6" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.837265 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/8eae2362-d971-4645-a71e-7bf6634a536d-etcd-service-ca\") pod \"etcd-operator-b45778765-l2nwx\" (UID: \"8eae2362-d971-4645-a71e-7bf6634a536d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-l2nwx" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.837298 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/66e8ed9b-99d5-4643-aa2a-4720ea4195b9-config-volume\") pod \"collect-profiles-29536860-dw5fg\" (UID: \"66e8ed9b-99d5-4643-aa2a-4720ea4195b9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536860-dw5fg" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.837326 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/483ac8b1-c17c-456d-9749-2ab085230e7a-signing-cabundle\") pod \"service-ca-9c57cc56f-rmtjp\" (UID: \"483ac8b1-c17c-456d-9749-2ab085230e7a\") " pod="openshift-service-ca/service-ca-9c57cc56f-rmtjp" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.837350 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3fbc309b-7d6b-4598-8854-c07a038d15bd-config\") pod \"service-ca-operator-777779d784-cm7vp\" (UID: \"3fbc309b-7d6b-4598-8854-c07a038d15bd\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cm7vp" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.837374 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/39da2f34-cd0c-460e-9b62-0e6662251261-node-bootstrap-token\") pod \"machine-config-server-n6ls6\" (UID: \"39da2f34-cd0c-460e-9b62-0e6662251261\") " pod="openshift-machine-config-operator/machine-config-server-n6ls6" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.837404 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.837161 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ba36da60-1e6a-4010-a017-802d1cbbae71-csi-data-dir\") pod \"csi-hostpathplugin-blwkv\" (UID: \"ba36da60-1e6a-4010-a017-802d1cbbae71\") " pod="hostpath-provisioner/csi-hostpathplugin-blwkv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.837496 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0008dd34-6337-445e-9db2-2b662d682a68-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-c7zb5\" (UID: \"0008dd34-6337-445e-9db2-2b662d682a68\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c7zb5" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.837445 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ba36da60-1e6a-4010-a017-802d1cbbae71-plugins-dir\") pod \"csi-hostpathplugin-blwkv\" (UID: \"ba36da60-1e6a-4010-a017-802d1cbbae71\") " pod="hostpath-provisioner/csi-hostpathplugin-blwkv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.837603 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ba36da60-1e6a-4010-a017-802d1cbbae71-socket-dir\") pod \"csi-hostpathplugin-blwkv\" (UID: \"ba36da60-1e6a-4010-a017-802d1cbbae71\") " pod="hostpath-provisioner/csi-hostpathplugin-blwkv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.837703 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ba36da60-1e6a-4010-a017-802d1cbbae71-socket-dir\") pod \"csi-hostpathplugin-blwkv\" (UID: \"ba36da60-1e6a-4010-a017-802d1cbbae71\") " pod="hostpath-provisioner/csi-hostpathplugin-blwkv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.837711 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnh2k\" (UniqueName: \"kubernetes.io/projected/173cb70c-93b2-4d78-a3df-a524bf604c37-kube-api-access-hnh2k\") pod \"packageserver-d55dfcdfc-p6lt7\" (UID: \"173cb70c-93b2-4d78-a3df-a524bf604c37\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p6lt7" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.837889 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0cc33e9e-5984-454d-880c-5187145a2c36-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hc86c\" (UID: \"0cc33e9e-5984-454d-880c-5187145a2c36\") " pod="openshift-marketplace/marketplace-operator-79b997595-hc86c" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.837963 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hc4kp\" (UniqueName: \"kubernetes.io/projected/0a2a9e61-d9b7-4891-8771-76137ccd27e1-kube-api-access-hc4kp\") pod \"kube-storage-version-migrator-operator-b67b599dd-fvzmk\" (UID: \"0a2a9e61-d9b7-4891-8771-76137ccd27e1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fvzmk" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.838011 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/749f12f7-ef9a-449d-a276-85ee4afe830e-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-nw2mh\" (UID: \"749f12f7-ef9a-449d-a276-85ee4afe830e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-nw2mh" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.838045 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c80e05b3-583f-4960-a0cc-351ff9ed3084-srv-cert\") pod \"catalog-operator-68c6474976-djxgw\" (UID: \"c80e05b3-583f-4960-a0cc-351ff9ed3084\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-djxgw" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.838098 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6skm\" (UniqueName: \"kubernetes.io/projected/f3526e22-b6e6-4a3a-95ec-67a831e5678a-kube-api-access-m6skm\") pod \"control-plane-machine-set-operator-78cbb6b69f-sbqxk\" (UID: \"f3526e22-b6e6-4a3a-95ec-67a831e5678a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sbqxk" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.838134 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09f36818-91cb-4ce3-bd5f-ccaf37766551-config-volume\") pod \"dns-default-gqjqp\" (UID: \"09f36818-91cb-4ce3-bd5f-ccaf37766551\") " pod="openshift-dns/dns-default-gqjqp" Feb 27 17:04:05 crc kubenswrapper[4700]: E0227 17:04:05.838172 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:06.338146263 +0000 UTC m=+206.323459040 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.838225 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8eae2362-d971-4645-a71e-7bf6634a536d-serving-cert\") pod \"etcd-operator-b45778765-l2nwx\" (UID: \"8eae2362-d971-4645-a71e-7bf6634a536d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-l2nwx" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.838271 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ba36da60-1e6a-4010-a017-802d1cbbae71-mountpoint-dir\") pod \"csi-hostpathplugin-blwkv\" (UID: \"ba36da60-1e6a-4010-a017-802d1cbbae71\") " pod="hostpath-provisioner/csi-hostpathplugin-blwkv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.838344 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7jtz\" (UniqueName: \"kubernetes.io/projected/e17a6bea-8ec8-40db-a566-44feacd2c584-kube-api-access-g7jtz\") pod \"machine-config-controller-84d6567774-2blr9\" (UID: \"e17a6bea-8ec8-40db-a566-44feacd2c584\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2blr9" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.838380 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/25827d09-2429-43d7-95a0-82279fd15a17-cert\") pod \"ingress-canary-w6ccg\" (UID: \"25827d09-2429-43d7-95a0-82279fd15a17\") " pod="openshift-ingress-canary/ingress-canary-w6ccg" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.838417 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7fb5109d-a02b-453b-abac-9f0637d9e8fc-metrics-tls\") pod \"dns-operator-744455d44c-jzpsv\" (UID: \"7fb5109d-a02b-453b-abac-9f0637d9e8fc\") " pod="openshift-dns-operator/dns-operator-744455d44c-jzpsv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.838455 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0008dd34-6337-445e-9db2-2b662d682a68-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-c7zb5\" (UID: \"0008dd34-6337-445e-9db2-2b662d682a68\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c7zb5" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.838518 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lltk4\" (UniqueName: \"kubernetes.io/projected/8eae2362-d971-4645-a71e-7bf6634a536d-kube-api-access-lltk4\") pod \"etcd-operator-b45778765-l2nwx\" (UID: \"8eae2362-d971-4645-a71e-7bf6634a536d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-l2nwx" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.838560 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0a2a9e61-d9b7-4891-8771-76137ccd27e1-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-fvzmk\" (UID: \"0a2a9e61-d9b7-4891-8771-76137ccd27e1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fvzmk" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.838596 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/cce840a6-7988-43de-a6d7-07187abb65c3-stats-auth\") pod \"router-default-5444994796-58xmr\" (UID: \"cce840a6-7988-43de-a6d7-07187abb65c3\") " pod="openshift-ingress/router-default-5444994796-58xmr" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.838631 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pw7l\" (UniqueName: \"kubernetes.io/projected/1c2bf6c9-eba6-499e-b214-c2336431714a-kube-api-access-8pw7l\") pod \"migrator-59844c95c7-sqthq\" (UID: \"1c2bf6c9-eba6-499e-b214-c2336431714a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sqthq" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.838679 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/eceb2456-36c4-4f96-83d8-495b3eb834a8-images\") pod \"machine-config-operator-74547568cd-kq896\" (UID: \"eceb2456-36c4-4f96-83d8-495b3eb834a8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kq896" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.838716 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ba36da60-1e6a-4010-a017-802d1cbbae71-registration-dir\") pod \"csi-hostpathplugin-blwkv\" (UID: \"ba36da60-1e6a-4010-a017-802d1cbbae71\") " pod="hostpath-provisioner/csi-hostpathplugin-blwkv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.838767 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n99md\" (UniqueName: \"kubernetes.io/projected/f3c625f5-89ca-4691-ab34-3b31516aecf6-kube-api-access-n99md\") pod \"ingress-operator-5b745b69d9-5svpj\" (UID: \"f3c625f5-89ca-4691-ab34-3b31516aecf6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5svpj" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.838808 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/8eae2362-d971-4645-a71e-7bf6634a536d-etcd-service-ca\") pod \"etcd-operator-b45778765-l2nwx\" (UID: \"8eae2362-d971-4645-a71e-7bf6634a536d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-l2nwx" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.838815 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/749f12f7-ef9a-449d-a276-85ee4afe830e-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-nw2mh\" (UID: \"749f12f7-ef9a-449d-a276-85ee4afe830e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-nw2mh" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.839279 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/8eae2362-d971-4645-a71e-7bf6634a536d-etcd-ca\") pod \"etcd-operator-b45778765-l2nwx\" (UID: \"8eae2362-d971-4645-a71e-7bf6634a536d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-l2nwx" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.839319 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e17a6bea-8ec8-40db-a566-44feacd2c584-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2blr9\" (UID: \"e17a6bea-8ec8-40db-a566-44feacd2c584\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2blr9" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.839453 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ba36da60-1e6a-4010-a017-802d1cbbae71-registration-dir\") pod \"csi-hostpathplugin-blwkv\" (UID: \"ba36da60-1e6a-4010-a017-802d1cbbae71\") " pod="hostpath-provisioner/csi-hostpathplugin-blwkv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.840565 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a2a9e61-d9b7-4891-8771-76137ccd27e1-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-fvzmk\" (UID: \"0a2a9e61-d9b7-4891-8771-76137ccd27e1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fvzmk" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.840757 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ba36da60-1e6a-4010-a017-802d1cbbae71-mountpoint-dir\") pod \"csi-hostpathplugin-blwkv\" (UID: \"ba36da60-1e6a-4010-a017-802d1cbbae71\") " pod="hostpath-provisioner/csi-hostpathplugin-blwkv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.841282 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/cce840a6-7988-43de-a6d7-07187abb65c3-default-certificate\") pod \"router-default-5444994796-58xmr\" (UID: \"cce840a6-7988-43de-a6d7-07187abb65c3\") " pod="openshift-ingress/router-default-5444994796-58xmr" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.841620 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c2a528ef-fbc8-4429-a4dd-f330c594fa4a-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-9mh7s\" (UID: \"c2a528ef-fbc8-4429-a4dd-f330c594fa4a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9mh7s" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.841732 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8dc9b0d3-c9e6-4532-bbea-0a29077b7eab-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-72b2b\" (UID: \"8dc9b0d3-c9e6-4532-bbea-0a29077b7eab\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-72b2b" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.842002 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8eae2362-d971-4645-a71e-7bf6634a536d-config\") pod \"etcd-operator-b45778765-l2nwx\" (UID: \"8eae2362-d971-4645-a71e-7bf6634a536d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-l2nwx" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.842087 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pknc\" (UniqueName: \"kubernetes.io/projected/ba36da60-1e6a-4010-a017-802d1cbbae71-kube-api-access-4pknc\") pod \"csi-hostpathplugin-blwkv\" (UID: \"ba36da60-1e6a-4010-a017-802d1cbbae71\") " pod="hostpath-provisioner/csi-hostpathplugin-blwkv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.842116 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbvwg\" (UniqueName: \"kubernetes.io/projected/09f36818-91cb-4ce3-bd5f-ccaf37766551-kube-api-access-nbvwg\") pod \"dns-default-gqjqp\" (UID: \"09f36818-91cb-4ce3-bd5f-ccaf37766551\") " pod="openshift-dns/dns-default-gqjqp" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.842007 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0008dd34-6337-445e-9db2-2b662d682a68-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-c7zb5\" (UID: \"0008dd34-6337-445e-9db2-2b662d682a68\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c7zb5" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.842170 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f3c625f5-89ca-4691-ab34-3b31516aecf6-bound-sa-token\") pod \"ingress-operator-5b745b69d9-5svpj\" (UID: \"f3c625f5-89ca-4691-ab34-3b31516aecf6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5svpj" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.842233 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0cc33e9e-5984-454d-880c-5187145a2c36-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hc86c\" (UID: \"0cc33e9e-5984-454d-880c-5187145a2c36\") " pod="openshift-marketplace/marketplace-operator-79b997595-hc86c" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.842283 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9zcl\" (UniqueName: \"kubernetes.io/projected/66e8ed9b-99d5-4643-aa2a-4720ea4195b9-kube-api-access-h9zcl\") pod \"collect-profiles-29536860-dw5fg\" (UID: \"66e8ed9b-99d5-4643-aa2a-4720ea4195b9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536860-dw5fg" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.842345 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/eceb2456-36c4-4f96-83d8-495b3eb834a8-proxy-tls\") pod \"machine-config-operator-74547568cd-kq896\" (UID: \"eceb2456-36c4-4f96-83d8-495b3eb834a8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kq896" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.842399 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n962r\" (UniqueName: \"kubernetes.io/projected/3166d7d3-7842-4655-b10b-24e5731d77e0-kube-api-access-n962r\") pod \"auto-csr-approver-29536864-kmfkv\" (UID: \"3166d7d3-7842-4655-b10b-24e5731d77e0\") " pod="openshift-infra/auto-csr-approver-29536864-kmfkv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.842428 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lgjw\" (UniqueName: \"kubernetes.io/projected/eceb2456-36c4-4f96-83d8-495b3eb834a8-kube-api-access-9lgjw\") pod \"machine-config-operator-74547568cd-kq896\" (UID: \"eceb2456-36c4-4f96-83d8-495b3eb834a8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kq896" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.842484 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbdkx\" (UniqueName: \"kubernetes.io/projected/cce840a6-7988-43de-a6d7-07187abb65c3-kube-api-access-kbdkx\") pod \"router-default-5444994796-58xmr\" (UID: \"cce840a6-7988-43de-a6d7-07187abb65c3\") " pod="openshift-ingress/router-default-5444994796-58xmr" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.842514 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2a528ef-fbc8-4429-a4dd-f330c594fa4a-config\") pod \"kube-apiserver-operator-766d6c64bb-9mh7s\" (UID: \"c2a528ef-fbc8-4429-a4dd-f330c594fa4a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9mh7s" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.842573 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qphmr\" (UniqueName: \"kubernetes.io/projected/0cc33e9e-5984-454d-880c-5187145a2c36-kube-api-access-qphmr\") pod \"marketplace-operator-79b997595-hc86c\" (UID: \"0cc33e9e-5984-454d-880c-5187145a2c36\") " pod="openshift-marketplace/marketplace-operator-79b997595-hc86c" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.842610 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtzjd\" (UniqueName: \"kubernetes.io/projected/87dc0d93-e849-4b96-a1dd-1abaf03dd7cb-kube-api-access-jtzjd\") pod \"multus-admission-controller-857f4d67dd-s8rlm\" (UID: \"87dc0d93-e849-4b96-a1dd-1abaf03dd7cb\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-s8rlm" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.842672 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zh67h\" (UniqueName: \"kubernetes.io/projected/25827d09-2429-43d7-95a0-82279fd15a17-kube-api-access-zh67h\") pod \"ingress-canary-w6ccg\" (UID: \"25827d09-2429-43d7-95a0-82279fd15a17\") " pod="openshift-ingress-canary/ingress-canary-w6ccg" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.842700 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lch6n\" (UniqueName: \"kubernetes.io/projected/3fbc309b-7d6b-4598-8854-c07a038d15bd-kube-api-access-lch6n\") pod \"service-ca-operator-777779d784-cm7vp\" (UID: \"3fbc309b-7d6b-4598-8854-c07a038d15bd\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cm7vp" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.842755 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/87dc0d93-e849-4b96-a1dd-1abaf03dd7cb-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-s8rlm\" (UID: \"87dc0d93-e849-4b96-a1dd-1abaf03dd7cb\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-s8rlm" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.842780 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/66e8ed9b-99d5-4643-aa2a-4720ea4195b9-secret-volume\") pod \"collect-profiles-29536860-dw5fg\" (UID: \"66e8ed9b-99d5-4643-aa2a-4720ea4195b9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536860-dw5fg" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.842834 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f3c625f5-89ca-4691-ab34-3b31516aecf6-trusted-ca\") pod \"ingress-operator-5b745b69d9-5svpj\" (UID: \"f3c625f5-89ca-4691-ab34-3b31516aecf6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5svpj" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.842863 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dn8cv\" (UniqueName: \"kubernetes.io/projected/543c318a-e2c5-4e19-8746-333817f1e361-kube-api-access-dn8cv\") pod \"olm-operator-6b444d44fb-kcngj\" (UID: \"543c318a-e2c5-4e19-8746-333817f1e361\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcngj" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.842892 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8eae2362-d971-4645-a71e-7bf6634a536d-etcd-client\") pod \"etcd-operator-b45778765-l2nwx\" (UID: \"8eae2362-d971-4645-a71e-7bf6634a536d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-l2nwx" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.842917 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e17a6bea-8ec8-40db-a566-44feacd2c584-proxy-tls\") pod \"machine-config-controller-84d6567774-2blr9\" (UID: \"e17a6bea-8ec8-40db-a566-44feacd2c584\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2blr9" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.843032 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/749f12f7-ef9a-449d-a276-85ee4afe830e-config\") pod \"kube-controller-manager-operator-78b949d7b-nw2mh\" (UID: \"749f12f7-ef9a-449d-a276-85ee4afe830e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-nw2mh" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.843074 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/543c318a-e2c5-4e19-8746-333817f1e361-profile-collector-cert\") pod \"olm-operator-6b444d44fb-kcngj\" (UID: \"543c318a-e2c5-4e19-8746-333817f1e361\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcngj" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.843127 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cce840a6-7988-43de-a6d7-07187abb65c3-service-ca-bundle\") pod \"router-default-5444994796-58xmr\" (UID: \"cce840a6-7988-43de-a6d7-07187abb65c3\") " pod="openshift-ingress/router-default-5444994796-58xmr" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.843219 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0008dd34-6337-445e-9db2-2b662d682a68-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-c7zb5\" (UID: \"0008dd34-6337-445e-9db2-2b662d682a68\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c7zb5" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.843260 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4786h\" (UniqueName: \"kubernetes.io/projected/8dc9b0d3-c9e6-4532-bbea-0a29077b7eab-kube-api-access-4786h\") pod \"package-server-manager-789f6589d5-72b2b\" (UID: \"8dc9b0d3-c9e6-4532-bbea-0a29077b7eab\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-72b2b" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.843300 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/eceb2456-36c4-4f96-83d8-495b3eb834a8-auth-proxy-config\") pod \"machine-config-operator-74547568cd-kq896\" (UID: \"eceb2456-36c4-4f96-83d8-495b3eb834a8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kq896" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.843335 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/173cb70c-93b2-4d78-a3df-a524bf604c37-tmpfs\") pod \"packageserver-d55dfcdfc-p6lt7\" (UID: \"173cb70c-93b2-4d78-a3df-a524bf604c37\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p6lt7" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.843371 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtn8v\" (UniqueName: \"kubernetes.io/projected/c80e05b3-583f-4960-a0cc-351ff9ed3084-kube-api-access-vtn8v\") pod \"catalog-operator-68c6474976-djxgw\" (UID: \"c80e05b3-583f-4960-a0cc-351ff9ed3084\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-djxgw" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.843405 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgwl5\" (UniqueName: \"kubernetes.io/projected/483ac8b1-c17c-456d-9749-2ab085230e7a-kube-api-access-zgwl5\") pod \"service-ca-9c57cc56f-rmtjp\" (UID: \"483ac8b1-c17c-456d-9749-2ab085230e7a\") " pod="openshift-service-ca/service-ca-9c57cc56f-rmtjp" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.844169 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/f3526e22-b6e6-4a3a-95ec-67a831e5678a-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-sbqxk\" (UID: \"f3526e22-b6e6-4a3a-95ec-67a831e5678a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sbqxk" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.844842 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/749f12f7-ef9a-449d-a276-85ee4afe830e-config\") pod \"kube-controller-manager-operator-78b949d7b-nw2mh\" (UID: \"749f12f7-ef9a-449d-a276-85ee4afe830e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-nw2mh" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.845227 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2a528ef-fbc8-4429-a4dd-f330c594fa4a-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-9mh7s\" (UID: \"c2a528ef-fbc8-4429-a4dd-f330c594fa4a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9mh7s" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.843251 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8eae2362-d971-4645-a71e-7bf6634a536d-config\") pod \"etcd-operator-b45778765-l2nwx\" (UID: \"8eae2362-d971-4645-a71e-7bf6634a536d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-l2nwx" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.845994 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cce840a6-7988-43de-a6d7-07187abb65c3-metrics-certs\") pod \"router-default-5444994796-58xmr\" (UID: \"cce840a6-7988-43de-a6d7-07187abb65c3\") " pod="openshift-ingress/router-default-5444994796-58xmr" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.846044 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/749f12f7-ef9a-449d-a276-85ee4afe830e-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-nw2mh\" (UID: \"749f12f7-ef9a-449d-a276-85ee4afe830e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-nw2mh" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.846044 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2a528ef-fbc8-4429-a4dd-f330c594fa4a-config\") pod \"kube-apiserver-operator-766d6c64bb-9mh7s\" (UID: \"c2a528ef-fbc8-4429-a4dd-f330c594fa4a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9mh7s" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.846978 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8eae2362-d971-4645-a71e-7bf6634a536d-serving-cert\") pod \"etcd-operator-b45778765-l2nwx\" (UID: \"8eae2362-d971-4645-a71e-7bf6634a536d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-l2nwx" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.847077 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cce840a6-7988-43de-a6d7-07187abb65c3-service-ca-bundle\") pod \"router-default-5444994796-58xmr\" (UID: \"cce840a6-7988-43de-a6d7-07187abb65c3\") " pod="openshift-ingress/router-default-5444994796-58xmr" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.847109 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/173cb70c-93b2-4d78-a3df-a524bf604c37-tmpfs\") pod \"packageserver-d55dfcdfc-p6lt7\" (UID: \"173cb70c-93b2-4d78-a3df-a524bf604c37\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p6lt7" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.847134 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0a2a9e61-d9b7-4891-8771-76137ccd27e1-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-fvzmk\" (UID: \"0a2a9e61-d9b7-4891-8771-76137ccd27e1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fvzmk" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.847363 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/eceb2456-36c4-4f96-83d8-495b3eb834a8-auth-proxy-config\") pod \"machine-config-operator-74547568cd-kq896\" (UID: \"eceb2456-36c4-4f96-83d8-495b3eb834a8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kq896" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.847850 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/cce840a6-7988-43de-a6d7-07187abb65c3-stats-auth\") pod \"router-default-5444994796-58xmr\" (UID: \"cce840a6-7988-43de-a6d7-07187abb65c3\") " pod="openshift-ingress/router-default-5444994796-58xmr" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.849178 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e17a6bea-8ec8-40db-a566-44feacd2c584-proxy-tls\") pod \"machine-config-controller-84d6567774-2blr9\" (UID: \"e17a6bea-8ec8-40db-a566-44feacd2c584\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2blr9" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.850096 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0008dd34-6337-445e-9db2-2b662d682a68-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-c7zb5\" (UID: \"0008dd34-6337-445e-9db2-2b662d682a68\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c7zb5" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.851547 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7fb5109d-a02b-453b-abac-9f0637d9e8fc-metrics-tls\") pod \"dns-operator-744455d44c-jzpsv\" (UID: \"7fb5109d-a02b-453b-abac-9f0637d9e8fc\") " pod="openshift-dns-operator/dns-operator-744455d44c-jzpsv" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.883653 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5przj\" (UniqueName: \"kubernetes.io/projected/0adc70e1-038b-45c1-80a8-9e1f938fc161-kube-api-access-5przj\") pod \"controller-manager-879f6c89f-95wjf\" (UID: \"0adc70e1-038b-45c1-80a8-9e1f938fc161\") " pod="openshift-controller-manager/controller-manager-879f6c89f-95wjf" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.895185 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xt2c\" (UniqueName: \"kubernetes.io/projected/af75f830-5407-4eb7-ae1b-b09e6ef72737-kube-api-access-9xt2c\") pod \"machine-approver-56656f9798-94pws\" (UID: \"af75f830-5407-4eb7-ae1b-b09e6ef72737\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-94pws" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.904317 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-94pws" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.915191 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szvzs\" (UniqueName: \"kubernetes.io/projected/39f2673f-09d0-4815-babb-a8f5961eee45-kube-api-access-szvzs\") pod \"console-operator-58897d9998-vj29g\" (UID: \"39f2673f-09d0-4815-babb-a8f5961eee45\") " pod="openshift-console-operator/console-operator-58897d9998-vj29g" Feb 27 17:04:05 crc kubenswrapper[4700]: W0227 17:04:05.929257 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf75f830_5407_4eb7_ae1b_b09e6ef72737.slice/crio-dd0ddd2dc838b76481d3f8180de9bed43d08ba8f58d4c20e1d9306b88292e73d WatchSource:0}: Error finding container dd0ddd2dc838b76481d3f8180de9bed43d08ba8f58d4c20e1d9306b88292e73d: Status 404 returned error can't find the container with id dd0ddd2dc838b76481d3f8180de9bed43d08ba8f58d4c20e1d9306b88292e73d Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.944688 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:05 crc kubenswrapper[4700]: E0227 17:04:05.944942 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:06.444911886 +0000 UTC m=+206.430224653 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.945719 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:05 crc kubenswrapper[4700]: E0227 17:04:05.946377 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:06.446356764 +0000 UTC m=+206.431669551 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.948082 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppmpb\" (UniqueName: \"kubernetes.io/projected/4ea98230-93e8-4ef6-86d6-e25b8448c632-kube-api-access-ppmpb\") pod \"machine-api-operator-5694c8668f-b9772\" (UID: \"4ea98230-93e8-4ef6-86d6-e25b8448c632\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-b9772" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.963442 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5m8t\" (UniqueName: \"kubernetes.io/projected/f9d75460-0c2b-4c82-89d0-e41f8b8c275c-kube-api-access-z5m8t\") pod \"apiserver-76f77b778f-6wfhz\" (UID: \"f9d75460-0c2b-4c82-89d0-e41f8b8c275c\") " pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.987230 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rffxp\" (UniqueName: \"kubernetes.io/projected/af74d952-25fb-4e1a-ab91-9c606b4b00ab-kube-api-access-rffxp\") pod \"downloads-7954f5f757-v4nr9\" (UID: \"af74d952-25fb-4e1a-ab91-9c606b4b00ab\") " pod="openshift-console/downloads-7954f5f757-v4nr9" Feb 27 17:04:05 crc kubenswrapper[4700]: I0227 17:04:05.996929 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4n8qd\" (UniqueName: \"kubernetes.io/projected/dd3b00b4-f42e-406c-9b88-b7e2602a2493-kube-api-access-4n8qd\") pod \"route-controller-manager-6576b87f9c-frd4g\" (UID: \"dd3b00b4-f42e-406c-9b88-b7e2602a2493\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-frd4g" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.016968 4700 request.go:700] Waited for 1.007223516s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication-operator/serviceaccounts/authentication-operator/token Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.018245 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ns985\" (UniqueName: \"kubernetes.io/projected/3fdb7434-0752-4484-b253-948bbde41112-kube-api-access-ns985\") pod \"apiserver-7bbb656c7d-2jbbg\" (UID: \"3fdb7434-0752-4484-b253-948bbde41112\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2jbbg" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.018721 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-95wjf" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.027201 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.036332 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-frd4g" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.044938 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-94pws" event={"ID":"af75f830-5407-4eb7-ae1b-b09e6ef72737","Type":"ContainerStarted","Data":"dd0ddd2dc838b76481d3f8180de9bed43d08ba8f58d4c20e1d9306b88292e73d"} Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.046733 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8k6kl\" (UniqueName: \"kubernetes.io/projected/a833ce54-1145-4b0e-8136-d1747e52673b-kube-api-access-8k6kl\") pod \"authentication-operator-69f744f599-cc8ct\" (UID: \"a833ce54-1145-4b0e-8136-d1747e52673b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cc8ct" Feb 27 17:04:06 crc kubenswrapper[4700]: E0227 17:04:06.046872 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:06.546852571 +0000 UTC m=+206.532165328 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.046802 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.047675 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:06 crc kubenswrapper[4700]: E0227 17:04:06.048155 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:06.548128895 +0000 UTC m=+206.533441682 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.057713 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.060823 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/eceb2456-36c4-4f96-83d8-495b3eb834a8-images\") pod \"machine-config-operator-74547568cd-kq896\" (UID: \"eceb2456-36c4-4f96-83d8-495b3eb834a8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kq896" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.063393 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-db9mr\" (UniqueName: \"kubernetes.io/projected/649e4fdc-b0b6-4231-8038-d4fa805db298-kube-api-access-db9mr\") pod \"oauth-openshift-558db77b4-zxt8n\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.077915 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.091687 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-vj29g" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.097940 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.103719 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.109511 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/eceb2456-36c4-4f96-83d8-495b3eb834a8-proxy-tls\") pod \"machine-config-operator-74547568cd-kq896\" (UID: \"eceb2456-36c4-4f96-83d8-495b3eb834a8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kq896" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.138486 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52cl4\" (UniqueName: \"kubernetes.io/projected/91eb41f7-eb34-496d-b65d-20dfc7abf7f0-kube-api-access-52cl4\") pod \"openshift-config-operator-7777fb866f-btxgl\" (UID: \"91eb41f7-eb34-496d-b65d-20dfc7abf7f0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-btxgl" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.149318 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:06 crc kubenswrapper[4700]: E0227 17:04:06.150124 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:06.650092671 +0000 UTC m=+206.635405428 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.156168 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.180760 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-cc8ct" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.180996 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.193366 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-b9772" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.196269 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.204893 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0cc33e9e-5984-454d-880c-5187145a2c36-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hc86c\" (UID: \"0cc33e9e-5984-454d-880c-5187145a2c36\") " pod="openshift-marketplace/marketplace-operator-79b997595-hc86c" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.225542 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.254578 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.255543 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:06 crc kubenswrapper[4700]: E0227 17:04:06.256154 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:06.756137864 +0000 UTC m=+206.741450611 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.258072 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-btxgl" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.258289 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.266678 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0cc33e9e-5984-454d-880c-5187145a2c36-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hc86c\" (UID: \"0cc33e9e-5984-454d-880c-5187145a2c36\") " pod="openshift-marketplace/marketplace-operator-79b997595-hc86c" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.273138 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-v4nr9" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.276470 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.277317 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2jbbg" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.277993 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-frd4g"] Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.298651 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.316320 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f3c625f5-89ca-4691-ab34-3b31516aecf6-metrics-tls\") pod \"ingress-operator-5b745b69d9-5svpj\" (UID: \"f3c625f5-89ca-4691-ab34-3b31516aecf6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5svpj" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.324825 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.329127 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f3c625f5-89ca-4691-ab34-3b31516aecf6-trusted-ca\") pod \"ingress-operator-5b745b69d9-5svpj\" (UID: \"f3c625f5-89ca-4691-ab34-3b31516aecf6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5svpj" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.336258 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.355612 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.356188 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:06 crc kubenswrapper[4700]: E0227 17:04:06.356433 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:06.856409006 +0000 UTC m=+206.841721743 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.356781 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:06 crc kubenswrapper[4700]: E0227 17:04:06.357214 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:06.857207607 +0000 UTC m=+206.842520354 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.375414 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.396668 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.410415 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.410535 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.420218 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.424049 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-b9772"] Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.436386 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.438482 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-cc8ct"] Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.440627 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/66e8ed9b-99d5-4643-aa2a-4720ea4195b9-secret-volume\") pod \"collect-profiles-29536860-dw5fg\" (UID: \"66e8ed9b-99d5-4643-aa2a-4720ea4195b9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536860-dw5fg" Feb 27 17:04:06 crc kubenswrapper[4700]: W0227 17:04:06.441496 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4ea98230_93e8_4ef6_86d6_e25b8448c632.slice/crio-f198239009e3526c526d740f04937cb99d77e0f54addc233d1ea4f14d66b9cbd WatchSource:0}: Error finding container f198239009e3526c526d740f04937cb99d77e0f54addc233d1ea4f14d66b9cbd: Status 404 returned error can't find the container with id f198239009e3526c526d740f04937cb99d77e0f54addc233d1ea4f14d66b9cbd Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.441948 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c80e05b3-583f-4960-a0cc-351ff9ed3084-profile-collector-cert\") pod \"catalog-operator-68c6474976-djxgw\" (UID: \"c80e05b3-583f-4960-a0cc-351ff9ed3084\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-djxgw" Feb 27 17:04:06 crc kubenswrapper[4700]: W0227 17:04:06.445782 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda833ce54_1145_4b0e_8136_d1747e52673b.slice/crio-2add863bf8daa5691173c4574e126378305205ac2cf06a9c285696ddfdce60d0 WatchSource:0}: Error finding container 2add863bf8daa5691173c4574e126378305205ac2cf06a9c285696ddfdce60d0: Status 404 returned error can't find the container with id 2add863bf8daa5691173c4574e126378305205ac2cf06a9c285696ddfdce60d0 Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.445807 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.450699 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/543c318a-e2c5-4e19-8746-333817f1e361-profile-collector-cert\") pod \"olm-operator-6b444d44fb-kcngj\" (UID: \"543c318a-e2c5-4e19-8746-333817f1e361\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcngj" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.456157 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.457418 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:06 crc kubenswrapper[4700]: E0227 17:04:06.457561 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:06.957546041 +0000 UTC m=+206.942858788 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.457776 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:06 crc kubenswrapper[4700]: E0227 17:04:06.458006 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:06.957998442 +0000 UTC m=+206.943311189 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.463414 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c80e05b3-583f-4960-a0cc-351ff9ed3084-srv-cert\") pod \"catalog-operator-68c6474976-djxgw\" (UID: \"c80e05b3-583f-4960-a0cc-351ff9ed3084\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-djxgw" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.474565 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.476104 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.495906 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.517055 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.517527 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-95wjf"] Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.521070 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-v4nr9"] Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.524120 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-6wfhz"] Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.528599 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/543c318a-e2c5-4e19-8746-333817f1e361-srv-cert\") pod \"olm-operator-6b444d44fb-kcngj\" (UID: \"543c318a-e2c5-4e19-8746-333817f1e361\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcngj" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.535809 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.546692 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8dc9b0d3-c9e6-4532-bbea-0a29077b7eab-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-72b2b\" (UID: \"8dc9b0d3-c9e6-4532-bbea-0a29077b7eab\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-72b2b" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.556345 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.558629 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:06 crc kubenswrapper[4700]: E0227 17:04:06.558720 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:07.058700586 +0000 UTC m=+207.044013343 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.559017 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:06 crc kubenswrapper[4700]: E0227 17:04:06.559571 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:07.059560628 +0000 UTC m=+207.044873375 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.560745 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/173cb70c-93b2-4d78-a3df-a524bf604c37-apiservice-cert\") pod \"packageserver-d55dfcdfc-p6lt7\" (UID: \"173cb70c-93b2-4d78-a3df-a524bf604c37\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p6lt7" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.567183 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-vj29g"] Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.569377 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-zxt8n"] Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.571066 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/173cb70c-93b2-4d78-a3df-a524bf604c37-webhook-cert\") pod \"packageserver-d55dfcdfc-p6lt7\" (UID: \"173cb70c-93b2-4d78-a3df-a524bf604c37\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p6lt7" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.576546 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 27 17:04:06 crc kubenswrapper[4700]: W0227 17:04:06.582006 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod649e4fdc_b0b6_4231_8038_d4fa805db298.slice/crio-4ebea6a0755943d73b4d40b4eab2dcb763761f099ed234a8e8891c07dbeb3bd9 WatchSource:0}: Error finding container 4ebea6a0755943d73b4d40b4eab2dcb763761f099ed234a8e8891c07dbeb3bd9: Status 404 returned error can't find the container with id 4ebea6a0755943d73b4d40b4eab2dcb763761f099ed234a8e8891c07dbeb3bd9 Feb 27 17:04:06 crc kubenswrapper[4700]: W0227 17:04:06.582891 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod39f2673f_09d0_4815_babb_a8f5961eee45.slice/crio-32198e51372cc525fd2f02fe180734aea6cc8d604c4465d432652aa7080a4881 WatchSource:0}: Error finding container 32198e51372cc525fd2f02fe180734aea6cc8d604c4465d432652aa7080a4881: Status 404 returned error can't find the container with id 32198e51372cc525fd2f02fe180734aea6cc8d604c4465d432652aa7080a4881 Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.591132 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/87dc0d93-e849-4b96-a1dd-1abaf03dd7cb-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-s8rlm\" (UID: \"87dc0d93-e849-4b96-a1dd-1abaf03dd7cb\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-s8rlm" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.595788 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.618153 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.636904 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.656491 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.661004 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:06 crc kubenswrapper[4700]: E0227 17:04:06.661533 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:07.161508354 +0000 UTC m=+207.146821101 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.665178 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3fbc309b-7d6b-4598-8854-c07a038d15bd-serving-cert\") pod \"service-ca-operator-777779d784-cm7vp\" (UID: \"3fbc309b-7d6b-4598-8854-c07a038d15bd\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cm7vp" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.678015 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.678871 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3fbc309b-7d6b-4598-8854-c07a038d15bd-config\") pod \"service-ca-operator-777779d784-cm7vp\" (UID: \"3fbc309b-7d6b-4598-8854-c07a038d15bd\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cm7vp" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.680000 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-btxgl"] Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.680889 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-2jbbg"] Feb 27 17:04:06 crc kubenswrapper[4700]: W0227 17:04:06.690476 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3fdb7434_0752_4484_b253_948bbde41112.slice/crio-224c7ab65c2ed339adb24f3746e2b53c284e6328449973f9daad43fef9ee2d07 WatchSource:0}: Error finding container 224c7ab65c2ed339adb24f3746e2b53c284e6328449973f9daad43fef9ee2d07: Status 404 returned error can't find the container with id 224c7ab65c2ed339adb24f3746e2b53c284e6328449973f9daad43fef9ee2d07 Feb 27 17:04:06 crc kubenswrapper[4700]: W0227 17:04:06.692307 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91eb41f7_eb34_496d_b65d_20dfc7abf7f0.slice/crio-60f915c113f37c4d0ce0f2ce1507abbe29e6b7cd887a4e6731033016d1d5d060 WatchSource:0}: Error finding container 60f915c113f37c4d0ce0f2ce1507abbe29e6b7cd887a4e6731033016d1d5d060: Status 404 returned error can't find the container with id 60f915c113f37c4d0ce0f2ce1507abbe29e6b7cd887a4e6731033016d1d5d060 Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.695676 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.716244 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.736068 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.755823 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.763194 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:06 crc kubenswrapper[4700]: E0227 17:04:06.763532 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:07.263517581 +0000 UTC m=+207.248830328 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.775506 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.778847 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/483ac8b1-c17c-456d-9749-2ab085230e7a-signing-cabundle\") pod \"service-ca-9c57cc56f-rmtjp\" (UID: \"483ac8b1-c17c-456d-9749-2ab085230e7a\") " pod="openshift-service-ca/service-ca-9c57cc56f-rmtjp" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.795909 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.799170 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/483ac8b1-c17c-456d-9749-2ab085230e7a-signing-key\") pod \"service-ca-9c57cc56f-rmtjp\" (UID: \"483ac8b1-c17c-456d-9749-2ab085230e7a\") " pod="openshift-service-ca/service-ca-9c57cc56f-rmtjp" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.816514 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.829155 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/09f36818-91cb-4ce3-bd5f-ccaf37766551-metrics-tls\") pod \"dns-default-gqjqp\" (UID: \"09f36818-91cb-4ce3-bd5f-ccaf37766551\") " pod="openshift-dns/dns-default-gqjqp" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.836247 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 27 17:04:06 crc kubenswrapper[4700]: E0227 17:04:06.838294 4700 configmap.go:193] Couldn't get configMap openshift-dns/dns-default: failed to sync configmap cache: timed out waiting for the condition Feb 27 17:04:06 crc kubenswrapper[4700]: E0227 17:04:06.838355 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/09f36818-91cb-4ce3-bd5f-ccaf37766551-config-volume podName:09f36818-91cb-4ce3-bd5f-ccaf37766551 nodeName:}" failed. No retries permitted until 2026-02-27 17:04:07.338338065 +0000 UTC m=+207.323650822 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/09f36818-91cb-4ce3-bd5f-ccaf37766551-config-volume") pod "dns-default-gqjqp" (UID: "09f36818-91cb-4ce3-bd5f-ccaf37766551") : failed to sync configmap cache: timed out waiting for the condition Feb 27 17:04:06 crc kubenswrapper[4700]: E0227 17:04:06.838574 4700 secret.go:188] Couldn't get secret openshift-machine-config-operator/machine-config-server-tls: failed to sync secret cache: timed out waiting for the condition Feb 27 17:04:06 crc kubenswrapper[4700]: E0227 17:04:06.838615 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/39da2f34-cd0c-460e-9b62-0e6662251261-certs podName:39da2f34-cd0c-460e-9b62-0e6662251261 nodeName:}" failed. No retries permitted until 2026-02-27 17:04:07.338604642 +0000 UTC m=+207.323917399 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "certs" (UniqueName: "kubernetes.io/secret/39da2f34-cd0c-460e-9b62-0e6662251261-certs") pod "machine-config-server-n6ls6" (UID: "39da2f34-cd0c-460e-9b62-0e6662251261") : failed to sync secret cache: timed out waiting for the condition Feb 27 17:04:06 crc kubenswrapper[4700]: E0227 17:04:06.838640 4700 secret.go:188] Couldn't get secret openshift-machine-config-operator/node-bootstrapper-token: failed to sync secret cache: timed out waiting for the condition Feb 27 17:04:06 crc kubenswrapper[4700]: E0227 17:04:06.838649 4700 configmap.go:193] Couldn't get configMap openshift-operator-lifecycle-manager/collect-profiles-config: failed to sync configmap cache: timed out waiting for the condition Feb 27 17:04:06 crc kubenswrapper[4700]: E0227 17:04:06.838667 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/39da2f34-cd0c-460e-9b62-0e6662251261-node-bootstrap-token podName:39da2f34-cd0c-460e-9b62-0e6662251261 nodeName:}" failed. No retries permitted until 2026-02-27 17:04:07.338656914 +0000 UTC m=+207.323969671 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-bootstrap-token" (UniqueName: "kubernetes.io/secret/39da2f34-cd0c-460e-9b62-0e6662251261-node-bootstrap-token") pod "machine-config-server-n6ls6" (UID: "39da2f34-cd0c-460e-9b62-0e6662251261") : failed to sync secret cache: timed out waiting for the condition Feb 27 17:04:06 crc kubenswrapper[4700]: E0227 17:04:06.838743 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/66e8ed9b-99d5-4643-aa2a-4720ea4195b9-config-volume podName:66e8ed9b-99d5-4643-aa2a-4720ea4195b9 nodeName:}" failed. No retries permitted until 2026-02-27 17:04:07.338726325 +0000 UTC m=+207.324039072 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/66e8ed9b-99d5-4643-aa2a-4720ea4195b9-config-volume") pod "collect-profiles-29536860-dw5fg" (UID: "66e8ed9b-99d5-4643-aa2a-4720ea4195b9") : failed to sync configmap cache: timed out waiting for the condition Feb 27 17:04:06 crc kubenswrapper[4700]: E0227 17:04:06.839837 4700 secret.go:188] Couldn't get secret openshift-ingress-canary/canary-serving-cert: failed to sync secret cache: timed out waiting for the condition Feb 27 17:04:06 crc kubenswrapper[4700]: E0227 17:04:06.839921 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/25827d09-2429-43d7-95a0-82279fd15a17-cert podName:25827d09-2429-43d7-95a0-82279fd15a17 nodeName:}" failed. No retries permitted until 2026-02-27 17:04:07.339896556 +0000 UTC m=+207.325209403 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/25827d09-2429-43d7-95a0-82279fd15a17-cert") pod "ingress-canary-w6ccg" (UID: "25827d09-2429-43d7-95a0-82279fd15a17") : failed to sync secret cache: timed out waiting for the condition Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.856395 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.864304 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:06 crc kubenswrapper[4700]: E0227 17:04:06.864618 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:07.364593634 +0000 UTC m=+207.349906381 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.876506 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.896100 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.916133 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.937534 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.958137 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.966398 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:06 crc kubenswrapper[4700]: E0227 17:04:06.966778 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:07.466767536 +0000 UTC m=+207.452080283 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.976190 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 27 17:04:06 crc kubenswrapper[4700]: I0227 17:04:06.996276 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.016119 4700 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.035198 4700 request.go:700] Waited for 1.95185445s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/hostpath-provisioner/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.036638 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.051471 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-cc8ct" event={"ID":"a833ce54-1145-4b0e-8136-d1747e52673b","Type":"ContainerStarted","Data":"c37e656205b984fce339775ccba9e0c7d543646f3adce61aab135f3f79f51436"} Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.051511 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-cc8ct" event={"ID":"a833ce54-1145-4b0e-8136-d1747e52673b","Type":"ContainerStarted","Data":"2add863bf8daa5691173c4574e126378305205ac2cf06a9c285696ddfdce60d0"} Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.053979 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-v4nr9" event={"ID":"af74d952-25fb-4e1a-ab91-9c606b4b00ab","Type":"ContainerStarted","Data":"46c89826935f8ced3aca9ee3e1c0386986d9bc241d4a58eefeae5dad147de219"} Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.054002 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-v4nr9" event={"ID":"af74d952-25fb-4e1a-ab91-9c606b4b00ab","Type":"ContainerStarted","Data":"a69490d166e1c8837f9c9c31509a38f0a7041e85d784d8a3f5dbc4674dd8bb26"} Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.054311 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-v4nr9" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.056782 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.056817 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" event={"ID":"649e4fdc-b0b6-4231-8038-d4fa805db298","Type":"ContainerStarted","Data":"67075e64d69ab0d1b77c77d9b6745e62b45b9a116ec8ae12ea0d941fa0990695"} Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.056838 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" event={"ID":"649e4fdc-b0b6-4231-8038-d4fa805db298","Type":"ContainerStarted","Data":"4ebea6a0755943d73b4d40b4eab2dcb763761f099ed234a8e8891c07dbeb3bd9"} Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.056862 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.056986 4700 patch_prober.go:28] interesting pod/downloads-7954f5f757-v4nr9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.057009 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-v4nr9" podUID="af74d952-25fb-4e1a-ab91-9c606b4b00ab" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.058363 4700 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-zxt8n container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.7:6443/healthz\": dial tcp 10.217.0.7:6443: connect: connection refused" start-of-body= Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.058413 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" podUID="649e4fdc-b0b6-4231-8038-d4fa805db298" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.7:6443/healthz\": dial tcp 10.217.0.7:6443: connect: connection refused" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.058639 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-frd4g" event={"ID":"dd3b00b4-f42e-406c-9b88-b7e2602a2493","Type":"ContainerStarted","Data":"fafd8c78604391a3fac7b7881221641300fc14e478b5ab8ecb89e5134e14f55e"} Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.058671 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-frd4g" event={"ID":"dd3b00b4-f42e-406c-9b88-b7e2602a2493","Type":"ContainerStarted","Data":"801b7d4b3b7d83db15627b7303668b6b067dbd1f72b52266e7f2dfa9cd5a0c92"} Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.059123 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-frd4g" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.061076 4700 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-frd4g container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.061144 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-frd4g" podUID="dd3b00b4-f42e-406c-9b88-b7e2602a2493" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.063007 4700 generic.go:334] "Generic (PLEG): container finished" podID="91eb41f7-eb34-496d-b65d-20dfc7abf7f0" containerID="0608c7f9054368ed0a2fa6ef75b776740a2e12cdd084b93fa488676268023437" exitCode=0 Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.063047 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-btxgl" event={"ID":"91eb41f7-eb34-496d-b65d-20dfc7abf7f0","Type":"ContainerDied","Data":"0608c7f9054368ed0a2fa6ef75b776740a2e12cdd084b93fa488676268023437"} Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.063117 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-btxgl" event={"ID":"91eb41f7-eb34-496d-b65d-20dfc7abf7f0","Type":"ContainerStarted","Data":"60f915c113f37c4d0ce0f2ce1507abbe29e6b7cd887a4e6731033016d1d5d060"} Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.065043 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-95wjf" event={"ID":"0adc70e1-038b-45c1-80a8-9e1f938fc161","Type":"ContainerStarted","Data":"2146b15eb795e733313e22c0781d75ed1fa03e1dfb05280f0e106cacb14b9849"} Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.065065 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-95wjf" event={"ID":"0adc70e1-038b-45c1-80a8-9e1f938fc161","Type":"ContainerStarted","Data":"7a27884621b4d957d07b6c1e5a3f7d0c5ca82073175333bb74ca9e4a3965d19d"} Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.065523 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-95wjf" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.067043 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:07 crc kubenswrapper[4700]: E0227 17:04:07.067268 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:07.567237453 +0000 UTC m=+207.552550210 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.067441 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:07 crc kubenswrapper[4700]: E0227 17:04:07.067983 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:07.567971832 +0000 UTC m=+207.553284589 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.070881 4700 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-95wjf container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.070927 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-95wjf" podUID="0adc70e1-038b-45c1-80a8-9e1f938fc161" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.073132 4700 generic.go:334] "Generic (PLEG): container finished" podID="f9d75460-0c2b-4c82-89d0-e41f8b8c275c" containerID="3e20c38fa1bfbca6fb70148694f420d730abb150adbd5380970a5b5c9eb74e11" exitCode=0 Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.073188 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" event={"ID":"f9d75460-0c2b-4c82-89d0-e41f8b8c275c","Type":"ContainerDied","Data":"3e20c38fa1bfbca6fb70148694f420d730abb150adbd5380970a5b5c9eb74e11"} Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.073410 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" event={"ID":"f9d75460-0c2b-4c82-89d0-e41f8b8c275c","Type":"ContainerStarted","Data":"6df6390dc9a006bee95c967cc4fee2f31141ef1669c889d38962bd7dd3511fbb"} Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.076902 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.078585 4700 generic.go:334] "Generic (PLEG): container finished" podID="3fdb7434-0752-4484-b253-948bbde41112" containerID="fbfb17d9df7da3a910fad8d36e873f078f7d3d1eb22e93096c065a15098fdaca" exitCode=0 Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.078774 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2jbbg" event={"ID":"3fdb7434-0752-4484-b253-948bbde41112","Type":"ContainerDied","Data":"fbfb17d9df7da3a910fad8d36e873f078f7d3d1eb22e93096c065a15098fdaca"} Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.078827 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2jbbg" event={"ID":"3fdb7434-0752-4484-b253-948bbde41112","Type":"ContainerStarted","Data":"224c7ab65c2ed339adb24f3746e2b53c284e6328449973f9daad43fef9ee2d07"} Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.081020 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-vj29g" event={"ID":"39f2673f-09d0-4815-babb-a8f5961eee45","Type":"ContainerStarted","Data":"1cc5fecfbbb49e7c36e367abdecae5d7ce3843dff45eefc808650889ffcd5886"} Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.081061 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-vj29g" event={"ID":"39f2673f-09d0-4815-babb-a8f5961eee45","Type":"ContainerStarted","Data":"32198e51372cc525fd2f02fe180734aea6cc8d604c4465d432652aa7080a4881"} Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.081434 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-vj29g" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.083354 4700 patch_prober.go:28] interesting pod/console-operator-58897d9998-vj29g container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/readyz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.083441 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-vj29g" podUID="39f2673f-09d0-4815-babb-a8f5961eee45" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.8:8443/readyz\": dial tcp 10.217.0.8:8443: connect: connection refused" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.083880 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-b9772" event={"ID":"4ea98230-93e8-4ef6-86d6-e25b8448c632","Type":"ContainerStarted","Data":"bae3cba779541c4ce71736627b069da8e41dbfa1aaa5ca6796ccfe1f413380cd"} Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.083909 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-b9772" event={"ID":"4ea98230-93e8-4ef6-86d6-e25b8448c632","Type":"ContainerStarted","Data":"1537a805c3d262ceb5ec7caeedba7677df009e465ca807621fce541a58dc3f0c"} Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.083919 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-b9772" event={"ID":"4ea98230-93e8-4ef6-86d6-e25b8448c632","Type":"ContainerStarted","Data":"f198239009e3526c526d740f04937cb99d77e0f54addc233d1ea4f14d66b9cbd"} Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.090221 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-94pws" event={"ID":"af75f830-5407-4eb7-ae1b-b09e6ef72737","Type":"ContainerStarted","Data":"fa470d59b35c6deb5432b58cfdc5b34960554d40f45034bcc9086dadadad3407"} Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.090270 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-94pws" event={"ID":"af75f830-5407-4eb7-ae1b-b09e6ef72737","Type":"ContainerStarted","Data":"b2b974bf48a3cd193d114b4e38bc02ed9ea138d630f11bdfb20a5f5dc4c9ac2a"} Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.097412 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.116988 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.137640 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.168960 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:07 crc kubenswrapper[4700]: E0227 17:04:07.169178 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:07.669147918 +0000 UTC m=+207.654460705 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.171950 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:07 crc kubenswrapper[4700]: E0227 17:04:07.172642 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:07.672620389 +0000 UTC m=+207.657933146 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.191178 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zp8mz\" (UniqueName: \"kubernetes.io/projected/26f6041e-592f-48c6-93e5-35517c775186-kube-api-access-zp8mz\") pod \"cluster-samples-operator-665b6dd947-b6vld\" (UID: \"26f6041e-592f-48c6-93e5-35517c775186\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b6vld" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.199076 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lp2mq\" (UniqueName: \"kubernetes.io/projected/f2257cba-c045-4865-aecf-03642416f2cb-kube-api-access-lp2mq\") pod \"openshift-apiserver-operator-796bbdcf4f-w298r\" (UID: \"f2257cba-c045-4865-aecf-03642416f2cb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-w298r" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.215552 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6f4c\" (UniqueName: \"kubernetes.io/projected/facc8439-0b4d-4228-95b2-d01d48ee1c53-kube-api-access-d6f4c\") pod \"console-f9d7485db-kq9sb\" (UID: \"facc8439-0b4d-4228-95b2-d01d48ee1c53\") " pod="openshift-console/console-f9d7485db-kq9sb" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.230380 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swq2d\" (UniqueName: \"kubernetes.io/projected/ac062b00-1a72-44b4-90c9-de1573e3e217-kube-api-access-swq2d\") pod \"cluster-image-registry-operator-dc59b4c8b-66kw4\" (UID: \"ac062b00-1a72-44b4-90c9-de1573e3e217\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-66kw4" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.252562 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ac062b00-1a72-44b4-90c9-de1573e3e217-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-66kw4\" (UID: \"ac062b00-1a72-44b4-90c9-de1573e3e217\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-66kw4" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.275159 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ee57d6e4-3550-46d0-947e-fe6db5cf3291-bound-sa-token\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.276640 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:07 crc kubenswrapper[4700]: E0227 17:04:07.276753 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:07.776734662 +0000 UTC m=+207.762047409 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.277022 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:07 crc kubenswrapper[4700]: E0227 17:04:07.277569 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:07.777547353 +0000 UTC m=+207.762860090 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.290044 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skzrc\" (UniqueName: \"kubernetes.io/projected/f84fe6b6-37ca-406b-a8dd-d2f40638c832-kube-api-access-skzrc\") pod \"openshift-controller-manager-operator-756b6f6bc6-b5tz8\" (UID: \"f84fe6b6-37ca-406b-a8dd-d2f40638c832\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-b5tz8" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.319350 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6pzj\" (UniqueName: \"kubernetes.io/projected/ee57d6e4-3550-46d0-947e-fe6db5cf3291-kube-api-access-z6pzj\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.329537 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsszp\" (UniqueName: \"kubernetes.io/projected/7fb5109d-a02b-453b-abac-9f0637d9e8fc-kube-api-access-fsszp\") pod \"dns-operator-744455d44c-jzpsv\" (UID: \"7fb5109d-a02b-453b-abac-9f0637d9e8fc\") " pod="openshift-dns-operator/dns-operator-744455d44c-jzpsv" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.342091 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-w298r" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.354621 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98wkm\" (UniqueName: \"kubernetes.io/projected/39da2f34-cd0c-460e-9b62-0e6662251261-kube-api-access-98wkm\") pod \"machine-config-server-n6ls6\" (UID: \"39da2f34-cd0c-460e-9b62-0e6662251261\") " pod="openshift-machine-config-operator/machine-config-server-n6ls6" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.367201 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-66kw4" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.373539 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hc4kp\" (UniqueName: \"kubernetes.io/projected/0a2a9e61-d9b7-4891-8771-76137ccd27e1-kube-api-access-hc4kp\") pod \"kube-storage-version-migrator-operator-b67b599dd-fvzmk\" (UID: \"0a2a9e61-d9b7-4891-8771-76137ccd27e1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fvzmk" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.379547 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.379771 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/39da2f34-cd0c-460e-9b62-0e6662251261-certs\") pod \"machine-config-server-n6ls6\" (UID: \"39da2f34-cd0c-460e-9b62-0e6662251261\") " pod="openshift-machine-config-operator/machine-config-server-n6ls6" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.379801 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/66e8ed9b-99d5-4643-aa2a-4720ea4195b9-config-volume\") pod \"collect-profiles-29536860-dw5fg\" (UID: \"66e8ed9b-99d5-4643-aa2a-4720ea4195b9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536860-dw5fg" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.379823 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/39da2f34-cd0c-460e-9b62-0e6662251261-node-bootstrap-token\") pod \"machine-config-server-n6ls6\" (UID: \"39da2f34-cd0c-460e-9b62-0e6662251261\") " pod="openshift-machine-config-operator/machine-config-server-n6ls6" Feb 27 17:04:07 crc kubenswrapper[4700]: E0227 17:04:07.379918 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:07.879882619 +0000 UTC m=+207.865195366 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.380033 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09f36818-91cb-4ce3-bd5f-ccaf37766551-config-volume\") pod \"dns-default-gqjqp\" (UID: \"09f36818-91cb-4ce3-bd5f-ccaf37766551\") " pod="openshift-dns/dns-default-gqjqp" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.380084 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/25827d09-2429-43d7-95a0-82279fd15a17-cert\") pod \"ingress-canary-w6ccg\" (UID: \"25827d09-2429-43d7-95a0-82279fd15a17\") " pod="openshift-ingress-canary/ingress-canary-w6ccg" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.380662 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/66e8ed9b-99d5-4643-aa2a-4720ea4195b9-config-volume\") pod \"collect-profiles-29536860-dw5fg\" (UID: \"66e8ed9b-99d5-4643-aa2a-4720ea4195b9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536860-dw5fg" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.381702 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09f36818-91cb-4ce3-bd5f-ccaf37766551-config-volume\") pod \"dns-default-gqjqp\" (UID: \"09f36818-91cb-4ce3-bd5f-ccaf37766551\") " pod="openshift-dns/dns-default-gqjqp" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.384402 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/39da2f34-cd0c-460e-9b62-0e6662251261-node-bootstrap-token\") pod \"machine-config-server-n6ls6\" (UID: \"39da2f34-cd0c-460e-9b62-0e6662251261\") " pod="openshift-machine-config-operator/machine-config-server-n6ls6" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.385263 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/39da2f34-cd0c-460e-9b62-0e6662251261-certs\") pod \"machine-config-server-n6ls6\" (UID: \"39da2f34-cd0c-460e-9b62-0e6662251261\") " pod="openshift-machine-config-operator/machine-config-server-n6ls6" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.385994 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/25827d09-2429-43d7-95a0-82279fd15a17-cert\") pod \"ingress-canary-w6ccg\" (UID: \"25827d09-2429-43d7-95a0-82279fd15a17\") " pod="openshift-ingress-canary/ingress-canary-w6ccg" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.390634 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/749f12f7-ef9a-449d-a276-85ee4afe830e-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-nw2mh\" (UID: \"749f12f7-ef9a-449d-a276-85ee4afe830e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-nw2mh" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.395685 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b6vld" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.410616 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnh2k\" (UniqueName: \"kubernetes.io/projected/173cb70c-93b2-4d78-a3df-a524bf604c37-kube-api-access-hnh2k\") pod \"packageserver-d55dfcdfc-p6lt7\" (UID: \"173cb70c-93b2-4d78-a3df-a524bf604c37\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p6lt7" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.427560 4700 ???:1] "http: TLS handshake error from 192.168.126.11:58570: no serving certificate available for the kubelet" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.435903 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6skm\" (UniqueName: \"kubernetes.io/projected/f3526e22-b6e6-4a3a-95ec-67a831e5678a-kube-api-access-m6skm\") pod \"control-plane-machine-set-operator-78cbb6b69f-sbqxk\" (UID: \"f3526e22-b6e6-4a3a-95ec-67a831e5678a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sbqxk" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.440812 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-n6ls6" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.451532 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-b5tz8" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.458561 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lltk4\" (UniqueName: \"kubernetes.io/projected/8eae2362-d971-4645-a71e-7bf6634a536d-kube-api-access-lltk4\") pod \"etcd-operator-b45778765-l2nwx\" (UID: \"8eae2362-d971-4645-a71e-7bf6634a536d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-l2nwx" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.481771 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7jtz\" (UniqueName: \"kubernetes.io/projected/e17a6bea-8ec8-40db-a566-44feacd2c584-kube-api-access-g7jtz\") pod \"machine-config-controller-84d6567774-2blr9\" (UID: \"e17a6bea-8ec8-40db-a566-44feacd2c584\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2blr9" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.484973 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-kq9sb" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.486019 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:07 crc kubenswrapper[4700]: E0227 17:04:07.486433 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:07.986420945 +0000 UTC m=+207.971733692 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.496185 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n99md\" (UniqueName: \"kubernetes.io/projected/f3c625f5-89ca-4691-ab34-3b31516aecf6-kube-api-access-n99md\") pod \"ingress-operator-5b745b69d9-5svpj\" (UID: \"f3c625f5-89ca-4691-ab34-3b31516aecf6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5svpj" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.503767 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-l2nwx" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.511818 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-nw2mh" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.516847 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pw7l\" (UniqueName: \"kubernetes.io/projected/1c2bf6c9-eba6-499e-b214-c2336431714a-kube-api-access-8pw7l\") pod \"migrator-59844c95c7-sqthq\" (UID: \"1c2bf6c9-eba6-499e-b214-c2336431714a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sqthq" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.522775 4700 ???:1] "http: TLS handshake error from 192.168.126.11:58586: no serving certificate available for the kubelet" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.529020 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c2a528ef-fbc8-4429-a4dd-f330c594fa4a-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-9mh7s\" (UID: \"c2a528ef-fbc8-4429-a4dd-f330c594fa4a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9mh7s" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.536894 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-jzpsv" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.548783 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sbqxk" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.551276 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2blr9" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.553408 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pknc\" (UniqueName: \"kubernetes.io/projected/ba36da60-1e6a-4010-a017-802d1cbbae71-kube-api-access-4pknc\") pod \"csi-hostpathplugin-blwkv\" (UID: \"ba36da60-1e6a-4010-a017-802d1cbbae71\") " pod="hostpath-provisioner/csi-hostpathplugin-blwkv" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.562729 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fvzmk" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.570952 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f3c625f5-89ca-4691-ab34-3b31516aecf6-bound-sa-token\") pod \"ingress-operator-5b745b69d9-5svpj\" (UID: \"f3c625f5-89ca-4691-ab34-3b31516aecf6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5svpj" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.582789 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5svpj" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.587005 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:07 crc kubenswrapper[4700]: E0227 17:04:07.587399 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:08.087382355 +0000 UTC m=+208.072695102 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.590157 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sqthq" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.591878 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbvwg\" (UniqueName: \"kubernetes.io/projected/09f36818-91cb-4ce3-bd5f-ccaf37766551-kube-api-access-nbvwg\") pod \"dns-default-gqjqp\" (UID: \"09f36818-91cb-4ce3-bd5f-ccaf37766551\") " pod="openshift-dns/dns-default-gqjqp" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.614989 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p6lt7" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.616302 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9zcl\" (UniqueName: \"kubernetes.io/projected/66e8ed9b-99d5-4643-aa2a-4720ea4195b9-kube-api-access-h9zcl\") pod \"collect-profiles-29536860-dw5fg\" (UID: \"66e8ed9b-99d5-4643-aa2a-4720ea4195b9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536860-dw5fg" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.628281 4700 ???:1] "http: TLS handshake error from 192.168.126.11:58594: no serving certificate available for the kubelet" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.637100 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgwl5\" (UniqueName: \"kubernetes.io/projected/483ac8b1-c17c-456d-9749-2ab085230e7a-kube-api-access-zgwl5\") pod \"service-ca-9c57cc56f-rmtjp\" (UID: \"483ac8b1-c17c-456d-9749-2ab085230e7a\") " pod="openshift-service-ca/service-ca-9c57cc56f-rmtjp" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.664290 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n962r\" (UniqueName: \"kubernetes.io/projected/3166d7d3-7842-4655-b10b-24e5731d77e0-kube-api-access-n962r\") pod \"auto-csr-approver-29536864-kmfkv\" (UID: \"3166d7d3-7842-4655-b10b-24e5731d77e0\") " pod="openshift-infra/auto-csr-approver-29536864-kmfkv" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.666250 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-w298r"] Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.668663 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-66kw4"] Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.671837 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-rmtjp" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.682324 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lch6n\" (UniqueName: \"kubernetes.io/projected/3fbc309b-7d6b-4598-8854-c07a038d15bd-kube-api-access-lch6n\") pod \"service-ca-operator-777779d784-cm7vp\" (UID: \"3fbc309b-7d6b-4598-8854-c07a038d15bd\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cm7vp" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.691383 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:07 crc kubenswrapper[4700]: E0227 17:04:07.691842 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:08.191828347 +0000 UTC m=+208.177141094 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.701130 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtzjd\" (UniqueName: \"kubernetes.io/projected/87dc0d93-e849-4b96-a1dd-1abaf03dd7cb-kube-api-access-jtzjd\") pod \"multus-admission-controller-857f4d67dd-s8rlm\" (UID: \"87dc0d93-e849-4b96-a1dd-1abaf03dd7cb\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-s8rlm" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.701420 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536864-kmfkv" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.713724 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536860-dw5fg" Feb 27 17:04:07 crc kubenswrapper[4700]: W0227 17:04:07.722833 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac062b00_1a72_44b4_90c9_de1573e3e217.slice/crio-c303e822ddd3eec2f924716b68ec0e19cce6c67430bb73c74a33ac49a61167d9 WatchSource:0}: Error finding container c303e822ddd3eec2f924716b68ec0e19cce6c67430bb73c74a33ac49a61167d9: Status 404 returned error can't find the container with id c303e822ddd3eec2f924716b68ec0e19cce6c67430bb73c74a33ac49a61167d9 Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.725070 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-gqjqp" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.730812 4700 ???:1] "http: TLS handshake error from 192.168.126.11:58596: no serving certificate available for the kubelet" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.731604 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qphmr\" (UniqueName: \"kubernetes.io/projected/0cc33e9e-5984-454d-880c-5187145a2c36-kube-api-access-qphmr\") pod \"marketplace-operator-79b997595-hc86c\" (UID: \"0cc33e9e-5984-454d-880c-5187145a2c36\") " pod="openshift-marketplace/marketplace-operator-79b997595-hc86c" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.736073 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0008dd34-6337-445e-9db2-2b662d682a68-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-c7zb5\" (UID: \"0008dd34-6337-445e-9db2-2b662d682a68\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c7zb5" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.755000 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-blwkv" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.765652 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4786h\" (UniqueName: \"kubernetes.io/projected/8dc9b0d3-c9e6-4532-bbea-0a29077b7eab-kube-api-access-4786h\") pod \"package-server-manager-789f6589d5-72b2b\" (UID: \"8dc9b0d3-c9e6-4532-bbea-0a29077b7eab\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-72b2b" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.784027 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b6vld"] Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.793103 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:07 crc kubenswrapper[4700]: E0227 17:04:07.794023 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:08.293995818 +0000 UTC m=+208.279308565 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.796714 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lgjw\" (UniqueName: \"kubernetes.io/projected/eceb2456-36c4-4f96-83d8-495b3eb834a8-kube-api-access-9lgjw\") pod \"machine-config-operator-74547568cd-kq896\" (UID: \"eceb2456-36c4-4f96-83d8-495b3eb834a8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kq896" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.798306 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9mh7s" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.802239 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zh67h\" (UniqueName: \"kubernetes.io/projected/25827d09-2429-43d7-95a0-82279fd15a17-kube-api-access-zh67h\") pod \"ingress-canary-w6ccg\" (UID: \"25827d09-2429-43d7-95a0-82279fd15a17\") " pod="openshift-ingress-canary/ingress-canary-w6ccg" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.818643 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtn8v\" (UniqueName: \"kubernetes.io/projected/c80e05b3-583f-4960-a0cc-351ff9ed3084-kube-api-access-vtn8v\") pod \"catalog-operator-68c6474976-djxgw\" (UID: \"c80e05b3-583f-4960-a0cc-351ff9ed3084\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-djxgw" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.829604 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c7zb5" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.836191 4700 ???:1] "http: TLS handshake error from 192.168.126.11:58602: no serving certificate available for the kubelet" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.839407 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-b5tz8"] Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.852908 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dn8cv\" (UniqueName: \"kubernetes.io/projected/543c318a-e2c5-4e19-8746-333817f1e361-kube-api-access-dn8cv\") pod \"olm-operator-6b444d44fb-kcngj\" (UID: \"543c318a-e2c5-4e19-8746-333817f1e361\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcngj" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.870735 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kq896" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.871679 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbdkx\" (UniqueName: \"kubernetes.io/projected/cce840a6-7988-43de-a6d7-07187abb65c3-kube-api-access-kbdkx\") pod \"router-default-5444994796-58xmr\" (UID: \"cce840a6-7988-43de-a6d7-07187abb65c3\") " pod="openshift-ingress/router-default-5444994796-58xmr" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.875902 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-hc86c" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.896139 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:07 crc kubenswrapper[4700]: E0227 17:04:07.896606 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:08.396593121 +0000 UTC m=+208.381905868 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.898098 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcngj" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.905616 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-72b2b" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.930952 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-djxgw" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.933882 4700 ???:1] "http: TLS handshake error from 192.168.126.11:58616: no serving certificate available for the kubelet" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.945933 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-s8rlm" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.961049 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-cm7vp" Feb 27 17:04:07 crc kubenswrapper[4700]: I0227 17:04:07.998141 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:07 crc kubenswrapper[4700]: E0227 17:04:07.998628 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:08.498607248 +0000 UTC m=+208.483919995 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.044440 4700 ???:1] "http: TLS handshake error from 192.168.126.11:58628: no serving certificate available for the kubelet" Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.076978 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-w6ccg" Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.101525 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:08 crc kubenswrapper[4700]: E0227 17:04:08.102010 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:08.601990752 +0000 UTC m=+208.587303499 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.121696 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-btxgl" event={"ID":"91eb41f7-eb34-496d-b65d-20dfc7abf7f0","Type":"ContainerStarted","Data":"b098603a8b33cfd21624b9e449a9096954f7000132a4fa6026d0684ca90830b5"} Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.122640 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-btxgl" Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.130383 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-w298r" event={"ID":"f2257cba-c045-4865-aecf-03642416f2cb","Type":"ContainerStarted","Data":"9c7a89af27ae59fc05755efc23cc0e98cb1184f0ff827611cb2a7448605e3468"} Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.148754 4700 ???:1] "http: TLS handshake error from 192.168.126.11:58636: no serving certificate available for the kubelet" Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.150217 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" event={"ID":"f9d75460-0c2b-4c82-89d0-e41f8b8c275c","Type":"ContainerStarted","Data":"66df162497e8bf047738f71c89c2068534701f7511feaaa5909fafb17ab0d00a"} Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.150263 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" event={"ID":"f9d75460-0c2b-4c82-89d0-e41f8b8c275c","Type":"ContainerStarted","Data":"637e57f482681203e70556617e4ed36a42a88f6f66fb9ceb59ce33009c11bc6d"} Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.152549 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-n6ls6" event={"ID":"39da2f34-cd0c-460e-9b62-0e6662251261","Type":"ContainerStarted","Data":"d853abf91e1c0dbbb61dfa4129d385345a800409dd6443c96cdd3250cfc9636c"} Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.152607 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-n6ls6" event={"ID":"39da2f34-cd0c-460e-9b62-0e6662251261","Type":"ContainerStarted","Data":"a87a5053d6f33dd08846dd1e20b574370f3790f6cda9e39d530a22047adc276d"} Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.157635 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-58xmr" Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.158340 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-66kw4" event={"ID":"ac062b00-1a72-44b4-90c9-de1573e3e217","Type":"ContainerStarted","Data":"c303e822ddd3eec2f924716b68ec0e19cce6c67430bb73c74a33ac49a61167d9"} Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.166123 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2jbbg" event={"ID":"3fdb7434-0752-4484-b253-948bbde41112","Type":"ContainerStarted","Data":"580853a93755f182fdbaa53ae7f664c8d6d8e22b8c404fb410444a5df720efe3"} Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.171045 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b6vld" event={"ID":"26f6041e-592f-48c6-93e5-35517c775186","Type":"ContainerStarted","Data":"6fe9d44916e77ea4ffb79c9a56c0abffac5606e206b88b03891e61d432755fa4"} Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.175771 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-nw2mh"] Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.178224 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-b5tz8" event={"ID":"f84fe6b6-37ca-406b-a8dd-d2f40638c832","Type":"ContainerStarted","Data":"c3bd0c015103bb9bfadee1b7f4b8fd87daed306a9df7ab5dc63b45d7ead479e5"} Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.181387 4700 patch_prober.go:28] interesting pod/downloads-7954f5f757-v4nr9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.181427 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-v4nr9" podUID="af74d952-25fb-4e1a-ab91-9c606b4b00ab" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.190091 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-95wjf" Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.191374 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-frd4g" Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.210416 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:08 crc kubenswrapper[4700]: E0227 17:04:08.210932 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:08.710911911 +0000 UTC m=+208.696224658 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.210958 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-kq9sb"] Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.313836 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:08 crc kubenswrapper[4700]: E0227 17:04:08.321676 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:08.821657348 +0000 UTC m=+208.806970095 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.390065 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.415941 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:08 crc kubenswrapper[4700]: E0227 17:04:08.416393 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:08.916367743 +0000 UTC m=+208.901680490 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.463224 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-l2nwx"] Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.521935 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:08 crc kubenswrapper[4700]: E0227 17:04:08.522544 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:09.022523199 +0000 UTC m=+209.007835946 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.557507 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" podStartSLOduration=154.557489226 podStartE2EDuration="2m34.557489226s" podCreationTimestamp="2026-02-27 17:01:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:08.494125723 +0000 UTC m=+208.479438480" watchObservedRunningTime="2026-02-27 17:04:08.557489226 +0000 UTC m=+208.542801973" Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.578603 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-btxgl" podStartSLOduration=154.57858234 podStartE2EDuration="2m34.57858234s" podCreationTimestamp="2026-02-27 17:01:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:08.561977224 +0000 UTC m=+208.547289991" watchObservedRunningTime="2026-02-27 17:04:08.57858234 +0000 UTC m=+208.563895087" Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.579215 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-jzpsv"] Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.622486 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:08 crc kubenswrapper[4700]: E0227 17:04:08.623068 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:09.123054077 +0000 UTC m=+209.108366824 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:08 crc kubenswrapper[4700]: W0227 17:04:08.637783 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8eae2362_d971_4645_a71e_7bf6634a536d.slice/crio-277ddba782dff16184678a040746c0afb212bff1aaaa27786881e7f8664fe72d WatchSource:0}: Error finding container 277ddba782dff16184678a040746c0afb212bff1aaaa27786881e7f8664fe72d: Status 404 returned error can't find the container with id 277ddba782dff16184678a040746c0afb212bff1aaaa27786881e7f8664fe72d Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.727386 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:08 crc kubenswrapper[4700]: E0227 17:04:08.727808 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:09.227796946 +0000 UTC m=+209.213109693 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.736810 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" podStartSLOduration=154.736793113 podStartE2EDuration="2m34.736793113s" podCreationTimestamp="2026-02-27 17:01:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:08.735987031 +0000 UTC m=+208.721299778" watchObservedRunningTime="2026-02-27 17:04:08.736793113 +0000 UTC m=+208.722105860" Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.773770 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-vj29g" podStartSLOduration=153.773756293 podStartE2EDuration="2m33.773756293s" podCreationTimestamp="2026-02-27 17:01:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:08.77290664 +0000 UTC m=+208.758219377" watchObservedRunningTime="2026-02-27 17:04:08.773756293 +0000 UTC m=+208.759069040" Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.777747 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2blr9"] Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.828619 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:08 crc kubenswrapper[4700]: E0227 17:04:08.829434 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:09.329419194 +0000 UTC m=+209.314731941 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.833924 4700 ???:1] "http: TLS handshake error from 192.168.126.11:58652: no serving certificate available for the kubelet" Feb 27 17:04:08 crc kubenswrapper[4700]: I0227 17:04:08.945645 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:08 crc kubenswrapper[4700]: E0227 17:04:08.946174 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:09.446154668 +0000 UTC m=+209.431467415 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.051661 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:09 crc kubenswrapper[4700]: E0227 17:04:09.051997 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:09.551976405 +0000 UTC m=+209.537289152 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.053584 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:09 crc kubenswrapper[4700]: E0227 17:04:09.053869 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:09.553858885 +0000 UTC m=+209.539171622 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.070518 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-vj29g" Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.165150 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:09 crc kubenswrapper[4700]: E0227 17:04:09.165415 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:09.665379162 +0000 UTC m=+209.650691909 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.165630 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:09 crc kubenswrapper[4700]: E0227 17:04:09.166114 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:09.666106741 +0000 UTC m=+209.651419488 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.199594 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b6vld" event={"ID":"26f6041e-592f-48c6-93e5-35517c775186","Type":"ContainerStarted","Data":"025f296bb8e8315d05beacab92b3812b905de931a64251686bd7b2d3109a0491"} Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.199640 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b6vld" event={"ID":"26f6041e-592f-48c6-93e5-35517c775186","Type":"ContainerStarted","Data":"6c194c13c1c3dae027905a57bc616383db0902d623e00400f951d32010e84f08"} Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.201028 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-b5tz8" event={"ID":"f84fe6b6-37ca-406b-a8dd-d2f40638c832","Type":"ContainerStarted","Data":"80ab6dd38b79b209b31a15330c1a3eaa9a0e3f2571d94e7243da792e8a15d9e3"} Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.201145 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-95wjf" podStartSLOduration=154.201103919 podStartE2EDuration="2m34.201103919s" podCreationTimestamp="2026-02-27 17:01:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:09.162990399 +0000 UTC m=+209.148303146" watchObservedRunningTime="2026-02-27 17:04:09.201103919 +0000 UTC m=+209.186416666" Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.201539 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2jbbg" podStartSLOduration=154.201477339 podStartE2EDuration="2m34.201477339s" podCreationTimestamp="2026-02-27 17:01:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:09.186680881 +0000 UTC m=+209.171993628" watchObservedRunningTime="2026-02-27 17:04:09.201477339 +0000 UTC m=+209.186790086" Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.223611 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2blr9" event={"ID":"e17a6bea-8ec8-40db-a566-44feacd2c584","Type":"ContainerStarted","Data":"0b9064c230d5828f2986afc2de7f08fe0ad95c55206dee237957b49044cbfc7c"} Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.227730 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-kq9sb" event={"ID":"facc8439-0b4d-4228-95b2-d01d48ee1c53","Type":"ContainerStarted","Data":"40567104d50ad623de5ae2b8117b2c4f946f7c9d2cd7b29ef18d788df2360bb2"} Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.227813 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-kq9sb" event={"ID":"facc8439-0b4d-4228-95b2-d01d48ee1c53","Type":"ContainerStarted","Data":"3c147cb863230b875f155f4e04d65e0368c2333d0ea58fd2097cc0b4a8f13f7e"} Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.256092 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-nw2mh" event={"ID":"749f12f7-ef9a-449d-a276-85ee4afe830e","Type":"ContainerStarted","Data":"4d11da938b302661299928e2960e482a8a04adee116a856422cae06b7e2d00ce"} Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.256140 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-nw2mh" event={"ID":"749f12f7-ef9a-449d-a276-85ee4afe830e","Type":"ContainerStarted","Data":"7ea03a07af292c53dd6c9612a0ae4921aeed5b6212ada2e5f2f3c5222b61d468"} Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.268817 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:09 crc kubenswrapper[4700]: E0227 17:04:09.269000 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:09.76896309 +0000 UTC m=+209.754275827 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.269076 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.269350 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-58xmr" event={"ID":"cce840a6-7988-43de-a6d7-07187abb65c3","Type":"ContainerStarted","Data":"68e4d9bbcf68c32a9d4b8c26f4f21b86ab84290a1b50d05819c1af4edad4ca0a"} Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.269387 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-58xmr" event={"ID":"cce840a6-7988-43de-a6d7-07187abb65c3","Type":"ContainerStarted","Data":"2ce50a4b9f97f15d98b4d0beaf3b08991640899b433c7a1e9f23b6a004fcb80c"} Feb 27 17:04:09 crc kubenswrapper[4700]: E0227 17:04:09.269526 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:09.769510345 +0000 UTC m=+209.754823092 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.282728 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-jzpsv" event={"ID":"7fb5109d-a02b-453b-abac-9f0637d9e8fc","Type":"ContainerStarted","Data":"6f19601273d3030b3cb0c748f2d4635466c84bd672f3f04d8885420395954047"} Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.296626 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-w298r" event={"ID":"f2257cba-c045-4865-aecf-03642416f2cb","Type":"ContainerStarted","Data":"9822abed6591fd3119523fdeb0da5fb9496d77b1b7b934397d48397f6988ed1b"} Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.318312 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-66kw4" event={"ID":"ac062b00-1a72-44b4-90c9-de1573e3e217","Type":"ContainerStarted","Data":"e0632f1d4c3cbea82883f80c4ddb0631c463bd6cae414651cd205369d1f5db6b"} Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.336645 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-l2nwx" event={"ID":"8eae2362-d971-4645-a71e-7bf6634a536d","Type":"ContainerStarted","Data":"277ddba782dff16184678a040746c0afb212bff1aaaa27786881e7f8664fe72d"} Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.370537 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:09 crc kubenswrapper[4700]: E0227 17:04:09.372171 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:09.872148189 +0000 UTC m=+209.857460936 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.379310 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-frd4g" podStartSLOduration=154.379296066 podStartE2EDuration="2m34.379296066s" podCreationTimestamp="2026-02-27 17:01:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:09.377825958 +0000 UTC m=+209.363138705" watchObservedRunningTime="2026-02-27 17:04:09.379296066 +0000 UTC m=+209.364608813" Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.473812 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:09 crc kubenswrapper[4700]: E0227 17:04:09.480037 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:09.98002223 +0000 UTC m=+209.965334977 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.574771 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:09 crc kubenswrapper[4700]: E0227 17:04:09.575475 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:10.075446315 +0000 UTC m=+210.060759062 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.648060 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p6lt7"] Feb 27 17:04:09 crc kubenswrapper[4700]: W0227 17:04:09.648271 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod173cb70c_93b2_4d78_a3df_a524bf604c37.slice/crio-388544b6f8cb7bd1e949ab32ac125fa57f2b3ef6b0d149b9f618665a3be774b0 WatchSource:0}: Error finding container 388544b6f8cb7bd1e949ab32ac125fa57f2b3ef6b0d149b9f618665a3be774b0: Status 404 returned error can't find the container with id 388544b6f8cb7bd1e949ab32ac125fa57f2b3ef6b0d149b9f618665a3be774b0 Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.668580 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-cc8ct" podStartSLOduration=155.668558128 podStartE2EDuration="2m35.668558128s" podCreationTimestamp="2026-02-27 17:01:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:09.627620484 +0000 UTC m=+209.612933231" watchObservedRunningTime="2026-02-27 17:04:09.668558128 +0000 UTC m=+209.653870875" Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.683711 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:09 crc kubenswrapper[4700]: E0227 17:04:09.684104 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:10.184073366 +0000 UTC m=+210.169386113 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.734957 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-sqthq"] Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.738599 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fvzmk"] Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.738657 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rmtjp"] Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.739119 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-gqjqp"] Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.765072 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c7zb5"] Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.790654 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:09 crc kubenswrapper[4700]: E0227 17:04:09.791199 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:10.291184897 +0000 UTC m=+210.276497644 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.792731 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-5svpj"] Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.795870 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-n6ls6" podStartSLOduration=5.7958603 podStartE2EDuration="5.7958603s" podCreationTimestamp="2026-02-27 17:04:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:09.790448968 +0000 UTC m=+209.775761715" watchObservedRunningTime="2026-02-27 17:04:09.7958603 +0000 UTC m=+209.781173047" Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.823930 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sbqxk"] Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.837350 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536860-dw5fg"] Feb 27 17:04:09 crc kubenswrapper[4700]: W0227 17:04:09.838269 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0008dd34_6337_445e_9db2_2b662d682a68.slice/crio-04efe68e63f507601602abcc92e0ccda2fe92a82149d5a8fb0760fc570042b83 WatchSource:0}: Error finding container 04efe68e63f507601602abcc92e0ccda2fe92a82149d5a8fb0760fc570042b83: Status 404 returned error can't find the container with id 04efe68e63f507601602abcc92e0ccda2fe92a82149d5a8fb0760fc570042b83 Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.844837 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-94pws" podStartSLOduration=155.844823365 podStartE2EDuration="2m35.844823365s" podCreationTimestamp="2026-02-27 17:01:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:09.843970093 +0000 UTC m=+209.829282840" watchObservedRunningTime="2026-02-27 17:04:09.844823365 +0000 UTC m=+209.830136112" Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.901424 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:09 crc kubenswrapper[4700]: E0227 17:04:09.936247 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:10.436208893 +0000 UTC m=+210.421521640 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.957874 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-s8rlm"] Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.961967 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-b9772" podStartSLOduration=154.961946019 podStartE2EDuration="2m34.961946019s" podCreationTimestamp="2026-02-27 17:01:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:09.958081708 +0000 UTC m=+209.943394455" watchObservedRunningTime="2026-02-27 17:04:09.961946019 +0000 UTC m=+209.947258766" Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.969808 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-kq896"] Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.991836 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9mh7s"] Feb 27 17:04:09 crc kubenswrapper[4700]: I0227 17:04:09.998846 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-v4nr9" podStartSLOduration=154.998829557 podStartE2EDuration="2m34.998829557s" podCreationTimestamp="2026-02-27 17:01:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:09.997683187 +0000 UTC m=+209.982995934" watchObservedRunningTime="2026-02-27 17:04:09.998829557 +0000 UTC m=+209.984142294" Feb 27 17:04:10 crc kubenswrapper[4700]: W0227 17:04:10.010777 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod87dc0d93_e849_4b96_a1dd_1abaf03dd7cb.slice/crio-51899053d8adbdb61472864c38acf3e5af7c5fb5df7bcad659f009651ae2466f WatchSource:0}: Error finding container 51899053d8adbdb61472864c38acf3e5af7c5fb5df7bcad659f009651ae2466f: Status 404 returned error can't find the container with id 51899053d8adbdb61472864c38acf3e5af7c5fb5df7bcad659f009651ae2466f Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.024063 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:10 crc kubenswrapper[4700]: E0227 17:04:10.024197 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:10.524172542 +0000 UTC m=+210.509485289 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.024433 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:10 crc kubenswrapper[4700]: E0227 17:04:10.024785 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:10.524772048 +0000 UTC m=+210.510084795 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.110811 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536864-kmfkv"] Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.133844 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-blwkv"] Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.134560 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:10 crc kubenswrapper[4700]: E0227 17:04:10.134990 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:10.63497636 +0000 UTC m=+210.620289117 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.160508 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-72b2b"] Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.160686 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-58xmr" Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.160726 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcngj"] Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.167597 4700 patch_prober.go:28] interesting pod/router-default-5444994796-58xmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 17:04:10 crc kubenswrapper[4700]: [-]has-synced failed: reason withheld Feb 27 17:04:10 crc kubenswrapper[4700]: [+]process-running ok Feb 27 17:04:10 crc kubenswrapper[4700]: healthz check failed Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.167640 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-58xmr" podUID="cce840a6-7988-43de-a6d7-07187abb65c3" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.195615 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-58xmr" podStartSLOduration=155.195600952 podStartE2EDuration="2m35.195600952s" podCreationTimestamp="2026-02-27 17:01:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:10.170991176 +0000 UTC m=+210.156303923" watchObservedRunningTime="2026-02-27 17:04:10.195600952 +0000 UTC m=+210.180913699" Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.213766 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-66kw4" podStartSLOduration=155.213746278 podStartE2EDuration="2m35.213746278s" podCreationTimestamp="2026-02-27 17:01:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:10.196555367 +0000 UTC m=+210.181868114" watchObservedRunningTime="2026-02-27 17:04:10.213746278 +0000 UTC m=+210.199059025" Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.215102 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-cm7vp"] Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.240083 4700 ???:1] "http: TLS handshake error from 192.168.126.11:58660: no serving certificate available for the kubelet" Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.240699 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:10 crc kubenswrapper[4700]: E0227 17:04:10.240992 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:10.740980663 +0000 UTC m=+210.726293410 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.285743 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-w6ccg"] Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.299215 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hc86c"] Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.327664 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-b5tz8" podStartSLOduration=155.327649298 podStartE2EDuration="2m35.327649298s" podCreationTimestamp="2026-02-27 17:01:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:10.325917842 +0000 UTC m=+210.311230589" watchObservedRunningTime="2026-02-27 17:04:10.327649298 +0000 UTC m=+210.312962045" Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.342568 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:10 crc kubenswrapper[4700]: E0227 17:04:10.342858 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:10.842843906 +0000 UTC m=+210.828156653 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.364269 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-w298r" podStartSLOduration=156.364255658 podStartE2EDuration="2m36.364255658s" podCreationTimestamp="2026-02-27 17:01:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:10.362389189 +0000 UTC m=+210.347701936" watchObservedRunningTime="2026-02-27 17:04:10.364255658 +0000 UTC m=+210.349568405" Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.364894 4700 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.368760 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-djxgw"] Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.376653 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-gqjqp" event={"ID":"09f36818-91cb-4ce3-bd5f-ccaf37766551","Type":"ContainerStarted","Data":"3182358f08d0273e10ed949d1571b441f6a1095ad441ab5d274a7dc23e98d99e"} Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.415566 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b6vld" podStartSLOduration=156.415551645 podStartE2EDuration="2m36.415551645s" podCreationTimestamp="2026-02-27 17:01:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:10.414021475 +0000 UTC m=+210.399334222" watchObservedRunningTime="2026-02-27 17:04:10.415551645 +0000 UTC m=+210.400864392" Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.415599 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5svpj" event={"ID":"f3c625f5-89ca-4691-ab34-3b31516aecf6","Type":"ContainerStarted","Data":"29c3655ee6528765a641b35ef3240bfa3a8ec3b756122ebc2bb5e7d57ce0374d"} Feb 27 17:04:10 crc kubenswrapper[4700]: W0227 17:04:10.428451 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0cc33e9e_5984_454d_880c_5187145a2c36.slice/crio-743e746f4f36b8dd56462dcd3823d64d50abdf0a2e6e2c23b57905962aa3d276 WatchSource:0}: Error finding container 743e746f4f36b8dd56462dcd3823d64d50abdf0a2e6e2c23b57905962aa3d276: Status 404 returned error can't find the container with id 743e746f4f36b8dd56462dcd3823d64d50abdf0a2e6e2c23b57905962aa3d276 Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.429978 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sqthq" event={"ID":"1c2bf6c9-eba6-499e-b214-c2336431714a","Type":"ContainerStarted","Data":"db9bba57e85c41f48f9cc72661cc172bd167df4e13f35b94fc6e0992f9003310"} Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.430010 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sqthq" event={"ID":"1c2bf6c9-eba6-499e-b214-c2336431714a","Type":"ContainerStarted","Data":"0fc3de85bdac2eac3cb41b4c23d63f28a14eea1f9b48c6b9b56395ad4dc787e0"} Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.442848 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536860-dw5fg" event={"ID":"66e8ed9b-99d5-4643-aa2a-4720ea4195b9","Type":"ContainerStarted","Data":"e54141cd5fbadf9d90d3eddd6880b75ec71d3712eaee22ece681a851909fd9ca"} Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.451278 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:10 crc kubenswrapper[4700]: E0227 17:04:10.451567 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:10.95155692 +0000 UTC m=+210.936869667 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.453630 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-kq9sb" podStartSLOduration=155.453609154 podStartE2EDuration="2m35.453609154s" podCreationTimestamp="2026-02-27 17:01:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:10.443102558 +0000 UTC m=+210.428415295" watchObservedRunningTime="2026-02-27 17:04:10.453609154 +0000 UTC m=+210.438921901" Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.470149 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-nw2mh" podStartSLOduration=155.470132487 podStartE2EDuration="2m35.470132487s" podCreationTimestamp="2026-02-27 17:01:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:10.468251588 +0000 UTC m=+210.453564335" watchObservedRunningTime="2026-02-27 17:04:10.470132487 +0000 UTC m=+210.455445234" Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.547723 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-jzpsv" event={"ID":"7fb5109d-a02b-453b-abac-9f0637d9e8fc","Type":"ContainerStarted","Data":"f0cf69d08a7ee2e0c3ff8754a226a1f5f23cfc7c59013e47e9c12ea61c335520"} Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.547768 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-jzpsv" event={"ID":"7fb5109d-a02b-453b-abac-9f0637d9e8fc","Type":"ContainerStarted","Data":"d7f80375ce99aa334c8642ebb270a5a4b4410875aee8a2981dca33cb0082b852"} Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.551967 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:10 crc kubenswrapper[4700]: E0227 17:04:10.552310 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:11.052297124 +0000 UTC m=+211.037609871 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.566916 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9mh7s" event={"ID":"c2a528ef-fbc8-4429-a4dd-f330c594fa4a","Type":"ContainerStarted","Data":"4f162f637e87efc6189892a5fd985181c8f67bcaaf4d49e2ee580b7ab8849c25"} Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.578771 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-rmtjp" event={"ID":"483ac8b1-c17c-456d-9749-2ab085230e7a","Type":"ContainerStarted","Data":"23c8b5b5b8d885f9d313d4c6d80289eb04dbb04dbe9af1912ff642af762e6b7f"} Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.595560 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-jzpsv" podStartSLOduration=155.595546249 podStartE2EDuration="2m35.595546249s" podCreationTimestamp="2026-02-27 17:01:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:10.584892399 +0000 UTC m=+210.570205146" watchObservedRunningTime="2026-02-27 17:04:10.595546249 +0000 UTC m=+210.580858986" Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.601692 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sbqxk" event={"ID":"f3526e22-b6e6-4a3a-95ec-67a831e5678a","Type":"ContainerStarted","Data":"20c78e36fa1cb614d5a02b1621132adc4a848855d3bd720bf0e1d348812279a2"} Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.620691 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-rmtjp" podStartSLOduration=155.620674079 podStartE2EDuration="2m35.620674079s" podCreationTimestamp="2026-02-27 17:01:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:10.620252677 +0000 UTC m=+210.605565424" watchObservedRunningTime="2026-02-27 17:04:10.620674079 +0000 UTC m=+210.605986826" Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.641516 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536864-kmfkv" event={"ID":"3166d7d3-7842-4655-b10b-24e5731d77e0","Type":"ContainerStarted","Data":"3dede70b7b809e53b04393ff0100eaa5a0e255dd345fcbdad115b702a28680c3"} Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.655174 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:10 crc kubenswrapper[4700]: E0227 17:04:10.656879 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:11.156860348 +0000 UTC m=+211.142173085 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.677059 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c7zb5" event={"ID":"0008dd34-6337-445e-9db2-2b662d682a68","Type":"ContainerStarted","Data":"04efe68e63f507601602abcc92e0ccda2fe92a82149d5a8fb0760fc570042b83"} Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.694728 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2blr9" event={"ID":"e17a6bea-8ec8-40db-a566-44feacd2c584","Type":"ContainerStarted","Data":"910c4cc7d5d29dcc91ba3804409a7c772032c4947385ce2b7571beeb4e7be15d"} Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.694776 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2blr9" event={"ID":"e17a6bea-8ec8-40db-a566-44feacd2c584","Type":"ContainerStarted","Data":"3ea1b4aa7f0d5ca5e6c0c28b0d80cbc7d617c16a902bb63b2ba49a977ed27666"} Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.727495 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2blr9" podStartSLOduration=155.727478712 podStartE2EDuration="2m35.727478712s" podCreationTimestamp="2026-02-27 17:01:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:10.725360586 +0000 UTC m=+210.710673333" watchObservedRunningTime="2026-02-27 17:04:10.727478712 +0000 UTC m=+210.712791459" Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.735739 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kq896" event={"ID":"eceb2456-36c4-4f96-83d8-495b3eb834a8","Type":"ContainerStarted","Data":"e74c7d6ba18f5dc345e7197b3813c4f22825655a35460c7f852b15d9304a516c"} Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.749090 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fvzmk" event={"ID":"0a2a9e61-d9b7-4891-8771-76137ccd27e1","Type":"ContainerStarted","Data":"9e727a4e1f0eede772ff4b04a4191f2be73abec6e2a2f88ec845d3753f1f8f69"} Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.756196 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:10 crc kubenswrapper[4700]: E0227 17:04:10.757518 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:11.25750174 +0000 UTC m=+211.242814487 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.759120 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-s8rlm" event={"ID":"87dc0d93-e849-4b96-a1dd-1abaf03dd7cb","Type":"ContainerStarted","Data":"51899053d8adbdb61472864c38acf3e5af7c5fb5df7bcad659f009651ae2466f"} Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.780075 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p6lt7" event={"ID":"173cb70c-93b2-4d78-a3df-a524bf604c37","Type":"ContainerStarted","Data":"388544b6f8cb7bd1e949ab32ac125fa57f2b3ef6b0d149b9f618665a3be774b0"} Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.781125 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p6lt7" Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.785522 4700 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-p6lt7 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" start-of-body= Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.785552 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p6lt7" podUID="173cb70c-93b2-4d78-a3df-a524bf604c37" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.803398 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-l2nwx" event={"ID":"8eae2362-d971-4645-a71e-7bf6634a536d","Type":"ContainerStarted","Data":"d84f831d0dd0604370bbd2bbfa68b03eb9a44ccdb67ec048335f9b21deaa2f1a"} Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.821769 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p6lt7" podStartSLOduration=155.821749216 podStartE2EDuration="2m35.821749216s" podCreationTimestamp="2026-02-27 17:01:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:10.820522314 +0000 UTC m=+210.805835051" watchObservedRunningTime="2026-02-27 17:04:10.821749216 +0000 UTC m=+210.807061963" Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.853862 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-l2nwx" podStartSLOduration=155.853845189 podStartE2EDuration="2m35.853845189s" podCreationTimestamp="2026-02-27 17:01:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:10.851944539 +0000 UTC m=+210.837257286" watchObservedRunningTime="2026-02-27 17:04:10.853845189 +0000 UTC m=+210.839157936" Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.858632 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:10 crc kubenswrapper[4700]: E0227 17:04:10.867015 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:11.366999114 +0000 UTC m=+211.352311861 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:10 crc kubenswrapper[4700]: I0227 17:04:10.969033 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:10 crc kubenswrapper[4700]: E0227 17:04:10.969407 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:11.469387071 +0000 UTC m=+211.454699808 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.032193 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.032276 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.060564 4700 patch_prober.go:28] interesting pod/apiserver-76f77b778f-6wfhz container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Feb 27 17:04:11 crc kubenswrapper[4700]: [+]log ok Feb 27 17:04:11 crc kubenswrapper[4700]: [+]etcd ok Feb 27 17:04:11 crc kubenswrapper[4700]: [+]poststarthook/start-apiserver-admission-initializer ok Feb 27 17:04:11 crc kubenswrapper[4700]: [+]poststarthook/generic-apiserver-start-informers ok Feb 27 17:04:11 crc kubenswrapper[4700]: [+]poststarthook/max-in-flight-filter ok Feb 27 17:04:11 crc kubenswrapper[4700]: [+]poststarthook/storage-object-count-tracker-hook ok Feb 27 17:04:11 crc kubenswrapper[4700]: [+]poststarthook/image.openshift.io-apiserver-caches ok Feb 27 17:04:11 crc kubenswrapper[4700]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Feb 27 17:04:11 crc kubenswrapper[4700]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Feb 27 17:04:11 crc kubenswrapper[4700]: [+]poststarthook/project.openshift.io-projectcache ok Feb 27 17:04:11 crc kubenswrapper[4700]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Feb 27 17:04:11 crc kubenswrapper[4700]: [+]poststarthook/openshift.io-startinformers ok Feb 27 17:04:11 crc kubenswrapper[4700]: [+]poststarthook/openshift.io-restmapperupdater ok Feb 27 17:04:11 crc kubenswrapper[4700]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Feb 27 17:04:11 crc kubenswrapper[4700]: livez check failed Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.060626 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" podUID="f9d75460-0c2b-4c82-89d0-e41f8b8c275c" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.072845 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:11 crc kubenswrapper[4700]: E0227 17:04:11.074897 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:11.57488699 +0000 UTC m=+211.560199737 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.168700 4700 patch_prober.go:28] interesting pod/router-default-5444994796-58xmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 17:04:11 crc kubenswrapper[4700]: [-]has-synced failed: reason withheld Feb 27 17:04:11 crc kubenswrapper[4700]: [+]process-running ok Feb 27 17:04:11 crc kubenswrapper[4700]: healthz check failed Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.169070 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-58xmr" podUID="cce840a6-7988-43de-a6d7-07187abb65c3" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.173997 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:11 crc kubenswrapper[4700]: E0227 17:04:11.174411 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:11.674396722 +0000 UTC m=+211.659709469 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.275093 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:11 crc kubenswrapper[4700]: E0227 17:04:11.275755 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:11.775743762 +0000 UTC m=+211.761056509 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.277778 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2jbbg" Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.277825 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2jbbg" Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.318859 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2jbbg" Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.376847 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:11 crc kubenswrapper[4700]: E0227 17:04:11.377190 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:11.877154294 +0000 UTC m=+211.862467041 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.388839 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:11 crc kubenswrapper[4700]: E0227 17:04:11.390022 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:11.890007771 +0000 UTC m=+211.875320518 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.490202 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:11 crc kubenswrapper[4700]: E0227 17:04:11.490657 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:11.990634822 +0000 UTC m=+211.975947569 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.591237 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:11 crc kubenswrapper[4700]: E0227 17:04:11.591731 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:12.091720735 +0000 UTC m=+212.077033482 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.692931 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:11 crc kubenswrapper[4700]: E0227 17:04:11.693057 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:12.193029335 +0000 UTC m=+212.178342082 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.693286 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:11 crc kubenswrapper[4700]: E0227 17:04:11.693611 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:12.193596989 +0000 UTC m=+212.178909736 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.794797 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:11 crc kubenswrapper[4700]: E0227 17:04:11.795407 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:12.295391011 +0000 UTC m=+212.280703758 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.826827 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sbqxk" event={"ID":"f3526e22-b6e6-4a3a-95ec-67a831e5678a","Type":"ContainerStarted","Data":"6ab6450285684d7802eb39bbfdceccbd93a6761542d5ea2ce3274a26201e21c5"} Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.848517 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9mh7s" event={"ID":"c2a528ef-fbc8-4429-a4dd-f330c594fa4a","Type":"ContainerStarted","Data":"a49235983695dd19e9a4133f278e3ae220114cefc167a38a2c7bed12212d31ea"} Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.852070 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcngj" event={"ID":"543c318a-e2c5-4e19-8746-333817f1e361","Type":"ContainerStarted","Data":"b59291ffcde4f62fe470458820384da4a7db0ee80c96d810a8de1617ff5e44d6"} Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.852096 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcngj" event={"ID":"543c318a-e2c5-4e19-8746-333817f1e361","Type":"ContainerStarted","Data":"6d9fbf2fcf396782d60e335746af2756fad67281a3c79044e7565806bd2a0a86"} Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.852679 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcngj" Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.855143 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sbqxk" podStartSLOduration=156.855133129 podStartE2EDuration="2m36.855133129s" podCreationTimestamp="2026-02-27 17:01:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:11.853499536 +0000 UTC m=+211.838812283" watchObservedRunningTime="2026-02-27 17:04:11.855133129 +0000 UTC m=+211.840445866" Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.861494 4700 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-kcngj container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.861540 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcngj" podUID="543c318a-e2c5-4e19-8746-333817f1e361" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.906997 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-gqjqp" event={"ID":"09f36818-91cb-4ce3-bd5f-ccaf37766551","Type":"ContainerStarted","Data":"8e8a09bfbaa62e315c68e66ae872bb6d11e774369d7c01725828287d9f93cc99"} Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.907165 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:11 crc kubenswrapper[4700]: E0227 17:04:11.907642 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:12.407631537 +0000 UTC m=+212.392944284 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.914812 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p6lt7" event={"ID":"173cb70c-93b2-4d78-a3df-a524bf604c37","Type":"ContainerStarted","Data":"e3954ea5906cc840045f87cc18bc12b5158fe497a11b9957544003711c406715"} Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.934148 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-rmtjp" event={"ID":"483ac8b1-c17c-456d-9749-2ab085230e7a","Type":"ContainerStarted","Data":"24fd01100e4a94752abfe0612d98bc27454b6d7d5c7e3500c5f9b30603cd7026"} Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.946359 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcngj" podStartSLOduration=156.946328233 podStartE2EDuration="2m36.946328233s" podCreationTimestamp="2026-02-27 17:01:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:11.885553358 +0000 UTC m=+211.870866105" watchObservedRunningTime="2026-02-27 17:04:11.946328233 +0000 UTC m=+211.931640970" Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.950312 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-cm7vp" event={"ID":"3fbc309b-7d6b-4598-8854-c07a038d15bd","Type":"ContainerStarted","Data":"ba73d21eaefad20974d6089532fe8c9480492716f0c40d19c0335412b5fbda1d"} Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.950392 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-cm7vp" event={"ID":"3fbc309b-7d6b-4598-8854-c07a038d15bd","Type":"ContainerStarted","Data":"f55eb1377e8f2d4ddb8f9a8d0e885a2c7d9027a688ab71dbe9285618ac6f6a98"} Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.964439 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-blwkv" event={"ID":"ba36da60-1e6a-4010-a017-802d1cbbae71","Type":"ContainerStarted","Data":"f1e3e00b31be24d401fe3cf4baa6fdf255017d4c59fbb85e7dba236e5d58b5a7"} Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.976645 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9mh7s" podStartSLOduration=156.976624648 podStartE2EDuration="2m36.976624648s" podCreationTimestamp="2026-02-27 17:01:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:11.947192095 +0000 UTC m=+211.932504832" watchObservedRunningTime="2026-02-27 17:04:11.976624648 +0000 UTC m=+211.961937395" Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.982846 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kq896" event={"ID":"eceb2456-36c4-4f96-83d8-495b3eb834a8","Type":"ContainerStarted","Data":"28d6b924456056d107bbda3c4f1d97bbfd685a7ec0570a54f24d026bcbed7407"} Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.982893 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kq896" event={"ID":"eceb2456-36c4-4f96-83d8-495b3eb834a8","Type":"ContainerStarted","Data":"f3702c67f1cc5e9e6a6d72d92dc99f2b7aef5cfa4cd99a6133c2f6a329790ba3"} Feb 27 17:04:11 crc kubenswrapper[4700]: I0227 17:04:11.999382 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-cm7vp" podStartSLOduration=156.999362035 podStartE2EDuration="2m36.999362035s" podCreationTimestamp="2026-02-27 17:01:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:11.977129411 +0000 UTC m=+211.962442158" watchObservedRunningTime="2026-02-27 17:04:11.999362035 +0000 UTC m=+211.984674792" Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.001287 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-95wjf"] Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.009125 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:12 crc kubenswrapper[4700]: E0227 17:04:12.010537 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:12.510517278 +0000 UTC m=+212.495830015 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.015628 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kq896" podStartSLOduration=157.015604331 podStartE2EDuration="2m37.015604331s" podCreationTimestamp="2026-02-27 17:01:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:12.009174212 +0000 UTC m=+211.994486959" watchObservedRunningTime="2026-02-27 17:04:12.015604331 +0000 UTC m=+212.000917068" Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.030868 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-s8rlm" event={"ID":"87dc0d93-e849-4b96-a1dd-1abaf03dd7cb","Type":"ContainerStarted","Data":"1b72713b41db790875b5576c4d9d0efc69730d05cf8259c68eac350475e73687"} Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.043694 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-frd4g"] Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.078304 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sqthq" event={"ID":"1c2bf6c9-eba6-499e-b214-c2336431714a","Type":"ContainerStarted","Data":"6c569f19c51b79bb687c8478c5aa1821ca04088990226e76ce8ff139597d2f7d"} Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.104741 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c7zb5" event={"ID":"0008dd34-6337-445e-9db2-2b662d682a68","Type":"ContainerStarted","Data":"b40db64b0f0dccae922941ad347f4aee649d20e7368af115df5b421510907c01"} Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.114369 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:12 crc kubenswrapper[4700]: E0227 17:04:12.114829 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:12.614812454 +0000 UTC m=+212.600125201 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.114829 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hc86c" event={"ID":"0cc33e9e-5984-454d-880c-5187145a2c36","Type":"ContainerStarted","Data":"d4afb43ef0068fad5a80a64fec6abe581f5e3d7507a435c30c279abc4eaff145"} Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.114883 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hc86c" event={"ID":"0cc33e9e-5984-454d-880c-5187145a2c36","Type":"ContainerStarted","Data":"743e746f4f36b8dd56462dcd3823d64d50abdf0a2e6e2c23b57905962aa3d276"} Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.115266 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-hc86c" Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.121954 4700 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-hc86c container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.122298 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-hc86c" podUID="0cc33e9e-5984-454d-880c-5187145a2c36" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.136377 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sqthq" podStartSLOduration=157.136357699 podStartE2EDuration="2m37.136357699s" podCreationTimestamp="2026-02-27 17:01:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:12.105950581 +0000 UTC m=+212.091263348" watchObservedRunningTime="2026-02-27 17:04:12.136357699 +0000 UTC m=+212.121670446" Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.138184 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5svpj" event={"ID":"f3c625f5-89ca-4691-ab34-3b31516aecf6","Type":"ContainerStarted","Data":"c790cdf00d99827419eea4dccccd3252f90e3313dad132ea17613303c7bf078a"} Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.138243 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5svpj" event={"ID":"f3c625f5-89ca-4691-ab34-3b31516aecf6","Type":"ContainerStarted","Data":"1c7e0e5eb0a0779e2aeb51f7bf66357790a9cdd3cdf717a009039043cd625341"} Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.179399 4700 patch_prober.go:28] interesting pod/router-default-5444994796-58xmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 17:04:12 crc kubenswrapper[4700]: [-]has-synced failed: reason withheld Feb 27 17:04:12 crc kubenswrapper[4700]: [+]process-running ok Feb 27 17:04:12 crc kubenswrapper[4700]: healthz check failed Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.179482 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-58xmr" podUID="cce840a6-7988-43de-a6d7-07187abb65c3" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.190786 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-hc86c" podStartSLOduration=157.190768168 podStartE2EDuration="2m37.190768168s" podCreationTimestamp="2026-02-27 17:01:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:12.188444927 +0000 UTC m=+212.173757674" watchObservedRunningTime="2026-02-27 17:04:12.190768168 +0000 UTC m=+212.176080915" Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.197424 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c7zb5" podStartSLOduration=157.197394551 podStartE2EDuration="2m37.197394551s" podCreationTimestamp="2026-02-27 17:01:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:12.135772754 +0000 UTC m=+212.121085501" watchObservedRunningTime="2026-02-27 17:04:12.197394551 +0000 UTC m=+212.182707288" Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.197677 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-djxgw" event={"ID":"c80e05b3-583f-4960-a0cc-351ff9ed3084","Type":"ContainerStarted","Data":"911f5ca4aa7259bb2fca9e7ad37838300f2627ef1463f39031241f911ceca50a"} Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.197706 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-djxgw" event={"ID":"c80e05b3-583f-4960-a0cc-351ff9ed3084","Type":"ContainerStarted","Data":"2fd629d2c40d8bc468448251b716a0142a744ed7b541994c2528e1c16b733bd4"} Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.197980 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-djxgw" Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.210137 4700 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-djxgw container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.211198 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-djxgw" podUID="c80e05b3-583f-4960-a0cc-351ff9ed3084" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.218247 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:12 crc kubenswrapper[4700]: E0227 17:04:12.223417 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:12.723148247 +0000 UTC m=+212.708460994 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.235428 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5svpj" podStartSLOduration=157.235408229 podStartE2EDuration="2m37.235408229s" podCreationTimestamp="2026-02-27 17:01:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:12.23315069 +0000 UTC m=+212.218463447" watchObservedRunningTime="2026-02-27 17:04:12.235408229 +0000 UTC m=+212.220720976" Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.244766 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-72b2b" event={"ID":"8dc9b0d3-c9e6-4532-bbea-0a29077b7eab","Type":"ContainerStarted","Data":"bc26515123c2f038f58e8d48122127078eadd01ad588f91491ba4d1f8a1e9a85"} Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.244826 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-72b2b" event={"ID":"8dc9b0d3-c9e6-4532-bbea-0a29077b7eab","Type":"ContainerStarted","Data":"2b0c3d6264791085301722573c2ad812b3aad823357d6d6a223f840d99336940"} Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.244837 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-72b2b" event={"ID":"8dc9b0d3-c9e6-4532-bbea-0a29077b7eab","Type":"ContainerStarted","Data":"3f45d00b1462c00c301b1eeb1199ded12a6100ebe823deace5186493ae7718e9"} Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.245188 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-72b2b" Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.278233 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fvzmk" event={"ID":"0a2a9e61-d9b7-4891-8771-76137ccd27e1","Type":"ContainerStarted","Data":"9ad6b1d15e53f5c4ccd10cedbf08412e34895b2fee0d82e32179920faa8eae11"} Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.296252 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-btxgl" Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.300815 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-w6ccg" event={"ID":"25827d09-2429-43d7-95a0-82279fd15a17","Type":"ContainerStarted","Data":"e09b8fc7b03aeaf023c18150593db0ca575deac0f47d03b0d137b1ee685752f1"} Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.300856 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-w6ccg" event={"ID":"25827d09-2429-43d7-95a0-82279fd15a17","Type":"ContainerStarted","Data":"b0e8832f93d5cf9b2beb9a709543a8bce4b1ff8a37af7b6bec0241da62e0827a"} Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.306710 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-djxgw" podStartSLOduration=157.30667766 podStartE2EDuration="2m37.30667766s" podCreationTimestamp="2026-02-27 17:01:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:12.276019185 +0000 UTC m=+212.261331952" watchObservedRunningTime="2026-02-27 17:04:12.30667766 +0000 UTC m=+212.291990407" Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.309122 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-72b2b" podStartSLOduration=157.309107354 podStartE2EDuration="2m37.309107354s" podCreationTimestamp="2026-02-27 17:01:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:12.293236937 +0000 UTC m=+212.278549694" watchObservedRunningTime="2026-02-27 17:04:12.309107354 +0000 UTC m=+212.294420101" Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.317684 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536860-dw5fg" event={"ID":"66e8ed9b-99d5-4643-aa2a-4720ea4195b9","Type":"ContainerStarted","Data":"9cb12c74550ad13f426db5679571715af91b50eaa633f56ce6a1c7a8c66ae2d3"} Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.318880 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-frd4g" podUID="dd3b00b4-f42e-406c-9b88-b7e2602a2493" containerName="route-controller-manager" containerID="cri-o://fafd8c78604391a3fac7b7881221641300fc14e478b5ab8ecb89e5134e14f55e" gracePeriod=30 Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.323516 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:12 crc kubenswrapper[4700]: E0227 17:04:12.325835 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:12.825817902 +0000 UTC m=+212.811130649 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.326779 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-95wjf" podUID="0adc70e1-038b-45c1-80a8-9e1f938fc161" containerName="controller-manager" containerID="cri-o://2146b15eb795e733313e22c0781d75ed1fa03e1dfb05280f0e106cacb14b9849" gracePeriod=30 Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.334652 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2jbbg" Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.337520 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fvzmk" podStartSLOduration=157.337492819 podStartE2EDuration="2m37.337492819s" podCreationTimestamp="2026-02-27 17:01:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:12.312983365 +0000 UTC m=+212.298296102" watchObservedRunningTime="2026-02-27 17:04:12.337492819 +0000 UTC m=+212.322805566" Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.380690 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-w6ccg" podStartSLOduration=8.380669022 podStartE2EDuration="8.380669022s" podCreationTimestamp="2026-02-27 17:04:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:12.378845984 +0000 UTC m=+212.364158731" watchObservedRunningTime="2026-02-27 17:04:12.380669022 +0000 UTC m=+212.365981759" Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.426373 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:12 crc kubenswrapper[4700]: E0227 17:04:12.427142 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:12.92709706 +0000 UTC m=+212.912409807 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.459272 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29536860-dw5fg" podStartSLOduration=158.459250304 podStartE2EDuration="2m38.459250304s" podCreationTimestamp="2026-02-27 17:01:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:12.457813437 +0000 UTC m=+212.443126184" watchObservedRunningTime="2026-02-27 17:04:12.459250304 +0000 UTC m=+212.444563041" Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.534164 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:12 crc kubenswrapper[4700]: E0227 17:04:12.534509 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:13.034496099 +0000 UTC m=+213.019808836 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.635777 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:12 crc kubenswrapper[4700]: E0227 17:04:12.635963 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:13.135936072 +0000 UTC m=+213.121248819 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.636063 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:12 crc kubenswrapper[4700]: E0227 17:04:12.636433 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:13.136414804 +0000 UTC m=+213.121727551 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.739992 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:12 crc kubenswrapper[4700]: E0227 17:04:12.740289 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:13.24026691 +0000 UTC m=+213.225579657 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.740653 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:12 crc kubenswrapper[4700]: E0227 17:04:12.740942 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:13.240930698 +0000 UTC m=+213.226243445 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.841737 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:12 crc kubenswrapper[4700]: E0227 17:04:12.842153 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:13.342138164 +0000 UTC m=+213.327450911 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.878157 4700 ???:1] "http: TLS handshake error from 192.168.126.11:58664: no serving certificate available for the kubelet" Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.885851 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-frd4g" Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.910500 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-95wjf" Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.917574 4700 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-p6lt7 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.917627 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p6lt7" podUID="173cb70c-93b2-4d78-a3df-a524bf604c37" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.943215 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4n8qd\" (UniqueName: \"kubernetes.io/projected/dd3b00b4-f42e-406c-9b88-b7e2602a2493-kube-api-access-4n8qd\") pod \"dd3b00b4-f42e-406c-9b88-b7e2602a2493\" (UID: \"dd3b00b4-f42e-406c-9b88-b7e2602a2493\") " Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.943270 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd3b00b4-f42e-406c-9b88-b7e2602a2493-serving-cert\") pod \"dd3b00b4-f42e-406c-9b88-b7e2602a2493\" (UID: \"dd3b00b4-f42e-406c-9b88-b7e2602a2493\") " Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.943489 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dd3b00b4-f42e-406c-9b88-b7e2602a2493-client-ca\") pod \"dd3b00b4-f42e-406c-9b88-b7e2602a2493\" (UID: \"dd3b00b4-f42e-406c-9b88-b7e2602a2493\") " Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.943509 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd3b00b4-f42e-406c-9b88-b7e2602a2493-config\") pod \"dd3b00b4-f42e-406c-9b88-b7e2602a2493\" (UID: \"dd3b00b4-f42e-406c-9b88-b7e2602a2493\") " Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.943740 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:12 crc kubenswrapper[4700]: E0227 17:04:12.944151 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:13.444138541 +0000 UTC m=+213.429451278 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.944408 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd3b00b4-f42e-406c-9b88-b7e2602a2493-client-ca" (OuterVolumeSpecName: "client-ca") pod "dd3b00b4-f42e-406c-9b88-b7e2602a2493" (UID: "dd3b00b4-f42e-406c-9b88-b7e2602a2493"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.944758 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd3b00b4-f42e-406c-9b88-b7e2602a2493-config" (OuterVolumeSpecName: "config") pod "dd3b00b4-f42e-406c-9b88-b7e2602a2493" (UID: "dd3b00b4-f42e-406c-9b88-b7e2602a2493"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.954688 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd3b00b4-f42e-406c-9b88-b7e2602a2493-kube-api-access-4n8qd" (OuterVolumeSpecName: "kube-api-access-4n8qd") pod "dd3b00b4-f42e-406c-9b88-b7e2602a2493" (UID: "dd3b00b4-f42e-406c-9b88-b7e2602a2493"). InnerVolumeSpecName "kube-api-access-4n8qd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:04:12 crc kubenswrapper[4700]: I0227 17:04:12.961772 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd3b00b4-f42e-406c-9b88-b7e2602a2493-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "dd3b00b4-f42e-406c-9b88-b7e2602a2493" (UID: "dd3b00b4-f42e-406c-9b88-b7e2602a2493"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.044984 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0adc70e1-038b-45c1-80a8-9e1f938fc161-config\") pod \"0adc70e1-038b-45c1-80a8-9e1f938fc161\" (UID: \"0adc70e1-038b-45c1-80a8-9e1f938fc161\") " Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.045099 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0adc70e1-038b-45c1-80a8-9e1f938fc161-proxy-ca-bundles\") pod \"0adc70e1-038b-45c1-80a8-9e1f938fc161\" (UID: \"0adc70e1-038b-45c1-80a8-9e1f938fc161\") " Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.045142 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0adc70e1-038b-45c1-80a8-9e1f938fc161-serving-cert\") pod \"0adc70e1-038b-45c1-80a8-9e1f938fc161\" (UID: \"0adc70e1-038b-45c1-80a8-9e1f938fc161\") " Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.045245 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.045308 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5przj\" (UniqueName: \"kubernetes.io/projected/0adc70e1-038b-45c1-80a8-9e1f938fc161-kube-api-access-5przj\") pod \"0adc70e1-038b-45c1-80a8-9e1f938fc161\" (UID: \"0adc70e1-038b-45c1-80a8-9e1f938fc161\") " Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.045336 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0adc70e1-038b-45c1-80a8-9e1f938fc161-client-ca\") pod \"0adc70e1-038b-45c1-80a8-9e1f938fc161\" (UID: \"0adc70e1-038b-45c1-80a8-9e1f938fc161\") " Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.045596 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4n8qd\" (UniqueName: \"kubernetes.io/projected/dd3b00b4-f42e-406c-9b88-b7e2602a2493-kube-api-access-4n8qd\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.045607 4700 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd3b00b4-f42e-406c-9b88-b7e2602a2493-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.045616 4700 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dd3b00b4-f42e-406c-9b88-b7e2602a2493-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.045624 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd3b00b4-f42e-406c-9b88-b7e2602a2493-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.046359 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0adc70e1-038b-45c1-80a8-9e1f938fc161-client-ca" (OuterVolumeSpecName: "client-ca") pod "0adc70e1-038b-45c1-80a8-9e1f938fc161" (UID: "0adc70e1-038b-45c1-80a8-9e1f938fc161"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:04:13 crc kubenswrapper[4700]: E0227 17:04:13.046426 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:13.546412505 +0000 UTC m=+213.531725252 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.046560 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0adc70e1-038b-45c1-80a8-9e1f938fc161-config" (OuterVolumeSpecName: "config") pod "0adc70e1-038b-45c1-80a8-9e1f938fc161" (UID: "0adc70e1-038b-45c1-80a8-9e1f938fc161"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.046995 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0adc70e1-038b-45c1-80a8-9e1f938fc161-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "0adc70e1-038b-45c1-80a8-9e1f938fc161" (UID: "0adc70e1-038b-45c1-80a8-9e1f938fc161"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.058091 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0adc70e1-038b-45c1-80a8-9e1f938fc161-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0adc70e1-038b-45c1-80a8-9e1f938fc161" (UID: "0adc70e1-038b-45c1-80a8-9e1f938fc161"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.061026 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0adc70e1-038b-45c1-80a8-9e1f938fc161-kube-api-access-5przj" (OuterVolumeSpecName: "kube-api-access-5przj") pod "0adc70e1-038b-45c1-80a8-9e1f938fc161" (UID: "0adc70e1-038b-45c1-80a8-9e1f938fc161"). InnerVolumeSpecName "kube-api-access-5przj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.147143 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.147502 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5przj\" (UniqueName: \"kubernetes.io/projected/0adc70e1-038b-45c1-80a8-9e1f938fc161-kube-api-access-5przj\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.147515 4700 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0adc70e1-038b-45c1-80a8-9e1f938fc161-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.147525 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0adc70e1-038b-45c1-80a8-9e1f938fc161-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.147534 4700 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0adc70e1-038b-45c1-80a8-9e1f938fc161-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.147541 4700 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0adc70e1-038b-45c1-80a8-9e1f938fc161-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:13 crc kubenswrapper[4700]: E0227 17:04:13.147781 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:13.647769226 +0000 UTC m=+213.633081973 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.170044 4700 patch_prober.go:28] interesting pod/router-default-5444994796-58xmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 17:04:13 crc kubenswrapper[4700]: [-]has-synced failed: reason withheld Feb 27 17:04:13 crc kubenswrapper[4700]: [+]process-running ok Feb 27 17:04:13 crc kubenswrapper[4700]: healthz check failed Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.170109 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-58xmr" podUID="cce840a6-7988-43de-a6d7-07187abb65c3" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.248265 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:13 crc kubenswrapper[4700]: E0227 17:04:13.248481 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:13.748438868 +0000 UTC m=+213.733751615 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.248590 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:13 crc kubenswrapper[4700]: E0227 17:04:13.248889 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:13.74887672 +0000 UTC m=+213.734189467 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.328917 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-s8rlm" event={"ID":"87dc0d93-e849-4b96-a1dd-1abaf03dd7cb","Type":"ContainerStarted","Data":"4f72a97a3f21b383a5de9b8e56b9de7d1e0703fc9d404071a567aae04043dd99"} Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.336367 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-gqjqp" event={"ID":"09f36818-91cb-4ce3-bd5f-ccaf37766551","Type":"ContainerStarted","Data":"9a637b6e1e44e51defb18555d2e22d05d12bcd40d1b928480439a65454e7ed7f"} Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.336519 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-gqjqp" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.337700 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-blwkv" event={"ID":"ba36da60-1e6a-4010-a017-802d1cbbae71","Type":"ContainerStarted","Data":"593d61b4b3b1f121801752f4e1e9bbe4d9f8a86375e84639442cac06b43dbaaf"} Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.344132 4700 generic.go:334] "Generic (PLEG): container finished" podID="0adc70e1-038b-45c1-80a8-9e1f938fc161" containerID="2146b15eb795e733313e22c0781d75ed1fa03e1dfb05280f0e106cacb14b9849" exitCode=0 Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.344195 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-95wjf" event={"ID":"0adc70e1-038b-45c1-80a8-9e1f938fc161","Type":"ContainerDied","Data":"2146b15eb795e733313e22c0781d75ed1fa03e1dfb05280f0e106cacb14b9849"} Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.344217 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-95wjf" event={"ID":"0adc70e1-038b-45c1-80a8-9e1f938fc161","Type":"ContainerDied","Data":"7a27884621b4d957d07b6c1e5a3f7d0c5ca82073175333bb74ca9e4a3965d19d"} Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.344233 4700 scope.go:117] "RemoveContainer" containerID="2146b15eb795e733313e22c0781d75ed1fa03e1dfb05280f0e106cacb14b9849" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.344332 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-95wjf" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.349771 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:13 crc kubenswrapper[4700]: E0227 17:04:13.349942 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:13.849913542 +0000 UTC m=+213.835226289 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.350217 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:13 crc kubenswrapper[4700]: E0227 17:04:13.350549 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:13.850533938 +0000 UTC m=+213.835846735 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.350908 4700 generic.go:334] "Generic (PLEG): container finished" podID="dd3b00b4-f42e-406c-9b88-b7e2602a2493" containerID="fafd8c78604391a3fac7b7881221641300fc14e478b5ab8ecb89e5134e14f55e" exitCode=0 Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.351371 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-frd4g" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.351758 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-frd4g" event={"ID":"dd3b00b4-f42e-406c-9b88-b7e2602a2493","Type":"ContainerDied","Data":"fafd8c78604391a3fac7b7881221641300fc14e478b5ab8ecb89e5134e14f55e"} Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.351778 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-frd4g" event={"ID":"dd3b00b4-f42e-406c-9b88-b7e2602a2493","Type":"ContainerDied","Data":"801b7d4b3b7d83db15627b7303668b6b067dbd1f72b52266e7f2dfa9cd5a0c92"} Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.356944 4700 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-hc86c container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.357001 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-hc86c" podUID="0cc33e9e-5984-454d-880c-5187145a2c36" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.359861 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-s8rlm" podStartSLOduration=158.359833732 podStartE2EDuration="2m38.359833732s" podCreationTimestamp="2026-02-27 17:01:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:13.357194213 +0000 UTC m=+213.342506960" watchObservedRunningTime="2026-02-27 17:04:13.359833732 +0000 UTC m=+213.345146479" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.365377 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-djxgw" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.370753 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p6lt7" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.385390 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcngj" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.385407 4700 scope.go:117] "RemoveContainer" containerID="2146b15eb795e733313e22c0781d75ed1fa03e1dfb05280f0e106cacb14b9849" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.397502 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-frd4g"] Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.400315 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-frd4g"] Feb 27 17:04:13 crc kubenswrapper[4700]: E0227 17:04:13.400856 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2146b15eb795e733313e22c0781d75ed1fa03e1dfb05280f0e106cacb14b9849\": container with ID starting with 2146b15eb795e733313e22c0781d75ed1fa03e1dfb05280f0e106cacb14b9849 not found: ID does not exist" containerID="2146b15eb795e733313e22c0781d75ed1fa03e1dfb05280f0e106cacb14b9849" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.400912 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2146b15eb795e733313e22c0781d75ed1fa03e1dfb05280f0e106cacb14b9849"} err="failed to get container status \"2146b15eb795e733313e22c0781d75ed1fa03e1dfb05280f0e106cacb14b9849\": rpc error: code = NotFound desc = could not find container \"2146b15eb795e733313e22c0781d75ed1fa03e1dfb05280f0e106cacb14b9849\": container with ID starting with 2146b15eb795e733313e22c0781d75ed1fa03e1dfb05280f0e106cacb14b9849 not found: ID does not exist" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.400952 4700 scope.go:117] "RemoveContainer" containerID="fafd8c78604391a3fac7b7881221641300fc14e478b5ab8ecb89e5134e14f55e" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.427382 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-gqjqp" podStartSLOduration=9.427364344 podStartE2EDuration="9.427364344s" podCreationTimestamp="2026-02-27 17:04:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:13.425888866 +0000 UTC m=+213.411201613" watchObservedRunningTime="2026-02-27 17:04:13.427364344 +0000 UTC m=+213.412677091" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.444709 4700 scope.go:117] "RemoveContainer" containerID="fafd8c78604391a3fac7b7881221641300fc14e478b5ab8ecb89e5134e14f55e" Feb 27 17:04:13 crc kubenswrapper[4700]: E0227 17:04:13.445266 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fafd8c78604391a3fac7b7881221641300fc14e478b5ab8ecb89e5134e14f55e\": container with ID starting with fafd8c78604391a3fac7b7881221641300fc14e478b5ab8ecb89e5134e14f55e not found: ID does not exist" containerID="fafd8c78604391a3fac7b7881221641300fc14e478b5ab8ecb89e5134e14f55e" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.445297 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fafd8c78604391a3fac7b7881221641300fc14e478b5ab8ecb89e5134e14f55e"} err="failed to get container status \"fafd8c78604391a3fac7b7881221641300fc14e478b5ab8ecb89e5134e14f55e\": rpc error: code = NotFound desc = could not find container \"fafd8c78604391a3fac7b7881221641300fc14e478b5ab8ecb89e5134e14f55e\": container with ID starting with fafd8c78604391a3fac7b7881221641300fc14e478b5ab8ecb89e5134e14f55e not found: ID does not exist" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.452363 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:13 crc kubenswrapper[4700]: E0227 17:04:13.454049 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:13.954019934 +0000 UTC m=+213.939332681 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.541022 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-95wjf"] Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.541077 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-95wjf"] Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.557243 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:13 crc kubenswrapper[4700]: E0227 17:04:13.557580 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:14.057567862 +0000 UTC m=+214.042880609 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.658747 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:13 crc kubenswrapper[4700]: E0227 17:04:13.658958 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:14.158930782 +0000 UTC m=+214.144243529 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.659113 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:13 crc kubenswrapper[4700]: E0227 17:04:13.659490 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:14.159483137 +0000 UTC m=+214.144795884 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.701836 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-67594fc686-h5hwx"] Feb 27 17:04:13 crc kubenswrapper[4700]: E0227 17:04:13.702084 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0adc70e1-038b-45c1-80a8-9e1f938fc161" containerName="controller-manager" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.702103 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="0adc70e1-038b-45c1-80a8-9e1f938fc161" containerName="controller-manager" Feb 27 17:04:13 crc kubenswrapper[4700]: E0227 17:04:13.702114 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd3b00b4-f42e-406c-9b88-b7e2602a2493" containerName="route-controller-manager" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.702121 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd3b00b4-f42e-406c-9b88-b7e2602a2493" containerName="route-controller-manager" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.702204 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="0adc70e1-038b-45c1-80a8-9e1f938fc161" containerName="controller-manager" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.702217 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd3b00b4-f42e-406c-9b88-b7e2602a2493" containerName="route-controller-manager" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.702625 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-67594fc686-h5hwx" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.704996 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-85f5d5cff6-pprdc"] Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.705881 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.706294 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.706533 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.706675 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.706794 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.706912 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.713354 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-85f5d5cff6-pprdc" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.723282 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.723357 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.723634 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.723663 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.724359 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.724835 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.726274 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-85f5d5cff6-pprdc"] Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.731593 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.760711 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:13 crc kubenswrapper[4700]: E0227 17:04:13.760873 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:14.260848787 +0000 UTC m=+214.246161534 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.761012 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/033db5e9-76ee-4a35-a464-ae157bd38f89-proxy-ca-bundles\") pod \"controller-manager-85f5d5cff6-pprdc\" (UID: \"033db5e9-76ee-4a35-a464-ae157bd38f89\") " pod="openshift-controller-manager/controller-manager-85f5d5cff6-pprdc" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.761049 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gntlm\" (UniqueName: \"kubernetes.io/projected/033db5e9-76ee-4a35-a464-ae157bd38f89-kube-api-access-gntlm\") pod \"controller-manager-85f5d5cff6-pprdc\" (UID: \"033db5e9-76ee-4a35-a464-ae157bd38f89\") " pod="openshift-controller-manager/controller-manager-85f5d5cff6-pprdc" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.761076 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7fb68227-bec3-4159-81a3-9ac7929d74cf-serving-cert\") pod \"route-controller-manager-67594fc686-h5hwx\" (UID: \"7fb68227-bec3-4159-81a3-9ac7929d74cf\") " pod="openshift-route-controller-manager/route-controller-manager-67594fc686-h5hwx" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.761098 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/033db5e9-76ee-4a35-a464-ae157bd38f89-client-ca\") pod \"controller-manager-85f5d5cff6-pprdc\" (UID: \"033db5e9-76ee-4a35-a464-ae157bd38f89\") " pod="openshift-controller-manager/controller-manager-85f5d5cff6-pprdc" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.761135 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fb68227-bec3-4159-81a3-9ac7929d74cf-config\") pod \"route-controller-manager-67594fc686-h5hwx\" (UID: \"7fb68227-bec3-4159-81a3-9ac7929d74cf\") " pod="openshift-route-controller-manager/route-controller-manager-67594fc686-h5hwx" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.761194 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/033db5e9-76ee-4a35-a464-ae157bd38f89-serving-cert\") pod \"controller-manager-85f5d5cff6-pprdc\" (UID: \"033db5e9-76ee-4a35-a464-ae157bd38f89\") " pod="openshift-controller-manager/controller-manager-85f5d5cff6-pprdc" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.761214 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/033db5e9-76ee-4a35-a464-ae157bd38f89-config\") pod \"controller-manager-85f5d5cff6-pprdc\" (UID: \"033db5e9-76ee-4a35-a464-ae157bd38f89\") " pod="openshift-controller-manager/controller-manager-85f5d5cff6-pprdc" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.761240 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.761273 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6v4rn\" (UniqueName: \"kubernetes.io/projected/7fb68227-bec3-4159-81a3-9ac7929d74cf-kube-api-access-6v4rn\") pod \"route-controller-manager-67594fc686-h5hwx\" (UID: \"7fb68227-bec3-4159-81a3-9ac7929d74cf\") " pod="openshift-route-controller-manager/route-controller-manager-67594fc686-h5hwx" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.761308 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7fb68227-bec3-4159-81a3-9ac7929d74cf-client-ca\") pod \"route-controller-manager-67594fc686-h5hwx\" (UID: \"7fb68227-bec3-4159-81a3-9ac7929d74cf\") " pod="openshift-route-controller-manager/route-controller-manager-67594fc686-h5hwx" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.761390 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-67594fc686-h5hwx"] Feb 27 17:04:13 crc kubenswrapper[4700]: E0227 17:04:13.761879 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:14.261861034 +0000 UTC m=+214.247173781 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.862809 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:13 crc kubenswrapper[4700]: E0227 17:04:13.862985 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:14.362960267 +0000 UTC m=+214.348273014 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.863335 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7fb68227-bec3-4159-81a3-9ac7929d74cf-serving-cert\") pod \"route-controller-manager-67594fc686-h5hwx\" (UID: \"7fb68227-bec3-4159-81a3-9ac7929d74cf\") " pod="openshift-route-controller-manager/route-controller-manager-67594fc686-h5hwx" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.864065 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/033db5e9-76ee-4a35-a464-ae157bd38f89-client-ca\") pod \"controller-manager-85f5d5cff6-pprdc\" (UID: \"033db5e9-76ee-4a35-a464-ae157bd38f89\") " pod="openshift-controller-manager/controller-manager-85f5d5cff6-pprdc" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.864089 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fb68227-bec3-4159-81a3-9ac7929d74cf-config\") pod \"route-controller-manager-67594fc686-h5hwx\" (UID: \"7fb68227-bec3-4159-81a3-9ac7929d74cf\") " pod="openshift-route-controller-manager/route-controller-manager-67594fc686-h5hwx" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.864141 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/033db5e9-76ee-4a35-a464-ae157bd38f89-serving-cert\") pod \"controller-manager-85f5d5cff6-pprdc\" (UID: \"033db5e9-76ee-4a35-a464-ae157bd38f89\") " pod="openshift-controller-manager/controller-manager-85f5d5cff6-pprdc" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.864183 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/033db5e9-76ee-4a35-a464-ae157bd38f89-config\") pod \"controller-manager-85f5d5cff6-pprdc\" (UID: \"033db5e9-76ee-4a35-a464-ae157bd38f89\") " pod="openshift-controller-manager/controller-manager-85f5d5cff6-pprdc" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.864207 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.864256 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6v4rn\" (UniqueName: \"kubernetes.io/projected/7fb68227-bec3-4159-81a3-9ac7929d74cf-kube-api-access-6v4rn\") pod \"route-controller-manager-67594fc686-h5hwx\" (UID: \"7fb68227-bec3-4159-81a3-9ac7929d74cf\") " pod="openshift-route-controller-manager/route-controller-manager-67594fc686-h5hwx" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.864354 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7fb68227-bec3-4159-81a3-9ac7929d74cf-client-ca\") pod \"route-controller-manager-67594fc686-h5hwx\" (UID: \"7fb68227-bec3-4159-81a3-9ac7929d74cf\") " pod="openshift-route-controller-manager/route-controller-manager-67594fc686-h5hwx" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.864372 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/033db5e9-76ee-4a35-a464-ae157bd38f89-proxy-ca-bundles\") pod \"controller-manager-85f5d5cff6-pprdc\" (UID: \"033db5e9-76ee-4a35-a464-ae157bd38f89\") " pod="openshift-controller-manager/controller-manager-85f5d5cff6-pprdc" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.864396 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gntlm\" (UniqueName: \"kubernetes.io/projected/033db5e9-76ee-4a35-a464-ae157bd38f89-kube-api-access-gntlm\") pod \"controller-manager-85f5d5cff6-pprdc\" (UID: \"033db5e9-76ee-4a35-a464-ae157bd38f89\") " pod="openshift-controller-manager/controller-manager-85f5d5cff6-pprdc" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.864857 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/033db5e9-76ee-4a35-a464-ae157bd38f89-client-ca\") pod \"controller-manager-85f5d5cff6-pprdc\" (UID: \"033db5e9-76ee-4a35-a464-ae157bd38f89\") " pod="openshift-controller-manager/controller-manager-85f5d5cff6-pprdc" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.865331 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fb68227-bec3-4159-81a3-9ac7929d74cf-config\") pod \"route-controller-manager-67594fc686-h5hwx\" (UID: \"7fb68227-bec3-4159-81a3-9ac7929d74cf\") " pod="openshift-route-controller-manager/route-controller-manager-67594fc686-h5hwx" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.865528 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7fb68227-bec3-4159-81a3-9ac7929d74cf-client-ca\") pod \"route-controller-manager-67594fc686-h5hwx\" (UID: \"7fb68227-bec3-4159-81a3-9ac7929d74cf\") " pod="openshift-route-controller-manager/route-controller-manager-67594fc686-h5hwx" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.866066 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/033db5e9-76ee-4a35-a464-ae157bd38f89-proxy-ca-bundles\") pod \"controller-manager-85f5d5cff6-pprdc\" (UID: \"033db5e9-76ee-4a35-a464-ae157bd38f89\") " pod="openshift-controller-manager/controller-manager-85f5d5cff6-pprdc" Feb 27 17:04:13 crc kubenswrapper[4700]: E0227 17:04:13.866241 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:14.366228633 +0000 UTC m=+214.351541380 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.866565 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/033db5e9-76ee-4a35-a464-ae157bd38f89-config\") pod \"controller-manager-85f5d5cff6-pprdc\" (UID: \"033db5e9-76ee-4a35-a464-ae157bd38f89\") " pod="openshift-controller-manager/controller-manager-85f5d5cff6-pprdc" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.882553 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7fb68227-bec3-4159-81a3-9ac7929d74cf-serving-cert\") pod \"route-controller-manager-67594fc686-h5hwx\" (UID: \"7fb68227-bec3-4159-81a3-9ac7929d74cf\") " pod="openshift-route-controller-manager/route-controller-manager-67594fc686-h5hwx" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.883038 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/033db5e9-76ee-4a35-a464-ae157bd38f89-serving-cert\") pod \"controller-manager-85f5d5cff6-pprdc\" (UID: \"033db5e9-76ee-4a35-a464-ae157bd38f89\") " pod="openshift-controller-manager/controller-manager-85f5d5cff6-pprdc" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.887121 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gntlm\" (UniqueName: \"kubernetes.io/projected/033db5e9-76ee-4a35-a464-ae157bd38f89-kube-api-access-gntlm\") pod \"controller-manager-85f5d5cff6-pprdc\" (UID: \"033db5e9-76ee-4a35-a464-ae157bd38f89\") " pod="openshift-controller-manager/controller-manager-85f5d5cff6-pprdc" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.891231 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6v4rn\" (UniqueName: \"kubernetes.io/projected/7fb68227-bec3-4159-81a3-9ac7929d74cf-kube-api-access-6v4rn\") pod \"route-controller-manager-67594fc686-h5hwx\" (UID: \"7fb68227-bec3-4159-81a3-9ac7929d74cf\") " pod="openshift-route-controller-manager/route-controller-manager-67594fc686-h5hwx" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.965089 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:13 crc kubenswrapper[4700]: E0227 17:04:13.965343 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:14.465315334 +0000 UTC m=+214.450628081 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.965396 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.965443 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.965501 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.965563 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.965609 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:04:13 crc kubenswrapper[4700]: E0227 17:04:13.966338 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:14.46632661 +0000 UTC m=+214.451639357 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.966592 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.969934 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.970397 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:04:13 crc kubenswrapper[4700]: I0227 17:04:13.971062 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.015846 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-67594fc686-h5hwx" Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.026146 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-85f5d5cff6-pprdc" Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.067022 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.067309 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f5288317-4810-4321-baa0-31a118a4dc36-metrics-certs\") pod \"network-metrics-daemon-s8k74\" (UID: \"f5288317-4810-4321-baa0-31a118a4dc36\") " pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:04:14 crc kubenswrapper[4700]: E0227 17:04:14.068301 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:14.568284996 +0000 UTC m=+214.553597743 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.070583 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f5288317-4810-4321-baa0-31a118a4dc36-metrics-certs\") pod \"network-metrics-daemon-s8k74\" (UID: \"f5288317-4810-4321-baa0-31a118a4dc36\") " pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.101049 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-s8k74" Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.166784 4700 patch_prober.go:28] interesting pod/router-default-5444994796-58xmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 17:04:14 crc kubenswrapper[4700]: [-]has-synced failed: reason withheld Feb 27 17:04:14 crc kubenswrapper[4700]: [+]process-running ok Feb 27 17:04:14 crc kubenswrapper[4700]: healthz check failed Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.167097 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-58xmr" podUID="cce840a6-7988-43de-a6d7-07187abb65c3" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.168287 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:14 crc kubenswrapper[4700]: E0227 17:04:14.168597 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:14.668584829 +0000 UTC m=+214.653897576 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.209980 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.223569 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.281876 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:14 crc kubenswrapper[4700]: E0227 17:04:14.281946 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:14.781929054 +0000 UTC m=+214.767241801 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.282002 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.281876 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 17:04:14 crc kubenswrapper[4700]: E0227 17:04:14.282234 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:14.782225252 +0000 UTC m=+214.767537999 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.382634 4700 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.384621 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:14 crc kubenswrapper[4700]: E0227 17:04:14.385033 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:14.88501092 +0000 UTC m=+214.870323667 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.395762 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-blwkv" event={"ID":"ba36da60-1e6a-4010-a017-802d1cbbae71","Type":"ContainerStarted","Data":"ddc02e0e54cf40bb5431ffd1bccc44d3dcf6b74faef8dd00553acb614782f68d"} Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.430797 4700 generic.go:334] "Generic (PLEG): container finished" podID="66e8ed9b-99d5-4643-aa2a-4720ea4195b9" containerID="9cb12c74550ad13f426db5679571715af91b50eaa633f56ce6a1c7a8c66ae2d3" exitCode=0 Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.430962 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536860-dw5fg" event={"ID":"66e8ed9b-99d5-4643-aa2a-4720ea4195b9","Type":"ContainerDied","Data":"9cb12c74550ad13f426db5679571715af91b50eaa633f56ce6a1c7a8c66ae2d3"} Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.439778 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-hc86c" Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.486603 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:14 crc kubenswrapper[4700]: E0227 17:04:14.486991 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:14.986978336 +0000 UTC m=+214.972291083 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.565329 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9t75w"] Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.566413 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9t75w" Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.568614 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.572025 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9t75w"] Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.587240 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:14 crc kubenswrapper[4700]: E0227 17:04:14.588923 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:15.088902071 +0000 UTC m=+215.074214818 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.637592 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-67594fc686-h5hwx"] Feb 27 17:04:14 crc kubenswrapper[4700]: W0227 17:04:14.657813 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7fb68227_bec3_4159_81a3_9ac7929d74cf.slice/crio-bd8fd5ff966a88fca422973cae9526542d3edfe22ba52fa28683913d6b364c06 WatchSource:0}: Error finding container bd8fd5ff966a88fca422973cae9526542d3edfe22ba52fa28683913d6b364c06: Status 404 returned error can't find the container with id bd8fd5ff966a88fca422973cae9526542d3edfe22ba52fa28683913d6b364c06 Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.682040 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-85f5d5cff6-pprdc"] Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.688906 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.688943 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89845e2b-1804-445b-8462-36b2350ae663-catalog-content\") pod \"community-operators-9t75w\" (UID: \"89845e2b-1804-445b-8462-36b2350ae663\") " pod="openshift-marketplace/community-operators-9t75w" Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.688999 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsvvr\" (UniqueName: \"kubernetes.io/projected/89845e2b-1804-445b-8462-36b2350ae663-kube-api-access-qsvvr\") pod \"community-operators-9t75w\" (UID: \"89845e2b-1804-445b-8462-36b2350ae663\") " pod="openshift-marketplace/community-operators-9t75w" Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.689020 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89845e2b-1804-445b-8462-36b2350ae663-utilities\") pod \"community-operators-9t75w\" (UID: \"89845e2b-1804-445b-8462-36b2350ae663\") " pod="openshift-marketplace/community-operators-9t75w" Feb 27 17:04:14 crc kubenswrapper[4700]: E0227 17:04:14.689265 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:15.189254005 +0000 UTC m=+215.174566752 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:14 crc kubenswrapper[4700]: W0227 17:04:14.693010 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod033db5e9_76ee_4a35_a464_ae157bd38f89.slice/crio-be55c878c7aea505ddc8d0544faa51bad3fd624d512c58b9bd816f5fd30ff69d WatchSource:0}: Error finding container be55c878c7aea505ddc8d0544faa51bad3fd624d512c58b9bd816f5fd30ff69d: Status 404 returned error can't find the container with id be55c878c7aea505ddc8d0544faa51bad3fd624d512c58b9bd816f5fd30ff69d Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.694035 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-s8k74"] Feb 27 17:04:14 crc kubenswrapper[4700]: W0227 17:04:14.712059 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf5288317_4810_4321_baa0_31a118a4dc36.slice/crio-af3e40fd7a2460b3230f5d331a591320152eecc080d045511e97ab2c7062ca1d WatchSource:0}: Error finding container af3e40fd7a2460b3230f5d331a591320152eecc080d045511e97ab2c7062ca1d: Status 404 returned error can't find the container with id af3e40fd7a2460b3230f5d331a591320152eecc080d045511e97ab2c7062ca1d Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.761093 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5n6qh"] Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.762069 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5n6qh" Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.763682 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.774133 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5n6qh"] Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.791143 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.791247 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a72bdcb3-42e2-4a45-8c98-7b6e97afabab-catalog-content\") pod \"certified-operators-5n6qh\" (UID: \"a72bdcb3-42e2-4a45-8c98-7b6e97afabab\") " pod="openshift-marketplace/certified-operators-5n6qh" Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.791271 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsvvr\" (UniqueName: \"kubernetes.io/projected/89845e2b-1804-445b-8462-36b2350ae663-kube-api-access-qsvvr\") pod \"community-operators-9t75w\" (UID: \"89845e2b-1804-445b-8462-36b2350ae663\") " pod="openshift-marketplace/community-operators-9t75w" Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.791296 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89845e2b-1804-445b-8462-36b2350ae663-utilities\") pod \"community-operators-9t75w\" (UID: \"89845e2b-1804-445b-8462-36b2350ae663\") " pod="openshift-marketplace/community-operators-9t75w" Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.791321 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a72bdcb3-42e2-4a45-8c98-7b6e97afabab-utilities\") pod \"certified-operators-5n6qh\" (UID: \"a72bdcb3-42e2-4a45-8c98-7b6e97afabab\") " pod="openshift-marketplace/certified-operators-5n6qh" Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.791345 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89845e2b-1804-445b-8462-36b2350ae663-catalog-content\") pod \"community-operators-9t75w\" (UID: \"89845e2b-1804-445b-8462-36b2350ae663\") " pod="openshift-marketplace/community-operators-9t75w" Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.791379 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kwkl\" (UniqueName: \"kubernetes.io/projected/a72bdcb3-42e2-4a45-8c98-7b6e97afabab-kube-api-access-4kwkl\") pod \"certified-operators-5n6qh\" (UID: \"a72bdcb3-42e2-4a45-8c98-7b6e97afabab\") " pod="openshift-marketplace/certified-operators-5n6qh" Feb 27 17:04:14 crc kubenswrapper[4700]: E0227 17:04:14.791475 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:15.291447587 +0000 UTC m=+215.276760334 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.792295 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89845e2b-1804-445b-8462-36b2350ae663-utilities\") pod \"community-operators-9t75w\" (UID: \"89845e2b-1804-445b-8462-36b2350ae663\") " pod="openshift-marketplace/community-operators-9t75w" Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.796292 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89845e2b-1804-445b-8462-36b2350ae663-catalog-content\") pod \"community-operators-9t75w\" (UID: \"89845e2b-1804-445b-8462-36b2350ae663\") " pod="openshift-marketplace/community-operators-9t75w" Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.835066 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsvvr\" (UniqueName: \"kubernetes.io/projected/89845e2b-1804-445b-8462-36b2350ae663-kube-api-access-qsvvr\") pod \"community-operators-9t75w\" (UID: \"89845e2b-1804-445b-8462-36b2350ae663\") " pod="openshift-marketplace/community-operators-9t75w" Feb 27 17:04:14 crc kubenswrapper[4700]: W0227 17:04:14.847216 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-f98789fb3b251cad20d3d6f00785941b13220f2ad8b733d3e688eac5ff39ae2b WatchSource:0}: Error finding container f98789fb3b251cad20d3d6f00785941b13220f2ad8b733d3e688eac5ff39ae2b: Status 404 returned error can't find the container with id f98789fb3b251cad20d3d6f00785941b13220f2ad8b733d3e688eac5ff39ae2b Feb 27 17:04:14 crc kubenswrapper[4700]: W0227 17:04:14.853286 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-76390293bb187a24cd663d89757a7a3e662256fdf25423c189d9236f880e92b8 WatchSource:0}: Error finding container 76390293bb187a24cd663d89757a7a3e662256fdf25423c189d9236f880e92b8: Status 404 returned error can't find the container with id 76390293bb187a24cd663d89757a7a3e662256fdf25423c189d9236f880e92b8 Feb 27 17:04:14 crc kubenswrapper[4700]: W0227 17:04:14.854592 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-87705886a8e264a71a9e62c919298aa74c357efe8bee54c3485e5d2e7617ed53 WatchSource:0}: Error finding container 87705886a8e264a71a9e62c919298aa74c357efe8bee54c3485e5d2e7617ed53: Status 404 returned error can't find the container with id 87705886a8e264a71a9e62c919298aa74c357efe8bee54c3485e5d2e7617ed53 Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.892870 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kwkl\" (UniqueName: \"kubernetes.io/projected/a72bdcb3-42e2-4a45-8c98-7b6e97afabab-kube-api-access-4kwkl\") pod \"certified-operators-5n6qh\" (UID: \"a72bdcb3-42e2-4a45-8c98-7b6e97afabab\") " pod="openshift-marketplace/certified-operators-5n6qh" Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.892927 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a72bdcb3-42e2-4a45-8c98-7b6e97afabab-catalog-content\") pod \"certified-operators-5n6qh\" (UID: \"a72bdcb3-42e2-4a45-8c98-7b6e97afabab\") " pod="openshift-marketplace/certified-operators-5n6qh" Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.892963 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a72bdcb3-42e2-4a45-8c98-7b6e97afabab-utilities\") pod \"certified-operators-5n6qh\" (UID: \"a72bdcb3-42e2-4a45-8c98-7b6e97afabab\") " pod="openshift-marketplace/certified-operators-5n6qh" Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.892983 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:14 crc kubenswrapper[4700]: E0227 17:04:14.893251 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:15.393238929 +0000 UTC m=+215.378551676 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.894214 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9t75w" Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.896344 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a72bdcb3-42e2-4a45-8c98-7b6e97afabab-utilities\") pod \"certified-operators-5n6qh\" (UID: \"a72bdcb3-42e2-4a45-8c98-7b6e97afabab\") " pod="openshift-marketplace/certified-operators-5n6qh" Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.896534 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a72bdcb3-42e2-4a45-8c98-7b6e97afabab-catalog-content\") pod \"certified-operators-5n6qh\" (UID: \"a72bdcb3-42e2-4a45-8c98-7b6e97afabab\") " pod="openshift-marketplace/certified-operators-5n6qh" Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.909050 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kwkl\" (UniqueName: \"kubernetes.io/projected/a72bdcb3-42e2-4a45-8c98-7b6e97afabab-kube-api-access-4kwkl\") pod \"certified-operators-5n6qh\" (UID: \"a72bdcb3-42e2-4a45-8c98-7b6e97afabab\") " pod="openshift-marketplace/certified-operators-5n6qh" Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.966755 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fltgg"] Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.967802 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fltgg" Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.978264 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fltgg"] Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.993369 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:14 crc kubenswrapper[4700]: E0227 17:04:14.994546 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:15.494525637 +0000 UTC m=+215.479838384 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.997144 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0adc70e1-038b-45c1-80a8-9e1f938fc161" path="/var/lib/kubelet/pods/0adc70e1-038b-45c1-80a8-9e1f938fc161/volumes" Feb 27 17:04:14 crc kubenswrapper[4700]: I0227 17:04:14.998981 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd3b00b4-f42e-406c-9b88-b7e2602a2493" path="/var/lib/kubelet/pods/dd3b00b4-f42e-406c-9b88-b7e2602a2493/volumes" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.014988 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.018748 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.022430 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.022634 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.056378 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.097084 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.097136 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqkjd\" (UniqueName: \"kubernetes.io/projected/bc1137d0-2659-430b-8d76-d08017e7f133-kube-api-access-lqkjd\") pod \"community-operators-fltgg\" (UID: \"bc1137d0-2659-430b-8d76-d08017e7f133\") " pod="openshift-marketplace/community-operators-fltgg" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.097166 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc1137d0-2659-430b-8d76-d08017e7f133-utilities\") pod \"community-operators-fltgg\" (UID: \"bc1137d0-2659-430b-8d76-d08017e7f133\") " pod="openshift-marketplace/community-operators-fltgg" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.097198 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/49b065e4-14eb-4416-bd24-2164635669d7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"49b065e4-14eb-4416-bd24-2164635669d7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.097233 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc1137d0-2659-430b-8d76-d08017e7f133-catalog-content\") pod \"community-operators-fltgg\" (UID: \"bc1137d0-2659-430b-8d76-d08017e7f133\") " pod="openshift-marketplace/community-operators-fltgg" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.097272 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/49b065e4-14eb-4416-bd24-2164635669d7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"49b065e4-14eb-4416-bd24-2164635669d7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 17:04:15 crc kubenswrapper[4700]: E0227 17:04:15.097602 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 17:04:15.597591702 +0000 UTC m=+215.582904449 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k97gv" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.121315 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5n6qh" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.169535 4700 patch_prober.go:28] interesting pod/router-default-5444994796-58xmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 17:04:15 crc kubenswrapper[4700]: [-]has-synced failed: reason withheld Feb 27 17:04:15 crc kubenswrapper[4700]: [+]process-running ok Feb 27 17:04:15 crc kubenswrapper[4700]: healthz check failed Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.169584 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-58xmr" podUID="cce840a6-7988-43de-a6d7-07187abb65c3" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.172298 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-j8wlz"] Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.178393 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j8wlz" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.198157 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.198700 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqkjd\" (UniqueName: \"kubernetes.io/projected/bc1137d0-2659-430b-8d76-d08017e7f133-kube-api-access-lqkjd\") pod \"community-operators-fltgg\" (UID: \"bc1137d0-2659-430b-8d76-d08017e7f133\") " pod="openshift-marketplace/community-operators-fltgg" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.198737 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc1137d0-2659-430b-8d76-d08017e7f133-utilities\") pod \"community-operators-fltgg\" (UID: \"bc1137d0-2659-430b-8d76-d08017e7f133\") " pod="openshift-marketplace/community-operators-fltgg" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.198757 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/49b065e4-14eb-4416-bd24-2164635669d7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"49b065e4-14eb-4416-bd24-2164635669d7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.198778 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d67j9\" (UniqueName: \"kubernetes.io/projected/d1c12d33-6674-4c8a-b871-4d95efca7c23-kube-api-access-d67j9\") pod \"certified-operators-j8wlz\" (UID: \"d1c12d33-6674-4c8a-b871-4d95efca7c23\") " pod="openshift-marketplace/certified-operators-j8wlz" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.198801 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1c12d33-6674-4c8a-b871-4d95efca7c23-utilities\") pod \"certified-operators-j8wlz\" (UID: \"d1c12d33-6674-4c8a-b871-4d95efca7c23\") " pod="openshift-marketplace/certified-operators-j8wlz" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.198831 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc1137d0-2659-430b-8d76-d08017e7f133-catalog-content\") pod \"community-operators-fltgg\" (UID: \"bc1137d0-2659-430b-8d76-d08017e7f133\") " pod="openshift-marketplace/community-operators-fltgg" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.198854 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/49b065e4-14eb-4416-bd24-2164635669d7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"49b065e4-14eb-4416-bd24-2164635669d7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.198871 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1c12d33-6674-4c8a-b871-4d95efca7c23-catalog-content\") pod \"certified-operators-j8wlz\" (UID: \"d1c12d33-6674-4c8a-b871-4d95efca7c23\") " pod="openshift-marketplace/certified-operators-j8wlz" Feb 27 17:04:15 crc kubenswrapper[4700]: E0227 17:04:15.198973 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 17:04:15.698959243 +0000 UTC m=+215.684271990 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.199347 4700 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-02-27T17:04:14.382660558Z","Handler":null,"Name":""} Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.200165 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/49b065e4-14eb-4416-bd24-2164635669d7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"49b065e4-14eb-4416-bd24-2164635669d7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.201697 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc1137d0-2659-430b-8d76-d08017e7f133-utilities\") pod \"community-operators-fltgg\" (UID: \"bc1137d0-2659-430b-8d76-d08017e7f133\") " pod="openshift-marketplace/community-operators-fltgg" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.202317 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j8wlz"] Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.204271 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc1137d0-2659-430b-8d76-d08017e7f133-catalog-content\") pod \"community-operators-fltgg\" (UID: \"bc1137d0-2659-430b-8d76-d08017e7f133\") " pod="openshift-marketplace/community-operators-fltgg" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.232737 4700 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.233406 4700 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.242990 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqkjd\" (UniqueName: \"kubernetes.io/projected/bc1137d0-2659-430b-8d76-d08017e7f133-kube-api-access-lqkjd\") pod \"community-operators-fltgg\" (UID: \"bc1137d0-2659-430b-8d76-d08017e7f133\") " pod="openshift-marketplace/community-operators-fltgg" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.264712 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/49b065e4-14eb-4416-bd24-2164635669d7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"49b065e4-14eb-4416-bd24-2164635669d7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.294109 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fltgg" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.300247 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d67j9\" (UniqueName: \"kubernetes.io/projected/d1c12d33-6674-4c8a-b871-4d95efca7c23-kube-api-access-d67j9\") pod \"certified-operators-j8wlz\" (UID: \"d1c12d33-6674-4c8a-b871-4d95efca7c23\") " pod="openshift-marketplace/certified-operators-j8wlz" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.300285 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1c12d33-6674-4c8a-b871-4d95efca7c23-utilities\") pod \"certified-operators-j8wlz\" (UID: \"d1c12d33-6674-4c8a-b871-4d95efca7c23\") " pod="openshift-marketplace/certified-operators-j8wlz" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.300677 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1c12d33-6674-4c8a-b871-4d95efca7c23-catalog-content\") pod \"certified-operators-j8wlz\" (UID: \"d1c12d33-6674-4c8a-b871-4d95efca7c23\") " pod="openshift-marketplace/certified-operators-j8wlz" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.300729 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.303984 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1c12d33-6674-4c8a-b871-4d95efca7c23-utilities\") pod \"certified-operators-j8wlz\" (UID: \"d1c12d33-6674-4c8a-b871-4d95efca7c23\") " pod="openshift-marketplace/certified-operators-j8wlz" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.304034 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1c12d33-6674-4c8a-b871-4d95efca7c23-catalog-content\") pod \"certified-operators-j8wlz\" (UID: \"d1c12d33-6674-4c8a-b871-4d95efca7c23\") " pod="openshift-marketplace/certified-operators-j8wlz" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.306591 4700 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.306698 4700 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.339254 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d67j9\" (UniqueName: \"kubernetes.io/projected/d1c12d33-6674-4c8a-b871-4d95efca7c23-kube-api-access-d67j9\") pod \"certified-operators-j8wlz\" (UID: \"d1c12d33-6674-4c8a-b871-4d95efca7c23\") " pod="openshift-marketplace/certified-operators-j8wlz" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.354737 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.474037 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"4e46712edab6c70ffc5ec5aff649786be9873ef6c18a2f4832caf72bfc62fab7"} Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.474416 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"87705886a8e264a71a9e62c919298aa74c357efe8bee54c3485e5d2e7617ed53"} Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.474771 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.476834 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-67594fc686-h5hwx" event={"ID":"7fb68227-bec3-4159-81a3-9ac7929d74cf","Type":"ContainerStarted","Data":"11857bf45c0c8569ef62772879d1bcb1c50f716d270283c62982234fcf908eec"} Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.476870 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-67594fc686-h5hwx" event={"ID":"7fb68227-bec3-4159-81a3-9ac7929d74cf","Type":"ContainerStarted","Data":"bd8fd5ff966a88fca422973cae9526542d3edfe22ba52fa28683913d6b364c06"} Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.477450 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-67594fc686-h5hwx" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.480640 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-s8k74" event={"ID":"f5288317-4810-4321-baa0-31a118a4dc36","Type":"ContainerStarted","Data":"26f3d4046419522877bd68c325bf0679af1ef3b6092980eabdc66e2694f01f3e"} Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.480754 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-s8k74" event={"ID":"f5288317-4810-4321-baa0-31a118a4dc36","Type":"ContainerStarted","Data":"af3e40fd7a2460b3230f5d331a591320152eecc080d045511e97ab2c7062ca1d"} Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.486543 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-67594fc686-h5hwx" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.488332 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k97gv\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.488595 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-blwkv" event={"ID":"ba36da60-1e6a-4010-a017-802d1cbbae71","Type":"ContainerStarted","Data":"0f82aa7ac8fdc626a59b4bb88323221816f23c8b5e8699f447e0e9dfee50efc6"} Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.488620 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-blwkv" event={"ID":"ba36da60-1e6a-4010-a017-802d1cbbae71","Type":"ContainerStarted","Data":"961e920aeba45d132e7adaa8757d6b4695a4b27b0508efa7eb617c0cd926f3dd"} Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.491052 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"63ba6010684ca50e975ee9fdf8624be65c81d9cdbb585495e18e193cd8699f9a"} Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.491091 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"76390293bb187a24cd663d89757a7a3e662256fdf25423c189d9236f880e92b8"} Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.492606 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"5e507a8a3648e7f30707a626f2cd83f18a75e6008ac274a2626b041b2ec149e0"} Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.492630 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"f98789fb3b251cad20d3d6f00785941b13220f2ad8b733d3e688eac5ff39ae2b"} Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.495984 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-85f5d5cff6-pprdc" event={"ID":"033db5e9-76ee-4a35-a464-ae157bd38f89","Type":"ContainerStarted","Data":"f486e51340d3ac1b5e15ce74c492a7ae3f3a4f13c2a63dc816291bd9b41eecb9"} Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.496012 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-85f5d5cff6-pprdc" event={"ID":"033db5e9-76ee-4a35-a464-ae157bd38f89","Type":"ContainerStarted","Data":"be55c878c7aea505ddc8d0544faa51bad3fd624d512c58b9bd816f5fd30ff69d"} Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.497070 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-85f5d5cff6-pprdc" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.503325 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.511723 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-85f5d5cff6-pprdc" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.516530 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.525761 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9t75w"] Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.541590 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j8wlz" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.560880 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-blwkv" podStartSLOduration=11.560861562 podStartE2EDuration="11.560861562s" podCreationTimestamp="2026-02-27 17:04:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:15.544445661 +0000 UTC m=+215.529758408" watchObservedRunningTime="2026-02-27 17:04:15.560861562 +0000 UTC m=+215.546174309" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.561176 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-67594fc686-h5hwx" podStartSLOduration=3.56117255 podStartE2EDuration="3.56117255s" podCreationTimestamp="2026-02-27 17:04:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:15.558214592 +0000 UTC m=+215.543527339" watchObservedRunningTime="2026-02-27 17:04:15.56117255 +0000 UTC m=+215.546485287" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.578642 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-85f5d5cff6-pprdc" podStartSLOduration=3.578621948 podStartE2EDuration="3.578621948s" podCreationTimestamp="2026-02-27 17:04:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:15.576303877 +0000 UTC m=+215.561616624" watchObservedRunningTime="2026-02-27 17:04:15.578621948 +0000 UTC m=+215.563934695" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.595904 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:15 crc kubenswrapper[4700]: I0227 17:04:15.945687 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536860-dw5fg" Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.009535 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fltgg"] Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.009692 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/66e8ed9b-99d5-4643-aa2a-4720ea4195b9-config-volume\") pod \"66e8ed9b-99d5-4643-aa2a-4720ea4195b9\" (UID: \"66e8ed9b-99d5-4643-aa2a-4720ea4195b9\") " Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.009771 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9zcl\" (UniqueName: \"kubernetes.io/projected/66e8ed9b-99d5-4643-aa2a-4720ea4195b9-kube-api-access-h9zcl\") pod \"66e8ed9b-99d5-4643-aa2a-4720ea4195b9\" (UID: \"66e8ed9b-99d5-4643-aa2a-4720ea4195b9\") " Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.009852 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/66e8ed9b-99d5-4643-aa2a-4720ea4195b9-secret-volume\") pod \"66e8ed9b-99d5-4643-aa2a-4720ea4195b9\" (UID: \"66e8ed9b-99d5-4643-aa2a-4720ea4195b9\") " Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.010585 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66e8ed9b-99d5-4643-aa2a-4720ea4195b9-config-volume" (OuterVolumeSpecName: "config-volume") pod "66e8ed9b-99d5-4643-aa2a-4720ea4195b9" (UID: "66e8ed9b-99d5-4643-aa2a-4720ea4195b9"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.014883 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5n6qh"] Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.018159 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66e8ed9b-99d5-4643-aa2a-4720ea4195b9-kube-api-access-h9zcl" (OuterVolumeSpecName: "kube-api-access-h9zcl") pod "66e8ed9b-99d5-4643-aa2a-4720ea4195b9" (UID: "66e8ed9b-99d5-4643-aa2a-4720ea4195b9"). InnerVolumeSpecName "kube-api-access-h9zcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.018674 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66e8ed9b-99d5-4643-aa2a-4720ea4195b9-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "66e8ed9b-99d5-4643-aa2a-4720ea4195b9" (UID: "66e8ed9b-99d5-4643-aa2a-4720ea4195b9"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:04:16 crc kubenswrapper[4700]: W0227 17:04:16.024977 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbc1137d0_2659_430b_8d76_d08017e7f133.slice/crio-4239dc6fe2b6b1c90b3c6c4f883793b5d34c72d842095d1b21780c9c04dc6646 WatchSource:0}: Error finding container 4239dc6fe2b6b1c90b3c6c4f883793b5d34c72d842095d1b21780c9c04dc6646: Status 404 returned error can't find the container with id 4239dc6fe2b6b1c90b3c6c4f883793b5d34c72d842095d1b21780c9c04dc6646 Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.036051 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.041623 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-6wfhz" Feb 27 17:04:16 crc kubenswrapper[4700]: W0227 17:04:16.048583 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda72bdcb3_42e2_4a45_8c98_7b6e97afabab.slice/crio-dc416ff2c3de8978b2e6f3e281afa5c3b969ec0e08913bacb226109e67ad1237 WatchSource:0}: Error finding container dc416ff2c3de8978b2e6f3e281afa5c3b969ec0e08913bacb226109e67ad1237: Status 404 returned error can't find the container with id dc416ff2c3de8978b2e6f3e281afa5c3b969ec0e08913bacb226109e67ad1237 Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.112316 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9zcl\" (UniqueName: \"kubernetes.io/projected/66e8ed9b-99d5-4643-aa2a-4720ea4195b9-kube-api-access-h9zcl\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.112355 4700 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/66e8ed9b-99d5-4643-aa2a-4720ea4195b9-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.112366 4700 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/66e8ed9b-99d5-4643-aa2a-4720ea4195b9-config-volume\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.155821 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.164049 4700 patch_prober.go:28] interesting pod/router-default-5444994796-58xmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 17:04:16 crc kubenswrapper[4700]: [-]has-synced failed: reason withheld Feb 27 17:04:16 crc kubenswrapper[4700]: [+]process-running ok Feb 27 17:04:16 crc kubenswrapper[4700]: healthz check failed Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.164094 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-58xmr" podUID="cce840a6-7988-43de-a6d7-07187abb65c3" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.254415 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j8wlz"] Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.256859 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-k97gv"] Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.274086 4700 patch_prober.go:28] interesting pod/downloads-7954f5f757-v4nr9 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.274127 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-v4nr9" podUID="af74d952-25fb-4e1a-ab91-9c606b4b00ab" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.274174 4700 patch_prober.go:28] interesting pod/downloads-7954f5f757-v4nr9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.274202 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-v4nr9" podUID="af74d952-25fb-4e1a-ab91-9c606b4b00ab" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.504420 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"49b065e4-14eb-4416-bd24-2164635669d7","Type":"ContainerStarted","Data":"4e732c4a5499d47eef6bb2aabf3654b6ac091d0dd5661c94bd33b56d0aec763f"} Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.506354 4700 generic.go:334] "Generic (PLEG): container finished" podID="89845e2b-1804-445b-8462-36b2350ae663" containerID="b75bbb481dfa12e9f0984d0c52f051571dc47d60b84622ed1ebe8e65b6de2783" exitCode=0 Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.506396 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9t75w" event={"ID":"89845e2b-1804-445b-8462-36b2350ae663","Type":"ContainerDied","Data":"b75bbb481dfa12e9f0984d0c52f051571dc47d60b84622ed1ebe8e65b6de2783"} Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.506413 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9t75w" event={"ID":"89845e2b-1804-445b-8462-36b2350ae663","Type":"ContainerStarted","Data":"d8d53a3427897d1d1b8664f4d398b544f13f89bf8d97f79d4d2f9a125ecc3d8a"} Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.519641 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" event={"ID":"ee57d6e4-3550-46d0-947e-fe6db5cf3291","Type":"ContainerStarted","Data":"53a859b46b5d3051299ef73410d20f19b56dfdfc84bc08c14fd21bbda1c8a1e5"} Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.519695 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" event={"ID":"ee57d6e4-3550-46d0-947e-fe6db5cf3291","Type":"ContainerStarted","Data":"90e87d19965ec592df14b6d13c12968724f1a43bbfa6d69fbb0fcbbdf57aa07f"} Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.520368 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.525800 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-s8k74" event={"ID":"f5288317-4810-4321-baa0-31a118a4dc36","Type":"ContainerStarted","Data":"f194650601ff1c202c559e9e9ed609126a07a5182057fd53a6d58fa30906dfcc"} Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.534557 4700 generic.go:334] "Generic (PLEG): container finished" podID="d1c12d33-6674-4c8a-b871-4d95efca7c23" containerID="8f69a80ab7dcdcb9827cb4ec7403a9a897071125d1ea02b6f73a714cc2d75bec" exitCode=0 Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.534646 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j8wlz" event={"ID":"d1c12d33-6674-4c8a-b871-4d95efca7c23","Type":"ContainerDied","Data":"8f69a80ab7dcdcb9827cb4ec7403a9a897071125d1ea02b6f73a714cc2d75bec"} Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.534675 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j8wlz" event={"ID":"d1c12d33-6674-4c8a-b871-4d95efca7c23","Type":"ContainerStarted","Data":"0940a658aa8c9b190adf4d90142fd96c7e440f4bdfaf73f142c5e54198590b5f"} Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.541314 4700 generic.go:334] "Generic (PLEG): container finished" podID="a72bdcb3-42e2-4a45-8c98-7b6e97afabab" containerID="59d9e1f169d44712ad38898168ed8a7ece3561c0bc8f0930a718b175778cad8c" exitCode=0 Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.541390 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5n6qh" event={"ID":"a72bdcb3-42e2-4a45-8c98-7b6e97afabab","Type":"ContainerDied","Data":"59d9e1f169d44712ad38898168ed8a7ece3561c0bc8f0930a718b175778cad8c"} Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.541418 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5n6qh" event={"ID":"a72bdcb3-42e2-4a45-8c98-7b6e97afabab","Type":"ContainerStarted","Data":"dc416ff2c3de8978b2e6f3e281afa5c3b969ec0e08913bacb226109e67ad1237"} Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.544191 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536860-dw5fg" event={"ID":"66e8ed9b-99d5-4643-aa2a-4720ea4195b9","Type":"ContainerDied","Data":"e54141cd5fbadf9d90d3eddd6880b75ec71d3712eaee22ece681a851909fd9ca"} Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.544212 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e54141cd5fbadf9d90d3eddd6880b75ec71d3712eaee22ece681a851909fd9ca" Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.544274 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536860-dw5fg" Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.557071 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-s8k74" podStartSLOduration=162.557054418 podStartE2EDuration="2m42.557054418s" podCreationTimestamp="2026-02-27 17:01:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:16.554899631 +0000 UTC m=+216.540212378" watchObservedRunningTime="2026-02-27 17:04:16.557054418 +0000 UTC m=+216.542367165" Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.559640 4700 generic.go:334] "Generic (PLEG): container finished" podID="bc1137d0-2659-430b-8d76-d08017e7f133" containerID="038784b5566875c18e91eea7fd9f0fe0438c8f7d149a8a3a40fb10426e904368" exitCode=0 Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.560625 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fltgg" event={"ID":"bc1137d0-2659-430b-8d76-d08017e7f133","Type":"ContainerDied","Data":"038784b5566875c18e91eea7fd9f0fe0438c8f7d149a8a3a40fb10426e904368"} Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.560657 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fltgg" event={"ID":"bc1137d0-2659-430b-8d76-d08017e7f133","Type":"ContainerStarted","Data":"4239dc6fe2b6b1c90b3c6c4f883793b5d34c72d842095d1b21780c9c04dc6646"} Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.572863 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" podStartSLOduration=161.572840142 podStartE2EDuration="2m41.572840142s" podCreationTimestamp="2026-02-27 17:01:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:16.569235217 +0000 UTC m=+216.554547964" watchObservedRunningTime="2026-02-27 17:04:16.572840142 +0000 UTC m=+216.558152889" Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.767410 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gpb6h"] Feb 27 17:04:16 crc kubenswrapper[4700]: E0227 17:04:16.769924 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66e8ed9b-99d5-4643-aa2a-4720ea4195b9" containerName="collect-profiles" Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.769945 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="66e8ed9b-99d5-4643-aa2a-4720ea4195b9" containerName="collect-profiles" Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.770064 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="66e8ed9b-99d5-4643-aa2a-4720ea4195b9" containerName="collect-profiles" Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.771285 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gpb6h" Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.773720 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.774690 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gpb6h"] Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.925722 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2-utilities\") pod \"redhat-marketplace-gpb6h\" (UID: \"8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2\") " pod="openshift-marketplace/redhat-marketplace-gpb6h" Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.925832 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2-catalog-content\") pod \"redhat-marketplace-gpb6h\" (UID: \"8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2\") " pod="openshift-marketplace/redhat-marketplace-gpb6h" Feb 27 17:04:16 crc kubenswrapper[4700]: I0227 17:04:16.926033 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hq655\" (UniqueName: \"kubernetes.io/projected/8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2-kube-api-access-hq655\") pod \"redhat-marketplace-gpb6h\" (UID: \"8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2\") " pod="openshift-marketplace/redhat-marketplace-gpb6h" Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:16.999774 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:17.026935 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hq655\" (UniqueName: \"kubernetes.io/projected/8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2-kube-api-access-hq655\") pod \"redhat-marketplace-gpb6h\" (UID: \"8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2\") " pod="openshift-marketplace/redhat-marketplace-gpb6h" Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:17.026984 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2-utilities\") pod \"redhat-marketplace-gpb6h\" (UID: \"8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2\") " pod="openshift-marketplace/redhat-marketplace-gpb6h" Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:17.027031 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2-catalog-content\") pod \"redhat-marketplace-gpb6h\" (UID: \"8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2\") " pod="openshift-marketplace/redhat-marketplace-gpb6h" Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:17.027529 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2-catalog-content\") pod \"redhat-marketplace-gpb6h\" (UID: \"8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2\") " pod="openshift-marketplace/redhat-marketplace-gpb6h" Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:17.027894 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2-utilities\") pod \"redhat-marketplace-gpb6h\" (UID: \"8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2\") " pod="openshift-marketplace/redhat-marketplace-gpb6h" Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:17.055061 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hq655\" (UniqueName: \"kubernetes.io/projected/8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2-kube-api-access-hq655\") pod \"redhat-marketplace-gpb6h\" (UID: \"8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2\") " pod="openshift-marketplace/redhat-marketplace-gpb6h" Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:17.097016 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gpb6h" Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:17.161922 4700 patch_prober.go:28] interesting pod/router-default-5444994796-58xmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 17:04:17 crc kubenswrapper[4700]: [-]has-synced failed: reason withheld Feb 27 17:04:17 crc kubenswrapper[4700]: [+]process-running ok Feb 27 17:04:17 crc kubenswrapper[4700]: healthz check failed Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:17.161993 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-58xmr" podUID="cce840a6-7988-43de-a6d7-07187abb65c3" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:17.177362 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8bf7d"] Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:17.188689 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8bf7d"] Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:17.189102 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8bf7d" Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:17.330183 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fd07395-2d50-418c-a75a-64f1f149b216-utilities\") pod \"redhat-marketplace-8bf7d\" (UID: \"1fd07395-2d50-418c-a75a-64f1f149b216\") " pod="openshift-marketplace/redhat-marketplace-8bf7d" Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:17.330564 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fd07395-2d50-418c-a75a-64f1f149b216-catalog-content\") pod \"redhat-marketplace-8bf7d\" (UID: \"1fd07395-2d50-418c-a75a-64f1f149b216\") " pod="openshift-marketplace/redhat-marketplace-8bf7d" Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:17.330596 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwgkj\" (UniqueName: \"kubernetes.io/projected/1fd07395-2d50-418c-a75a-64f1f149b216-kube-api-access-rwgkj\") pod \"redhat-marketplace-8bf7d\" (UID: \"1fd07395-2d50-418c-a75a-64f1f149b216\") " pod="openshift-marketplace/redhat-marketplace-8bf7d" Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:17.353478 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gpb6h"] Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:17.431664 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fd07395-2d50-418c-a75a-64f1f149b216-utilities\") pod \"redhat-marketplace-8bf7d\" (UID: \"1fd07395-2d50-418c-a75a-64f1f149b216\") " pod="openshift-marketplace/redhat-marketplace-8bf7d" Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:17.431716 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fd07395-2d50-418c-a75a-64f1f149b216-catalog-content\") pod \"redhat-marketplace-8bf7d\" (UID: \"1fd07395-2d50-418c-a75a-64f1f149b216\") " pod="openshift-marketplace/redhat-marketplace-8bf7d" Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:17.431744 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwgkj\" (UniqueName: \"kubernetes.io/projected/1fd07395-2d50-418c-a75a-64f1f149b216-kube-api-access-rwgkj\") pod \"redhat-marketplace-8bf7d\" (UID: \"1fd07395-2d50-418c-a75a-64f1f149b216\") " pod="openshift-marketplace/redhat-marketplace-8bf7d" Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:17.432632 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fd07395-2d50-418c-a75a-64f1f149b216-utilities\") pod \"redhat-marketplace-8bf7d\" (UID: \"1fd07395-2d50-418c-a75a-64f1f149b216\") " pod="openshift-marketplace/redhat-marketplace-8bf7d" Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:17.432707 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fd07395-2d50-418c-a75a-64f1f149b216-catalog-content\") pod \"redhat-marketplace-8bf7d\" (UID: \"1fd07395-2d50-418c-a75a-64f1f149b216\") " pod="openshift-marketplace/redhat-marketplace-8bf7d" Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:17.450530 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwgkj\" (UniqueName: \"kubernetes.io/projected/1fd07395-2d50-418c-a75a-64f1f149b216-kube-api-access-rwgkj\") pod \"redhat-marketplace-8bf7d\" (UID: \"1fd07395-2d50-418c-a75a-64f1f149b216\") " pod="openshift-marketplace/redhat-marketplace-8bf7d" Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:17.486112 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-kq9sb" Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:17.486321 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-kq9sb" Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:17.488651 4700 patch_prober.go:28] interesting pod/console-f9d7485db-kq9sb container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.23:8443/health\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:17.488701 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-kq9sb" podUID="facc8439-0b4d-4228-95b2-d01d48ee1c53" containerName="console" probeResult="failure" output="Get \"https://10.217.0.23:8443/health\": dial tcp 10.217.0.23:8443: connect: connection refused" Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:17.509339 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8bf7d" Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:17.593671 4700 generic.go:334] "Generic (PLEG): container finished" podID="49b065e4-14eb-4416-bd24-2164635669d7" containerID="d68325b724caf03a0e909f0d8393e1cc1e23ae594029c2f407a98a5e9d6a6968" exitCode=0 Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:17.593998 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"49b065e4-14eb-4416-bd24-2164635669d7","Type":"ContainerDied","Data":"d68325b724caf03a0e909f0d8393e1cc1e23ae594029c2f407a98a5e9d6a6968"} Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:17.596792 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gpb6h" event={"ID":"8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2","Type":"ContainerStarted","Data":"864cd6a11e2a2f8e31a8c1fe61b497b560567f12393e162ef383a2df5b73c9c5"} Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:17.771978 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ltq9s"] Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:17.774564 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ltq9s" Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:17.777611 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ltq9s"] Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:17.778091 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:17.943537 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fppfr\" (UniqueName: \"kubernetes.io/projected/fe1b00ba-9936-45d5-9a6e-f4f93ce09b44-kube-api-access-fppfr\") pod \"redhat-operators-ltq9s\" (UID: \"fe1b00ba-9936-45d5-9a6e-f4f93ce09b44\") " pod="openshift-marketplace/redhat-operators-ltq9s" Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:17.943575 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe1b00ba-9936-45d5-9a6e-f4f93ce09b44-catalog-content\") pod \"redhat-operators-ltq9s\" (UID: \"fe1b00ba-9936-45d5-9a6e-f4f93ce09b44\") " pod="openshift-marketplace/redhat-operators-ltq9s" Feb 27 17:04:17 crc kubenswrapper[4700]: I0227 17:04:17.943605 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe1b00ba-9936-45d5-9a6e-f4f93ce09b44-utilities\") pod \"redhat-operators-ltq9s\" (UID: \"fe1b00ba-9936-45d5-9a6e-f4f93ce09b44\") " pod="openshift-marketplace/redhat-operators-ltq9s" Feb 27 17:04:18 crc kubenswrapper[4700]: I0227 17:04:18.040867 4700 ???:1] "http: TLS handshake error from 192.168.126.11:43080: no serving certificate available for the kubelet" Feb 27 17:04:18 crc kubenswrapper[4700]: I0227 17:04:18.044747 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fppfr\" (UniqueName: \"kubernetes.io/projected/fe1b00ba-9936-45d5-9a6e-f4f93ce09b44-kube-api-access-fppfr\") pod \"redhat-operators-ltq9s\" (UID: \"fe1b00ba-9936-45d5-9a6e-f4f93ce09b44\") " pod="openshift-marketplace/redhat-operators-ltq9s" Feb 27 17:04:18 crc kubenswrapper[4700]: I0227 17:04:18.045245 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe1b00ba-9936-45d5-9a6e-f4f93ce09b44-catalog-content\") pod \"redhat-operators-ltq9s\" (UID: \"fe1b00ba-9936-45d5-9a6e-f4f93ce09b44\") " pod="openshift-marketplace/redhat-operators-ltq9s" Feb 27 17:04:18 crc kubenswrapper[4700]: I0227 17:04:18.045311 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe1b00ba-9936-45d5-9a6e-f4f93ce09b44-utilities\") pod \"redhat-operators-ltq9s\" (UID: \"fe1b00ba-9936-45d5-9a6e-f4f93ce09b44\") " pod="openshift-marketplace/redhat-operators-ltq9s" Feb 27 17:04:18 crc kubenswrapper[4700]: I0227 17:04:18.045631 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe1b00ba-9936-45d5-9a6e-f4f93ce09b44-catalog-content\") pod \"redhat-operators-ltq9s\" (UID: \"fe1b00ba-9936-45d5-9a6e-f4f93ce09b44\") " pod="openshift-marketplace/redhat-operators-ltq9s" Feb 27 17:04:18 crc kubenswrapper[4700]: I0227 17:04:18.045770 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe1b00ba-9936-45d5-9a6e-f4f93ce09b44-utilities\") pod \"redhat-operators-ltq9s\" (UID: \"fe1b00ba-9936-45d5-9a6e-f4f93ce09b44\") " pod="openshift-marketplace/redhat-operators-ltq9s" Feb 27 17:04:18 crc kubenswrapper[4700]: I0227 17:04:18.075545 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fppfr\" (UniqueName: \"kubernetes.io/projected/fe1b00ba-9936-45d5-9a6e-f4f93ce09b44-kube-api-access-fppfr\") pod \"redhat-operators-ltq9s\" (UID: \"fe1b00ba-9936-45d5-9a6e-f4f93ce09b44\") " pod="openshift-marketplace/redhat-operators-ltq9s" Feb 27 17:04:18 crc kubenswrapper[4700]: I0227 17:04:18.093305 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ltq9s" Feb 27 17:04:18 crc kubenswrapper[4700]: I0227 17:04:18.103658 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 27 17:04:18 crc kubenswrapper[4700]: I0227 17:04:18.104246 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 17:04:18 crc kubenswrapper[4700]: I0227 17:04:18.106229 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 27 17:04:18 crc kubenswrapper[4700]: I0227 17:04:18.107539 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 27 17:04:18 crc kubenswrapper[4700]: I0227 17:04:18.126595 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 27 17:04:18 crc kubenswrapper[4700]: I0227 17:04:18.159213 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-58xmr" Feb 27 17:04:18 crc kubenswrapper[4700]: I0227 17:04:18.162151 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-f4s6x"] Feb 27 17:04:18 crc kubenswrapper[4700]: I0227 17:04:18.162492 4700 patch_prober.go:28] interesting pod/router-default-5444994796-58xmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 17:04:18 crc kubenswrapper[4700]: [-]has-synced failed: reason withheld Feb 27 17:04:18 crc kubenswrapper[4700]: [+]process-running ok Feb 27 17:04:18 crc kubenswrapper[4700]: healthz check failed Feb 27 17:04:18 crc kubenswrapper[4700]: I0227 17:04:18.162823 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-58xmr" podUID="cce840a6-7988-43de-a6d7-07187abb65c3" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 17:04:18 crc kubenswrapper[4700]: I0227 17:04:18.163237 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f4s6x" Feb 27 17:04:18 crc kubenswrapper[4700]: I0227 17:04:18.170617 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f4s6x"] Feb 27 17:04:18 crc kubenswrapper[4700]: I0227 17:04:18.248020 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f59b4462-8588-4be1-948c-0c1e3b84f21a-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"f59b4462-8588-4be1-948c-0c1e3b84f21a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 17:04:18 crc kubenswrapper[4700]: I0227 17:04:18.248110 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f59b4462-8588-4be1-948c-0c1e3b84f21a-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"f59b4462-8588-4be1-948c-0c1e3b84f21a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 17:04:18 crc kubenswrapper[4700]: I0227 17:04:18.351358 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f59b4462-8588-4be1-948c-0c1e3b84f21a-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"f59b4462-8588-4be1-948c-0c1e3b84f21a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 17:04:18 crc kubenswrapper[4700]: I0227 17:04:18.351503 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce-catalog-content\") pod \"redhat-operators-f4s6x\" (UID: \"8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce\") " pod="openshift-marketplace/redhat-operators-f4s6x" Feb 27 17:04:18 crc kubenswrapper[4700]: I0227 17:04:18.351530 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f59b4462-8588-4be1-948c-0c1e3b84f21a-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"f59b4462-8588-4be1-948c-0c1e3b84f21a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 17:04:18 crc kubenswrapper[4700]: I0227 17:04:18.351588 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce-utilities\") pod \"redhat-operators-f4s6x\" (UID: \"8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce\") " pod="openshift-marketplace/redhat-operators-f4s6x" Feb 27 17:04:18 crc kubenswrapper[4700]: I0227 17:04:18.351629 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpbh9\" (UniqueName: \"kubernetes.io/projected/8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce-kube-api-access-vpbh9\") pod \"redhat-operators-f4s6x\" (UID: \"8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce\") " pod="openshift-marketplace/redhat-operators-f4s6x" Feb 27 17:04:18 crc kubenswrapper[4700]: I0227 17:04:18.352051 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f59b4462-8588-4be1-948c-0c1e3b84f21a-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"f59b4462-8588-4be1-948c-0c1e3b84f21a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 17:04:18 crc kubenswrapper[4700]: I0227 17:04:18.380410 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f59b4462-8588-4be1-948c-0c1e3b84f21a-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"f59b4462-8588-4be1-948c-0c1e3b84f21a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 17:04:18 crc kubenswrapper[4700]: I0227 17:04:18.438332 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 17:04:18 crc kubenswrapper[4700]: I0227 17:04:18.453009 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce-catalog-content\") pod \"redhat-operators-f4s6x\" (UID: \"8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce\") " pod="openshift-marketplace/redhat-operators-f4s6x" Feb 27 17:04:18 crc kubenswrapper[4700]: I0227 17:04:18.453078 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce-utilities\") pod \"redhat-operators-f4s6x\" (UID: \"8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce\") " pod="openshift-marketplace/redhat-operators-f4s6x" Feb 27 17:04:18 crc kubenswrapper[4700]: I0227 17:04:18.453108 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpbh9\" (UniqueName: \"kubernetes.io/projected/8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce-kube-api-access-vpbh9\") pod \"redhat-operators-f4s6x\" (UID: \"8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce\") " pod="openshift-marketplace/redhat-operators-f4s6x" Feb 27 17:04:18 crc kubenswrapper[4700]: I0227 17:04:18.453777 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce-catalog-content\") pod \"redhat-operators-f4s6x\" (UID: \"8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce\") " pod="openshift-marketplace/redhat-operators-f4s6x" Feb 27 17:04:18 crc kubenswrapper[4700]: I0227 17:04:18.453871 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce-utilities\") pod \"redhat-operators-f4s6x\" (UID: \"8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce\") " pod="openshift-marketplace/redhat-operators-f4s6x" Feb 27 17:04:18 crc kubenswrapper[4700]: I0227 17:04:18.485863 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpbh9\" (UniqueName: \"kubernetes.io/projected/8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce-kube-api-access-vpbh9\") pod \"redhat-operators-f4s6x\" (UID: \"8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce\") " pod="openshift-marketplace/redhat-operators-f4s6x" Feb 27 17:04:18 crc kubenswrapper[4700]: I0227 17:04:18.785437 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f4s6x" Feb 27 17:04:18 crc kubenswrapper[4700]: I0227 17:04:18.882439 4700 ???:1] "http: TLS handshake error from 192.168.126.11:43094: no serving certificate available for the kubelet" Feb 27 17:04:19 crc kubenswrapper[4700]: I0227 17:04:19.162771 4700 patch_prober.go:28] interesting pod/router-default-5444994796-58xmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 17:04:19 crc kubenswrapper[4700]: [-]has-synced failed: reason withheld Feb 27 17:04:19 crc kubenswrapper[4700]: [+]process-running ok Feb 27 17:04:19 crc kubenswrapper[4700]: healthz check failed Feb 27 17:04:19 crc kubenswrapper[4700]: I0227 17:04:19.163113 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-58xmr" podUID="cce840a6-7988-43de-a6d7-07187abb65c3" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 17:04:20 crc kubenswrapper[4700]: I0227 17:04:20.160502 4700 patch_prober.go:28] interesting pod/router-default-5444994796-58xmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 17:04:20 crc kubenswrapper[4700]: [-]has-synced failed: reason withheld Feb 27 17:04:20 crc kubenswrapper[4700]: [+]process-running ok Feb 27 17:04:20 crc kubenswrapper[4700]: healthz check failed Feb 27 17:04:20 crc kubenswrapper[4700]: I0227 17:04:20.160597 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-58xmr" podUID="cce840a6-7988-43de-a6d7-07187abb65c3" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 17:04:21 crc kubenswrapper[4700]: I0227 17:04:21.160319 4700 patch_prober.go:28] interesting pod/router-default-5444994796-58xmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 17:04:21 crc kubenswrapper[4700]: [-]has-synced failed: reason withheld Feb 27 17:04:21 crc kubenswrapper[4700]: [+]process-running ok Feb 27 17:04:21 crc kubenswrapper[4700]: healthz check failed Feb 27 17:04:21 crc kubenswrapper[4700]: I0227 17:04:21.160378 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-58xmr" podUID="cce840a6-7988-43de-a6d7-07187abb65c3" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 17:04:22 crc kubenswrapper[4700]: I0227 17:04:22.161064 4700 patch_prober.go:28] interesting pod/router-default-5444994796-58xmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 17:04:22 crc kubenswrapper[4700]: [-]has-synced failed: reason withheld Feb 27 17:04:22 crc kubenswrapper[4700]: [+]process-running ok Feb 27 17:04:22 crc kubenswrapper[4700]: healthz check failed Feb 27 17:04:22 crc kubenswrapper[4700]: I0227 17:04:22.161179 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-58xmr" podUID="cce840a6-7988-43de-a6d7-07187abb65c3" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 17:04:22 crc kubenswrapper[4700]: I0227 17:04:22.733209 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-gqjqp" Feb 27 17:04:23 crc kubenswrapper[4700]: I0227 17:04:23.162284 4700 patch_prober.go:28] interesting pod/router-default-5444994796-58xmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 17:04:23 crc kubenswrapper[4700]: [-]has-synced failed: reason withheld Feb 27 17:04:23 crc kubenswrapper[4700]: [+]process-running ok Feb 27 17:04:23 crc kubenswrapper[4700]: healthz check failed Feb 27 17:04:23 crc kubenswrapper[4700]: I0227 17:04:23.162374 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-58xmr" podUID="cce840a6-7988-43de-a6d7-07187abb65c3" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 17:04:24 crc kubenswrapper[4700]: I0227 17:04:24.160834 4700 patch_prober.go:28] interesting pod/router-default-5444994796-58xmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 17:04:24 crc kubenswrapper[4700]: [-]has-synced failed: reason withheld Feb 27 17:04:24 crc kubenswrapper[4700]: [+]process-running ok Feb 27 17:04:24 crc kubenswrapper[4700]: healthz check failed Feb 27 17:04:24 crc kubenswrapper[4700]: I0227 17:04:24.160934 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-58xmr" podUID="cce840a6-7988-43de-a6d7-07187abb65c3" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 17:04:25 crc kubenswrapper[4700]: I0227 17:04:25.160746 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-58xmr" Feb 27 17:04:25 crc kubenswrapper[4700]: I0227 17:04:25.164038 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-58xmr" Feb 27 17:04:26 crc kubenswrapper[4700]: I0227 17:04:26.288096 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-v4nr9" Feb 27 17:04:27 crc kubenswrapper[4700]: I0227 17:04:27.491005 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-kq9sb" Feb 27 17:04:27 crc kubenswrapper[4700]: I0227 17:04:27.496214 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-kq9sb" Feb 27 17:04:28 crc kubenswrapper[4700]: I0227 17:04:28.326210 4700 ???:1] "http: TLS handshake error from 192.168.126.11:52584: no serving certificate available for the kubelet" Feb 27 17:04:28 crc kubenswrapper[4700]: I0227 17:04:28.766759 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 17:04:28 crc kubenswrapper[4700]: I0227 17:04:28.899857 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/49b065e4-14eb-4416-bd24-2164635669d7-kubelet-dir\") pod \"49b065e4-14eb-4416-bd24-2164635669d7\" (UID: \"49b065e4-14eb-4416-bd24-2164635669d7\") " Feb 27 17:04:28 crc kubenswrapper[4700]: I0227 17:04:28.900083 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/49b065e4-14eb-4416-bd24-2164635669d7-kube-api-access\") pod \"49b065e4-14eb-4416-bd24-2164635669d7\" (UID: \"49b065e4-14eb-4416-bd24-2164635669d7\") " Feb 27 17:04:28 crc kubenswrapper[4700]: I0227 17:04:28.900156 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/49b065e4-14eb-4416-bd24-2164635669d7-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "49b065e4-14eb-4416-bd24-2164635669d7" (UID: "49b065e4-14eb-4416-bd24-2164635669d7"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:04:28 crc kubenswrapper[4700]: I0227 17:04:28.900419 4700 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/49b065e4-14eb-4416-bd24-2164635669d7-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:28 crc kubenswrapper[4700]: I0227 17:04:28.907252 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49b065e4-14eb-4416-bd24-2164635669d7-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "49b065e4-14eb-4416-bd24-2164635669d7" (UID: "49b065e4-14eb-4416-bd24-2164635669d7"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:04:29 crc kubenswrapper[4700]: I0227 17:04:29.001141 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/49b065e4-14eb-4416-bd24-2164635669d7-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:29 crc kubenswrapper[4700]: E0227 17:04:29.652531 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 17:04:29 crc kubenswrapper[4700]: E0227 17:04:29.653353 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 17:04:29 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 17:04:29 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n962r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536864-kmfkv_openshift-infra(3166d7d3-7842-4655-b10b-24e5731d77e0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled Feb 27 17:04:29 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 17:04:29 crc kubenswrapper[4700]: E0227 17:04:29.654820 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-infra/auto-csr-approver-29536864-kmfkv" podUID="3166d7d3-7842-4655-b10b-24e5731d77e0" Feb 27 17:04:29 crc kubenswrapper[4700]: I0227 17:04:29.673206 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"49b065e4-14eb-4416-bd24-2164635669d7","Type":"ContainerDied","Data":"4e732c4a5499d47eef6bb2aabf3654b6ac091d0dd5661c94bd33b56d0aec763f"} Feb 27 17:04:29 crc kubenswrapper[4700]: I0227 17:04:29.673254 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 17:04:29 crc kubenswrapper[4700]: I0227 17:04:29.673276 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e732c4a5499d47eef6bb2aabf3654b6ac091d0dd5661c94bd33b56d0aec763f" Feb 27 17:04:29 crc kubenswrapper[4700]: E0227 17:04:29.674825 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536864-kmfkv" podUID="3166d7d3-7842-4655-b10b-24e5731d77e0" Feb 27 17:04:31 crc kubenswrapper[4700]: I0227 17:04:31.282662 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-85f5d5cff6-pprdc"] Feb 27 17:04:31 crc kubenswrapper[4700]: I0227 17:04:31.283265 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-85f5d5cff6-pprdc" podUID="033db5e9-76ee-4a35-a464-ae157bd38f89" containerName="controller-manager" containerID="cri-o://f486e51340d3ac1b5e15ce74c492a7ae3f3a4f13c2a63dc816291bd9b41eecb9" gracePeriod=30 Feb 27 17:04:31 crc kubenswrapper[4700]: I0227 17:04:31.299857 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-67594fc686-h5hwx"] Feb 27 17:04:31 crc kubenswrapper[4700]: I0227 17:04:31.300184 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-67594fc686-h5hwx" podUID="7fb68227-bec3-4159-81a3-9ac7929d74cf" containerName="route-controller-manager" containerID="cri-o://11857bf45c0c8569ef62772879d1bcb1c50f716d270283c62982234fcf908eec" gracePeriod=30 Feb 27 17:04:32 crc kubenswrapper[4700]: I0227 17:04:32.695230 4700 generic.go:334] "Generic (PLEG): container finished" podID="033db5e9-76ee-4a35-a464-ae157bd38f89" containerID="f486e51340d3ac1b5e15ce74c492a7ae3f3a4f13c2a63dc816291bd9b41eecb9" exitCode=0 Feb 27 17:04:32 crc kubenswrapper[4700]: I0227 17:04:32.695372 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-85f5d5cff6-pprdc" event={"ID":"033db5e9-76ee-4a35-a464-ae157bd38f89","Type":"ContainerDied","Data":"f486e51340d3ac1b5e15ce74c492a7ae3f3a4f13c2a63dc816291bd9b41eecb9"} Feb 27 17:04:32 crc kubenswrapper[4700]: I0227 17:04:32.697328 4700 generic.go:334] "Generic (PLEG): container finished" podID="7fb68227-bec3-4159-81a3-9ac7929d74cf" containerID="11857bf45c0c8569ef62772879d1bcb1c50f716d270283c62982234fcf908eec" exitCode=0 Feb 27 17:04:32 crc kubenswrapper[4700]: I0227 17:04:32.697358 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-67594fc686-h5hwx" event={"ID":"7fb68227-bec3-4159-81a3-9ac7929d74cf","Type":"ContainerDied","Data":"11857bf45c0c8569ef62772879d1bcb1c50f716d270283c62982234fcf908eec"} Feb 27 17:04:34 crc kubenswrapper[4700]: I0227 17:04:34.017108 4700 patch_prober.go:28] interesting pod/route-controller-manager-67594fc686-h5hwx container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.45:8443/healthz\": dial tcp 10.217.0.45:8443: connect: connection refused" start-of-body= Feb 27 17:04:34 crc kubenswrapper[4700]: I0227 17:04:34.017202 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-67594fc686-h5hwx" podUID="7fb68227-bec3-4159-81a3-9ac7929d74cf" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.45:8443/healthz\": dial tcp 10.217.0.45:8443: connect: connection refused" Feb 27 17:04:34 crc kubenswrapper[4700]: I0227 17:04:34.027698 4700 patch_prober.go:28] interesting pod/controller-manager-85f5d5cff6-pprdc container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.46:8443/healthz\": dial tcp 10.217.0.46:8443: connect: connection refused" start-of-body= Feb 27 17:04:34 crc kubenswrapper[4700]: I0227 17:04:34.027778 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-85f5d5cff6-pprdc" podUID="033db5e9-76ee-4a35-a464-ae157bd38f89" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.46:8443/healthz\": dial tcp 10.217.0.46:8443: connect: connection refused" Feb 27 17:04:35 crc kubenswrapper[4700]: I0227 17:04:35.604880 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:04:36 crc kubenswrapper[4700]: I0227 17:04:36.410645 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:04:36 crc kubenswrapper[4700]: I0227 17:04:36.410708 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:04:40 crc kubenswrapper[4700]: E0227 17:04:40.104642 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 27 17:04:40 crc kubenswrapper[4700]: E0227 17:04:40.105362 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4kwkl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-5n6qh_openshift-marketplace(a72bdcb3-42e2-4a45-8c98-7b6e97afabab): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 27 17:04:40 crc kubenswrapper[4700]: E0227 17:04:40.106497 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-5n6qh" podUID="a72bdcb3-42e2-4a45-8c98-7b6e97afabab" Feb 27 17:04:41 crc kubenswrapper[4700]: E0227 17:04:41.548720 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-5n6qh" podUID="a72bdcb3-42e2-4a45-8c98-7b6e97afabab" Feb 27 17:04:41 crc kubenswrapper[4700]: E0227 17:04:41.632433 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 27 17:04:41 crc kubenswrapper[4700]: E0227 17:04:41.632629 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qsvvr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-9t75w_openshift-marketplace(89845e2b-1804-445b-8462-36b2350ae663): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 27 17:04:41 crc kubenswrapper[4700]: E0227 17:04:41.633877 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-9t75w" podUID="89845e2b-1804-445b-8462-36b2350ae663" Feb 27 17:04:41 crc kubenswrapper[4700]: E0227 17:04:41.691295 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 27 17:04:41 crc kubenswrapper[4700]: E0227 17:04:41.691644 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lqkjd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-fltgg_openshift-marketplace(bc1137d0-2659-430b-8d76-d08017e7f133): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 27 17:04:41 crc kubenswrapper[4700]: E0227 17:04:41.692807 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-fltgg" podUID="bc1137d0-2659-430b-8d76-d08017e7f133" Feb 27 17:04:41 crc kubenswrapper[4700]: E0227 17:04:41.717063 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 27 17:04:41 crc kubenswrapper[4700]: E0227 17:04:41.717198 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-d67j9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-j8wlz_openshift-marketplace(d1c12d33-6674-4c8a-b871-4d95efca7c23): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 27 17:04:41 crc kubenswrapper[4700]: E0227 17:04:41.718381 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-j8wlz" podUID="d1c12d33-6674-4c8a-b871-4d95efca7c23" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.729691 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-67594fc686-h5hwx" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.731425 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-85f5d5cff6-pprdc" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.759409 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b557968-hzjth"] Feb 27 17:04:41 crc kubenswrapper[4700]: E0227 17:04:41.759731 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="033db5e9-76ee-4a35-a464-ae157bd38f89" containerName="controller-manager" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.759748 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="033db5e9-76ee-4a35-a464-ae157bd38f89" containerName="controller-manager" Feb 27 17:04:41 crc kubenswrapper[4700]: E0227 17:04:41.759758 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fb68227-bec3-4159-81a3-9ac7929d74cf" containerName="route-controller-manager" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.759766 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fb68227-bec3-4159-81a3-9ac7929d74cf" containerName="route-controller-manager" Feb 27 17:04:41 crc kubenswrapper[4700]: E0227 17:04:41.759776 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49b065e4-14eb-4416-bd24-2164635669d7" containerName="pruner" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.759782 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="49b065e4-14eb-4416-bd24-2164635669d7" containerName="pruner" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.759898 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fb68227-bec3-4159-81a3-9ac7929d74cf" containerName="route-controller-manager" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.759917 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="49b065e4-14eb-4416-bd24-2164635669d7" containerName="pruner" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.759926 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="033db5e9-76ee-4a35-a464-ae157bd38f89" containerName="controller-manager" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.760243 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6b557968-hzjth" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.767038 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-85f5d5cff6-pprdc" event={"ID":"033db5e9-76ee-4a35-a464-ae157bd38f89","Type":"ContainerDied","Data":"be55c878c7aea505ddc8d0544faa51bad3fd624d512c58b9bd816f5fd30ff69d"} Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.767205 4700 scope.go:117] "RemoveContainer" containerID="f486e51340d3ac1b5e15ce74c492a7ae3f3a4f13c2a63dc816291bd9b41eecb9" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.767333 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-85f5d5cff6-pprdc" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.768982 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b557968-hzjth"] Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.774886 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-67594fc686-h5hwx" event={"ID":"7fb68227-bec3-4159-81a3-9ac7929d74cf","Type":"ContainerDied","Data":"bd8fd5ff966a88fca422973cae9526542d3edfe22ba52fa28683913d6b364c06"} Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.774955 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-67594fc686-h5hwx" Feb 27 17:04:41 crc kubenswrapper[4700]: E0227 17:04:41.776083 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-fltgg" podUID="bc1137d0-2659-430b-8d76-d08017e7f133" Feb 27 17:04:41 crc kubenswrapper[4700]: E0227 17:04:41.778142 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-j8wlz" podUID="d1c12d33-6674-4c8a-b871-4d95efca7c23" Feb 27 17:04:41 crc kubenswrapper[4700]: E0227 17:04:41.778167 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-9t75w" podUID="89845e2b-1804-445b-8462-36b2350ae663" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.798938 4700 scope.go:117] "RemoveContainer" containerID="11857bf45c0c8569ef62772879d1bcb1c50f716d270283c62982234fcf908eec" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.818944 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6v4rn\" (UniqueName: \"kubernetes.io/projected/7fb68227-bec3-4159-81a3-9ac7929d74cf-kube-api-access-6v4rn\") pod \"7fb68227-bec3-4159-81a3-9ac7929d74cf\" (UID: \"7fb68227-bec3-4159-81a3-9ac7929d74cf\") " Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.819007 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/033db5e9-76ee-4a35-a464-ae157bd38f89-client-ca\") pod \"033db5e9-76ee-4a35-a464-ae157bd38f89\" (UID: \"033db5e9-76ee-4a35-a464-ae157bd38f89\") " Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.819039 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gntlm\" (UniqueName: \"kubernetes.io/projected/033db5e9-76ee-4a35-a464-ae157bd38f89-kube-api-access-gntlm\") pod \"033db5e9-76ee-4a35-a464-ae157bd38f89\" (UID: \"033db5e9-76ee-4a35-a464-ae157bd38f89\") " Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.819083 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fb68227-bec3-4159-81a3-9ac7929d74cf-config\") pod \"7fb68227-bec3-4159-81a3-9ac7929d74cf\" (UID: \"7fb68227-bec3-4159-81a3-9ac7929d74cf\") " Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.819110 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/033db5e9-76ee-4a35-a464-ae157bd38f89-serving-cert\") pod \"033db5e9-76ee-4a35-a464-ae157bd38f89\" (UID: \"033db5e9-76ee-4a35-a464-ae157bd38f89\") " Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.819157 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7fb68227-bec3-4159-81a3-9ac7929d74cf-serving-cert\") pod \"7fb68227-bec3-4159-81a3-9ac7929d74cf\" (UID: \"7fb68227-bec3-4159-81a3-9ac7929d74cf\") " Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.819187 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/033db5e9-76ee-4a35-a464-ae157bd38f89-proxy-ca-bundles\") pod \"033db5e9-76ee-4a35-a464-ae157bd38f89\" (UID: \"033db5e9-76ee-4a35-a464-ae157bd38f89\") " Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.819237 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7fb68227-bec3-4159-81a3-9ac7929d74cf-client-ca\") pod \"7fb68227-bec3-4159-81a3-9ac7929d74cf\" (UID: \"7fb68227-bec3-4159-81a3-9ac7929d74cf\") " Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.819280 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/033db5e9-76ee-4a35-a464-ae157bd38f89-config\") pod \"033db5e9-76ee-4a35-a464-ae157bd38f89\" (UID: \"033db5e9-76ee-4a35-a464-ae157bd38f89\") " Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.819649 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19761039-8fd5-41cc-9897-0fd64234cc55-config\") pod \"route-controller-manager-6b557968-hzjth\" (UID: \"19761039-8fd5-41cc-9897-0fd64234cc55\") " pod="openshift-route-controller-manager/route-controller-manager-6b557968-hzjth" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.819678 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/19761039-8fd5-41cc-9897-0fd64234cc55-client-ca\") pod \"route-controller-manager-6b557968-hzjth\" (UID: \"19761039-8fd5-41cc-9897-0fd64234cc55\") " pod="openshift-route-controller-manager/route-controller-manager-6b557968-hzjth" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.819702 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2h4r\" (UniqueName: \"kubernetes.io/projected/19761039-8fd5-41cc-9897-0fd64234cc55-kube-api-access-b2h4r\") pod \"route-controller-manager-6b557968-hzjth\" (UID: \"19761039-8fd5-41cc-9897-0fd64234cc55\") " pod="openshift-route-controller-manager/route-controller-manager-6b557968-hzjth" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.819795 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19761039-8fd5-41cc-9897-0fd64234cc55-serving-cert\") pod \"route-controller-manager-6b557968-hzjth\" (UID: \"19761039-8fd5-41cc-9897-0fd64234cc55\") " pod="openshift-route-controller-manager/route-controller-manager-6b557968-hzjth" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.820561 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/033db5e9-76ee-4a35-a464-ae157bd38f89-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "033db5e9-76ee-4a35-a464-ae157bd38f89" (UID: "033db5e9-76ee-4a35-a464-ae157bd38f89"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.821005 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fb68227-bec3-4159-81a3-9ac7929d74cf-client-ca" (OuterVolumeSpecName: "client-ca") pod "7fb68227-bec3-4159-81a3-9ac7929d74cf" (UID: "7fb68227-bec3-4159-81a3-9ac7929d74cf"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.821483 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/033db5e9-76ee-4a35-a464-ae157bd38f89-config" (OuterVolumeSpecName: "config") pod "033db5e9-76ee-4a35-a464-ae157bd38f89" (UID: "033db5e9-76ee-4a35-a464-ae157bd38f89"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.822744 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/033db5e9-76ee-4a35-a464-ae157bd38f89-client-ca" (OuterVolumeSpecName: "client-ca") pod "033db5e9-76ee-4a35-a464-ae157bd38f89" (UID: "033db5e9-76ee-4a35-a464-ae157bd38f89"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.823870 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fb68227-bec3-4159-81a3-9ac7929d74cf-config" (OuterVolumeSpecName: "config") pod "7fb68227-bec3-4159-81a3-9ac7929d74cf" (UID: "7fb68227-bec3-4159-81a3-9ac7929d74cf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.831217 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fb68227-bec3-4159-81a3-9ac7929d74cf-kube-api-access-6v4rn" (OuterVolumeSpecName: "kube-api-access-6v4rn") pod "7fb68227-bec3-4159-81a3-9ac7929d74cf" (UID: "7fb68227-bec3-4159-81a3-9ac7929d74cf"). InnerVolumeSpecName "kube-api-access-6v4rn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.832147 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fb68227-bec3-4159-81a3-9ac7929d74cf-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7fb68227-bec3-4159-81a3-9ac7929d74cf" (UID: "7fb68227-bec3-4159-81a3-9ac7929d74cf"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.832684 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/033db5e9-76ee-4a35-a464-ae157bd38f89-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "033db5e9-76ee-4a35-a464-ae157bd38f89" (UID: "033db5e9-76ee-4a35-a464-ae157bd38f89"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.836992 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/033db5e9-76ee-4a35-a464-ae157bd38f89-kube-api-access-gntlm" (OuterVolumeSpecName: "kube-api-access-gntlm") pod "033db5e9-76ee-4a35-a464-ae157bd38f89" (UID: "033db5e9-76ee-4a35-a464-ae157bd38f89"). InnerVolumeSpecName "kube-api-access-gntlm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.921129 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19761039-8fd5-41cc-9897-0fd64234cc55-config\") pod \"route-controller-manager-6b557968-hzjth\" (UID: \"19761039-8fd5-41cc-9897-0fd64234cc55\") " pod="openshift-route-controller-manager/route-controller-manager-6b557968-hzjth" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.921171 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/19761039-8fd5-41cc-9897-0fd64234cc55-client-ca\") pod \"route-controller-manager-6b557968-hzjth\" (UID: \"19761039-8fd5-41cc-9897-0fd64234cc55\") " pod="openshift-route-controller-manager/route-controller-manager-6b557968-hzjth" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.921195 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2h4r\" (UniqueName: \"kubernetes.io/projected/19761039-8fd5-41cc-9897-0fd64234cc55-kube-api-access-b2h4r\") pod \"route-controller-manager-6b557968-hzjth\" (UID: \"19761039-8fd5-41cc-9897-0fd64234cc55\") " pod="openshift-route-controller-manager/route-controller-manager-6b557968-hzjth" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.921257 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19761039-8fd5-41cc-9897-0fd64234cc55-serving-cert\") pod \"route-controller-manager-6b557968-hzjth\" (UID: \"19761039-8fd5-41cc-9897-0fd64234cc55\") " pod="openshift-route-controller-manager/route-controller-manager-6b557968-hzjth" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.921305 4700 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7fb68227-bec3-4159-81a3-9ac7929d74cf-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.921316 4700 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/033db5e9-76ee-4a35-a464-ae157bd38f89-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.921325 4700 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7fb68227-bec3-4159-81a3-9ac7929d74cf-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.921334 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/033db5e9-76ee-4a35-a464-ae157bd38f89-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.921343 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6v4rn\" (UniqueName: \"kubernetes.io/projected/7fb68227-bec3-4159-81a3-9ac7929d74cf-kube-api-access-6v4rn\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.921351 4700 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/033db5e9-76ee-4a35-a464-ae157bd38f89-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.921359 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gntlm\" (UniqueName: \"kubernetes.io/projected/033db5e9-76ee-4a35-a464-ae157bd38f89-kube-api-access-gntlm\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.921367 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fb68227-bec3-4159-81a3-9ac7929d74cf-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.921375 4700 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/033db5e9-76ee-4a35-a464-ae157bd38f89-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.922437 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/19761039-8fd5-41cc-9897-0fd64234cc55-client-ca\") pod \"route-controller-manager-6b557968-hzjth\" (UID: \"19761039-8fd5-41cc-9897-0fd64234cc55\") " pod="openshift-route-controller-manager/route-controller-manager-6b557968-hzjth" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.922945 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19761039-8fd5-41cc-9897-0fd64234cc55-config\") pod \"route-controller-manager-6b557968-hzjth\" (UID: \"19761039-8fd5-41cc-9897-0fd64234cc55\") " pod="openshift-route-controller-manager/route-controller-manager-6b557968-hzjth" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.929048 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19761039-8fd5-41cc-9897-0fd64234cc55-serving-cert\") pod \"route-controller-manager-6b557968-hzjth\" (UID: \"19761039-8fd5-41cc-9897-0fd64234cc55\") " pod="openshift-route-controller-manager/route-controller-manager-6b557968-hzjth" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.942694 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2h4r\" (UniqueName: \"kubernetes.io/projected/19761039-8fd5-41cc-9897-0fd64234cc55-kube-api-access-b2h4r\") pod \"route-controller-manager-6b557968-hzjth\" (UID: \"19761039-8fd5-41cc-9897-0fd64234cc55\") " pod="openshift-route-controller-manager/route-controller-manager-6b557968-hzjth" Feb 27 17:04:41 crc kubenswrapper[4700]: I0227 17:04:41.972382 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8bf7d"] Feb 27 17:04:41 crc kubenswrapper[4700]: W0227 17:04:41.978494 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1fd07395_2d50_418c_a75a_64f1f149b216.slice/crio-f4147df4bf7530581fb8acda7003127674d124861f8bc5aa3477842db094d73e WatchSource:0}: Error finding container f4147df4bf7530581fb8acda7003127674d124861f8bc5aa3477842db094d73e: Status 404 returned error can't find the container with id f4147df4bf7530581fb8acda7003127674d124861f8bc5aa3477842db094d73e Feb 27 17:04:42 crc kubenswrapper[4700]: I0227 17:04:42.014087 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ltq9s"] Feb 27 17:04:42 crc kubenswrapper[4700]: W0227 17:04:42.025804 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfe1b00ba_9936_45d5_9a6e_f4f93ce09b44.slice/crio-17dcd3d7a115422a23c0178e740471b6249e855ce09dd68bc02a4b6763747218 WatchSource:0}: Error finding container 17dcd3d7a115422a23c0178e740471b6249e855ce09dd68bc02a4b6763747218: Status 404 returned error can't find the container with id 17dcd3d7a115422a23c0178e740471b6249e855ce09dd68bc02a4b6763747218 Feb 27 17:04:42 crc kubenswrapper[4700]: I0227 17:04:42.065259 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 27 17:04:42 crc kubenswrapper[4700]: I0227 17:04:42.069217 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f4s6x"] Feb 27 17:04:42 crc kubenswrapper[4700]: W0227 17:04:42.080602 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8fd5ee7b_1468_4a27_8bb9_e0f81d7f0cce.slice/crio-225e070473cae716ff3eec8053e05d138c0c93db36fed520e374b603853a93ab WatchSource:0}: Error finding container 225e070473cae716ff3eec8053e05d138c0c93db36fed520e374b603853a93ab: Status 404 returned error can't find the container with id 225e070473cae716ff3eec8053e05d138c0c93db36fed520e374b603853a93ab Feb 27 17:04:42 crc kubenswrapper[4700]: I0227 17:04:42.104740 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6b557968-hzjth" Feb 27 17:04:42 crc kubenswrapper[4700]: I0227 17:04:42.108877 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-85f5d5cff6-pprdc"] Feb 27 17:04:42 crc kubenswrapper[4700]: I0227 17:04:42.115396 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-85f5d5cff6-pprdc"] Feb 27 17:04:42 crc kubenswrapper[4700]: I0227 17:04:42.118758 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-67594fc686-h5hwx"] Feb 27 17:04:42 crc kubenswrapper[4700]: I0227 17:04:42.121675 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-67594fc686-h5hwx"] Feb 27 17:04:42 crc kubenswrapper[4700]: I0227 17:04:42.282967 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b557968-hzjth"] Feb 27 17:04:42 crc kubenswrapper[4700]: W0227 17:04:42.290621 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod19761039_8fd5_41cc_9897_0fd64234cc55.slice/crio-7dce627c9011d2010d3dec362753570684ac5a7b504c5596e0a86d2a324cccb5 WatchSource:0}: Error finding container 7dce627c9011d2010d3dec362753570684ac5a7b504c5596e0a86d2a324cccb5: Status 404 returned error can't find the container with id 7dce627c9011d2010d3dec362753570684ac5a7b504c5596e0a86d2a324cccb5 Feb 27 17:04:42 crc kubenswrapper[4700]: I0227 17:04:42.792995 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f59b4462-8588-4be1-948c-0c1e3b84f21a","Type":"ContainerStarted","Data":"ebdd355d14aea7aa25729a781ae1d6323f44f488445c404d51d23e6e188a0ead"} Feb 27 17:04:42 crc kubenswrapper[4700]: I0227 17:04:42.794591 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f59b4462-8588-4be1-948c-0c1e3b84f21a","Type":"ContainerStarted","Data":"c04f46c3db3fdac3100be7a83bad0ed6a7e897ef9b360441bedf297409e4261b"} Feb 27 17:04:42 crc kubenswrapper[4700]: I0227 17:04:42.797452 4700 generic.go:334] "Generic (PLEG): container finished" podID="1fd07395-2d50-418c-a75a-64f1f149b216" containerID="ffc4f32ded5ffef823909e33af29dc3cbb744b87b3f9ce208943276adfbd5614" exitCode=0 Feb 27 17:04:42 crc kubenswrapper[4700]: I0227 17:04:42.797652 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8bf7d" event={"ID":"1fd07395-2d50-418c-a75a-64f1f149b216","Type":"ContainerDied","Data":"ffc4f32ded5ffef823909e33af29dc3cbb744b87b3f9ce208943276adfbd5614"} Feb 27 17:04:42 crc kubenswrapper[4700]: I0227 17:04:42.798448 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8bf7d" event={"ID":"1fd07395-2d50-418c-a75a-64f1f149b216","Type":"ContainerStarted","Data":"f4147df4bf7530581fb8acda7003127674d124861f8bc5aa3477842db094d73e"} Feb 27 17:04:42 crc kubenswrapper[4700]: I0227 17:04:42.801055 4700 generic.go:334] "Generic (PLEG): container finished" podID="8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2" containerID="2bfba51a7a76772d0f6968119fb575eba783f431af139fbda9ff262e0010306d" exitCode=0 Feb 27 17:04:42 crc kubenswrapper[4700]: I0227 17:04:42.801238 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gpb6h" event={"ID":"8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2","Type":"ContainerDied","Data":"2bfba51a7a76772d0f6968119fb575eba783f431af139fbda9ff262e0010306d"} Feb 27 17:04:42 crc kubenswrapper[4700]: I0227 17:04:42.805306 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6b557968-hzjth" event={"ID":"19761039-8fd5-41cc-9897-0fd64234cc55","Type":"ContainerStarted","Data":"20807e027c8ae00862950c840f6166200a2e20b3c3f1e7f373434e82cac81d77"} Feb 27 17:04:42 crc kubenswrapper[4700]: I0227 17:04:42.805424 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6b557968-hzjth" event={"ID":"19761039-8fd5-41cc-9897-0fd64234cc55","Type":"ContainerStarted","Data":"7dce627c9011d2010d3dec362753570684ac5a7b504c5596e0a86d2a324cccb5"} Feb 27 17:04:42 crc kubenswrapper[4700]: I0227 17:04:42.806098 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6b557968-hzjth" Feb 27 17:04:42 crc kubenswrapper[4700]: I0227 17:04:42.811479 4700 generic.go:334] "Generic (PLEG): container finished" podID="8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce" containerID="e15d3e5ae320d4ee0be90673b5bd412d4649023e638968b2f1d880715bb3499c" exitCode=0 Feb 27 17:04:42 crc kubenswrapper[4700]: I0227 17:04:42.811664 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f4s6x" event={"ID":"8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce","Type":"ContainerDied","Data":"e15d3e5ae320d4ee0be90673b5bd412d4649023e638968b2f1d880715bb3499c"} Feb 27 17:04:42 crc kubenswrapper[4700]: I0227 17:04:42.812113 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f4s6x" event={"ID":"8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce","Type":"ContainerStarted","Data":"225e070473cae716ff3eec8053e05d138c0c93db36fed520e374b603853a93ab"} Feb 27 17:04:42 crc kubenswrapper[4700]: I0227 17:04:42.823211 4700 generic.go:334] "Generic (PLEG): container finished" podID="fe1b00ba-9936-45d5-9a6e-f4f93ce09b44" containerID="f1e5ff4267abdf9ff0078425b5b6b2499ae863eb907438b538f3e53c30cd262a" exitCode=0 Feb 27 17:04:42 crc kubenswrapper[4700]: I0227 17:04:42.823301 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ltq9s" event={"ID":"fe1b00ba-9936-45d5-9a6e-f4f93ce09b44","Type":"ContainerDied","Data":"f1e5ff4267abdf9ff0078425b5b6b2499ae863eb907438b538f3e53c30cd262a"} Feb 27 17:04:42 crc kubenswrapper[4700]: I0227 17:04:42.823327 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ltq9s" event={"ID":"fe1b00ba-9936-45d5-9a6e-f4f93ce09b44","Type":"ContainerStarted","Data":"17dcd3d7a115422a23c0178e740471b6249e855ce09dd68bc02a4b6763747218"} Feb 27 17:04:42 crc kubenswrapper[4700]: I0227 17:04:42.825043 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=24.825018582 podStartE2EDuration="24.825018582s" podCreationTimestamp="2026-02-27 17:04:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:42.813612763 +0000 UTC m=+242.798925520" watchObservedRunningTime="2026-02-27 17:04:42.825018582 +0000 UTC m=+242.810331329" Feb 27 17:04:42 crc kubenswrapper[4700]: I0227 17:04:42.837142 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6b557968-hzjth" Feb 27 17:04:42 crc kubenswrapper[4700]: I0227 17:04:42.839520 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6b557968-hzjth" podStartSLOduration=11.839446551 podStartE2EDuration="11.839446551s" podCreationTimestamp="2026-02-27 17:04:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:42.839171364 +0000 UTC m=+242.824484111" watchObservedRunningTime="2026-02-27 17:04:42.839446551 +0000 UTC m=+242.824759308" Feb 27 17:04:42 crc kubenswrapper[4700]: I0227 17:04:42.995877 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="033db5e9-76ee-4a35-a464-ae157bd38f89" path="/var/lib/kubelet/pods/033db5e9-76ee-4a35-a464-ae157bd38f89/volumes" Feb 27 17:04:42 crc kubenswrapper[4700]: I0227 17:04:42.996675 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fb68227-bec3-4159-81a3-9ac7929d74cf" path="/var/lib/kubelet/pods/7fb68227-bec3-4159-81a3-9ac7929d74cf/volumes" Feb 27 17:04:43 crc kubenswrapper[4700]: I0227 17:04:43.654831 4700 csr.go:261] certificate signing request csr-l6w9w is approved, waiting to be issued Feb 27 17:04:43 crc kubenswrapper[4700]: I0227 17:04:43.659785 4700 csr.go:257] certificate signing request csr-l6w9w is issued Feb 27 17:04:43 crc kubenswrapper[4700]: I0227 17:04:43.831418 4700 generic.go:334] "Generic (PLEG): container finished" podID="f59b4462-8588-4be1-948c-0c1e3b84f21a" containerID="ebdd355d14aea7aa25729a781ae1d6323f44f488445c404d51d23e6e188a0ead" exitCode=0 Feb 27 17:04:43 crc kubenswrapper[4700]: I0227 17:04:43.831497 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f59b4462-8588-4be1-948c-0c1e3b84f21a","Type":"ContainerDied","Data":"ebdd355d14aea7aa25729a781ae1d6323f44f488445c404d51d23e6e188a0ead"} Feb 27 17:04:43 crc kubenswrapper[4700]: I0227 17:04:43.834765 4700 generic.go:334] "Generic (PLEG): container finished" podID="3166d7d3-7842-4655-b10b-24e5731d77e0" containerID="150cee9a20fea156a81b992257b8aa0474a7374f0bf2b8aefeb83d1ea77030d4" exitCode=0 Feb 27 17:04:43 crc kubenswrapper[4700]: I0227 17:04:43.834888 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536864-kmfkv" event={"ID":"3166d7d3-7842-4655-b10b-24e5731d77e0","Type":"ContainerDied","Data":"150cee9a20fea156a81b992257b8aa0474a7374f0bf2b8aefeb83d1ea77030d4"} Feb 27 17:04:44 crc kubenswrapper[4700]: I0227 17:04:44.661661 4700 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-11-08 14:48:21.848803095 +0000 UTC Feb 27 17:04:44 crc kubenswrapper[4700]: I0227 17:04:44.661713 4700 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6093h43m37.187094716s for next certificate rotation Feb 27 17:04:44 crc kubenswrapper[4700]: I0227 17:04:44.724039 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-bfcdcc75b-rw22n"] Feb 27 17:04:44 crc kubenswrapper[4700]: I0227 17:04:44.726705 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-bfcdcc75b-rw22n" Feb 27 17:04:44 crc kubenswrapper[4700]: I0227 17:04:44.729945 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 27 17:04:44 crc kubenswrapper[4700]: I0227 17:04:44.730202 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 27 17:04:44 crc kubenswrapper[4700]: I0227 17:04:44.730386 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 27 17:04:44 crc kubenswrapper[4700]: I0227 17:04:44.733903 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 27 17:04:44 crc kubenswrapper[4700]: I0227 17:04:44.734142 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 27 17:04:44 crc kubenswrapper[4700]: I0227 17:04:44.734949 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 27 17:04:44 crc kubenswrapper[4700]: I0227 17:04:44.750130 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 27 17:04:44 crc kubenswrapper[4700]: I0227 17:04:44.754947 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-bfcdcc75b-rw22n"] Feb 27 17:04:44 crc kubenswrapper[4700]: I0227 17:04:44.768426 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3707ad5e-da74-4541-9b24-da2763ac81a0-client-ca\") pod \"controller-manager-bfcdcc75b-rw22n\" (UID: \"3707ad5e-da74-4541-9b24-da2763ac81a0\") " pod="openshift-controller-manager/controller-manager-bfcdcc75b-rw22n" Feb 27 17:04:44 crc kubenswrapper[4700]: I0227 17:04:44.768482 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksv2v\" (UniqueName: \"kubernetes.io/projected/3707ad5e-da74-4541-9b24-da2763ac81a0-kube-api-access-ksv2v\") pod \"controller-manager-bfcdcc75b-rw22n\" (UID: \"3707ad5e-da74-4541-9b24-da2763ac81a0\") " pod="openshift-controller-manager/controller-manager-bfcdcc75b-rw22n" Feb 27 17:04:44 crc kubenswrapper[4700]: I0227 17:04:44.768508 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3707ad5e-da74-4541-9b24-da2763ac81a0-serving-cert\") pod \"controller-manager-bfcdcc75b-rw22n\" (UID: \"3707ad5e-da74-4541-9b24-da2763ac81a0\") " pod="openshift-controller-manager/controller-manager-bfcdcc75b-rw22n" Feb 27 17:04:44 crc kubenswrapper[4700]: I0227 17:04:44.768525 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3707ad5e-da74-4541-9b24-da2763ac81a0-proxy-ca-bundles\") pod \"controller-manager-bfcdcc75b-rw22n\" (UID: \"3707ad5e-da74-4541-9b24-da2763ac81a0\") " pod="openshift-controller-manager/controller-manager-bfcdcc75b-rw22n" Feb 27 17:04:44 crc kubenswrapper[4700]: I0227 17:04:44.768569 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3707ad5e-da74-4541-9b24-da2763ac81a0-config\") pod \"controller-manager-bfcdcc75b-rw22n\" (UID: \"3707ad5e-da74-4541-9b24-da2763ac81a0\") " pod="openshift-controller-manager/controller-manager-bfcdcc75b-rw22n" Feb 27 17:04:44 crc kubenswrapper[4700]: I0227 17:04:44.869699 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3707ad5e-da74-4541-9b24-da2763ac81a0-client-ca\") pod \"controller-manager-bfcdcc75b-rw22n\" (UID: \"3707ad5e-da74-4541-9b24-da2763ac81a0\") " pod="openshift-controller-manager/controller-manager-bfcdcc75b-rw22n" Feb 27 17:04:44 crc kubenswrapper[4700]: I0227 17:04:44.869936 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksv2v\" (UniqueName: \"kubernetes.io/projected/3707ad5e-da74-4541-9b24-da2763ac81a0-kube-api-access-ksv2v\") pod \"controller-manager-bfcdcc75b-rw22n\" (UID: \"3707ad5e-da74-4541-9b24-da2763ac81a0\") " pod="openshift-controller-manager/controller-manager-bfcdcc75b-rw22n" Feb 27 17:04:44 crc kubenswrapper[4700]: I0227 17:04:44.869961 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3707ad5e-da74-4541-9b24-da2763ac81a0-serving-cert\") pod \"controller-manager-bfcdcc75b-rw22n\" (UID: \"3707ad5e-da74-4541-9b24-da2763ac81a0\") " pod="openshift-controller-manager/controller-manager-bfcdcc75b-rw22n" Feb 27 17:04:44 crc kubenswrapper[4700]: I0227 17:04:44.869981 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3707ad5e-da74-4541-9b24-da2763ac81a0-proxy-ca-bundles\") pod \"controller-manager-bfcdcc75b-rw22n\" (UID: \"3707ad5e-da74-4541-9b24-da2763ac81a0\") " pod="openshift-controller-manager/controller-manager-bfcdcc75b-rw22n" Feb 27 17:04:44 crc kubenswrapper[4700]: I0227 17:04:44.870059 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3707ad5e-da74-4541-9b24-da2763ac81a0-config\") pod \"controller-manager-bfcdcc75b-rw22n\" (UID: \"3707ad5e-da74-4541-9b24-da2763ac81a0\") " pod="openshift-controller-manager/controller-manager-bfcdcc75b-rw22n" Feb 27 17:04:44 crc kubenswrapper[4700]: I0227 17:04:44.870626 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3707ad5e-da74-4541-9b24-da2763ac81a0-client-ca\") pod \"controller-manager-bfcdcc75b-rw22n\" (UID: \"3707ad5e-da74-4541-9b24-da2763ac81a0\") " pod="openshift-controller-manager/controller-manager-bfcdcc75b-rw22n" Feb 27 17:04:44 crc kubenswrapper[4700]: I0227 17:04:44.871181 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3707ad5e-da74-4541-9b24-da2763ac81a0-config\") pod \"controller-manager-bfcdcc75b-rw22n\" (UID: \"3707ad5e-da74-4541-9b24-da2763ac81a0\") " pod="openshift-controller-manager/controller-manager-bfcdcc75b-rw22n" Feb 27 17:04:44 crc kubenswrapper[4700]: I0227 17:04:44.871371 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3707ad5e-da74-4541-9b24-da2763ac81a0-proxy-ca-bundles\") pod \"controller-manager-bfcdcc75b-rw22n\" (UID: \"3707ad5e-da74-4541-9b24-da2763ac81a0\") " pod="openshift-controller-manager/controller-manager-bfcdcc75b-rw22n" Feb 27 17:04:44 crc kubenswrapper[4700]: I0227 17:04:44.876682 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3707ad5e-da74-4541-9b24-da2763ac81a0-serving-cert\") pod \"controller-manager-bfcdcc75b-rw22n\" (UID: \"3707ad5e-da74-4541-9b24-da2763ac81a0\") " pod="openshift-controller-manager/controller-manager-bfcdcc75b-rw22n" Feb 27 17:04:44 crc kubenswrapper[4700]: I0227 17:04:44.886684 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksv2v\" (UniqueName: \"kubernetes.io/projected/3707ad5e-da74-4541-9b24-da2763ac81a0-kube-api-access-ksv2v\") pod \"controller-manager-bfcdcc75b-rw22n\" (UID: \"3707ad5e-da74-4541-9b24-da2763ac81a0\") " pod="openshift-controller-manager/controller-manager-bfcdcc75b-rw22n" Feb 27 17:04:45 crc kubenswrapper[4700]: I0227 17:04:45.053629 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-bfcdcc75b-rw22n" Feb 27 17:04:45 crc kubenswrapper[4700]: I0227 17:04:45.149140 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536864-kmfkv" Feb 27 17:04:45 crc kubenswrapper[4700]: I0227 17:04:45.153384 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 17:04:45 crc kubenswrapper[4700]: I0227 17:04:45.173984 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n962r\" (UniqueName: \"kubernetes.io/projected/3166d7d3-7842-4655-b10b-24e5731d77e0-kube-api-access-n962r\") pod \"3166d7d3-7842-4655-b10b-24e5731d77e0\" (UID: \"3166d7d3-7842-4655-b10b-24e5731d77e0\") " Feb 27 17:04:45 crc kubenswrapper[4700]: I0227 17:04:45.185803 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3166d7d3-7842-4655-b10b-24e5731d77e0-kube-api-access-n962r" (OuterVolumeSpecName: "kube-api-access-n962r") pod "3166d7d3-7842-4655-b10b-24e5731d77e0" (UID: "3166d7d3-7842-4655-b10b-24e5731d77e0"). InnerVolumeSpecName "kube-api-access-n962r". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:04:45 crc kubenswrapper[4700]: I0227 17:04:45.274859 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f59b4462-8588-4be1-948c-0c1e3b84f21a-kubelet-dir\") pod \"f59b4462-8588-4be1-948c-0c1e3b84f21a\" (UID: \"f59b4462-8588-4be1-948c-0c1e3b84f21a\") " Feb 27 17:04:45 crc kubenswrapper[4700]: I0227 17:04:45.274958 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f59b4462-8588-4be1-948c-0c1e3b84f21a-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "f59b4462-8588-4be1-948c-0c1e3b84f21a" (UID: "f59b4462-8588-4be1-948c-0c1e3b84f21a"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:04:45 crc kubenswrapper[4700]: I0227 17:04:45.275300 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f59b4462-8588-4be1-948c-0c1e3b84f21a-kube-api-access\") pod \"f59b4462-8588-4be1-948c-0c1e3b84f21a\" (UID: \"f59b4462-8588-4be1-948c-0c1e3b84f21a\") " Feb 27 17:04:45 crc kubenswrapper[4700]: I0227 17:04:45.275564 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n962r\" (UniqueName: \"kubernetes.io/projected/3166d7d3-7842-4655-b10b-24e5731d77e0-kube-api-access-n962r\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:45 crc kubenswrapper[4700]: I0227 17:04:45.275586 4700 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f59b4462-8588-4be1-948c-0c1e3b84f21a-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:45 crc kubenswrapper[4700]: I0227 17:04:45.277426 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f59b4462-8588-4be1-948c-0c1e3b84f21a-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "f59b4462-8588-4be1-948c-0c1e3b84f21a" (UID: "f59b4462-8588-4be1-948c-0c1e3b84f21a"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:04:45 crc kubenswrapper[4700]: I0227 17:04:45.377130 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f59b4462-8588-4be1-948c-0c1e3b84f21a-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:45 crc kubenswrapper[4700]: I0227 17:04:45.446909 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-bfcdcc75b-rw22n"] Feb 27 17:04:45 crc kubenswrapper[4700]: I0227 17:04:45.662661 4700 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2027-01-16 12:56:10.195289652 +0000 UTC Feb 27 17:04:45 crc kubenswrapper[4700]: I0227 17:04:45.662722 4700 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 7747h51m24.53256964s for next certificate rotation Feb 27 17:04:45 crc kubenswrapper[4700]: I0227 17:04:45.849297 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 17:04:45 crc kubenswrapper[4700]: I0227 17:04:45.849305 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f59b4462-8588-4be1-948c-0c1e3b84f21a","Type":"ContainerDied","Data":"c04f46c3db3fdac3100be7a83bad0ed6a7e897ef9b360441bedf297409e4261b"} Feb 27 17:04:45 crc kubenswrapper[4700]: I0227 17:04:45.849342 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c04f46c3db3fdac3100be7a83bad0ed6a7e897ef9b360441bedf297409e4261b" Feb 27 17:04:45 crc kubenswrapper[4700]: I0227 17:04:45.851038 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536864-kmfkv" event={"ID":"3166d7d3-7842-4655-b10b-24e5731d77e0","Type":"ContainerDied","Data":"3dede70b7b809e53b04393ff0100eaa5a0e255dd345fcbdad115b702a28680c3"} Feb 27 17:04:45 crc kubenswrapper[4700]: I0227 17:04:45.851064 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3dede70b7b809e53b04393ff0100eaa5a0e255dd345fcbdad115b702a28680c3" Feb 27 17:04:45 crc kubenswrapper[4700]: I0227 17:04:45.851119 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536864-kmfkv" Feb 27 17:04:46 crc kubenswrapper[4700]: I0227 17:04:46.856836 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-bfcdcc75b-rw22n" event={"ID":"3707ad5e-da74-4541-9b24-da2763ac81a0","Type":"ContainerStarted","Data":"8ad842720502545d2fa44a8860dbb96b216044d6f02a6165a14ae13b4af30f1e"} Feb 27 17:04:47 crc kubenswrapper[4700]: I0227 17:04:47.863489 4700 generic.go:334] "Generic (PLEG): container finished" podID="1fd07395-2d50-418c-a75a-64f1f149b216" containerID="13aac29224c0ea229469303fa0edd0e95d593e718293b729d5aab6c3dc7c7a1c" exitCode=0 Feb 27 17:04:47 crc kubenswrapper[4700]: I0227 17:04:47.863970 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8bf7d" event={"ID":"1fd07395-2d50-418c-a75a-64f1f149b216","Type":"ContainerDied","Data":"13aac29224c0ea229469303fa0edd0e95d593e718293b729d5aab6c3dc7c7a1c"} Feb 27 17:04:47 crc kubenswrapper[4700]: I0227 17:04:47.865930 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-bfcdcc75b-rw22n" event={"ID":"3707ad5e-da74-4541-9b24-da2763ac81a0","Type":"ContainerStarted","Data":"4bc4f220b01758f44932a743fe9f79351b73253f802ec53240bdbb4452696787"} Feb 27 17:04:47 crc kubenswrapper[4700]: I0227 17:04:47.866517 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-bfcdcc75b-rw22n" Feb 27 17:04:47 crc kubenswrapper[4700]: I0227 17:04:47.871020 4700 generic.go:334] "Generic (PLEG): container finished" podID="8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2" containerID="fa9e2fbb1c27f16e3b6698982e627ab8b7cbee499d71e03e21ab93e57fcb9d66" exitCode=0 Feb 27 17:04:47 crc kubenswrapper[4700]: I0227 17:04:47.871046 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gpb6h" event={"ID":"8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2","Type":"ContainerDied","Data":"fa9e2fbb1c27f16e3b6698982e627ab8b7cbee499d71e03e21ab93e57fcb9d66"} Feb 27 17:04:47 crc kubenswrapper[4700]: I0227 17:04:47.871629 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-bfcdcc75b-rw22n" Feb 27 17:04:47 crc kubenswrapper[4700]: I0227 17:04:47.904144 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-bfcdcc75b-rw22n" podStartSLOduration=16.904128451 podStartE2EDuration="16.904128451s" podCreationTimestamp="2026-02-27 17:04:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:47.903029472 +0000 UTC m=+247.888342219" watchObservedRunningTime="2026-02-27 17:04:47.904128451 +0000 UTC m=+247.889441198" Feb 27 17:04:47 crc kubenswrapper[4700]: I0227 17:04:47.915090 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-72b2b" Feb 27 17:04:50 crc kubenswrapper[4700]: I0227 17:04:50.512606 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 27 17:04:50 crc kubenswrapper[4700]: E0227 17:04:50.513300 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f59b4462-8588-4be1-948c-0c1e3b84f21a" containerName="pruner" Feb 27 17:04:50 crc kubenswrapper[4700]: I0227 17:04:50.513316 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="f59b4462-8588-4be1-948c-0c1e3b84f21a" containerName="pruner" Feb 27 17:04:50 crc kubenswrapper[4700]: E0227 17:04:50.513335 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3166d7d3-7842-4655-b10b-24e5731d77e0" containerName="oc" Feb 27 17:04:50 crc kubenswrapper[4700]: I0227 17:04:50.513341 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="3166d7d3-7842-4655-b10b-24e5731d77e0" containerName="oc" Feb 27 17:04:50 crc kubenswrapper[4700]: I0227 17:04:50.513516 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="3166d7d3-7842-4655-b10b-24e5731d77e0" containerName="oc" Feb 27 17:04:50 crc kubenswrapper[4700]: I0227 17:04:50.513528 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="f59b4462-8588-4be1-948c-0c1e3b84f21a" containerName="pruner" Feb 27 17:04:50 crc kubenswrapper[4700]: I0227 17:04:50.514040 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 17:04:50 crc kubenswrapper[4700]: I0227 17:04:50.516182 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 27 17:04:50 crc kubenswrapper[4700]: I0227 17:04:50.516321 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 27 17:04:50 crc kubenswrapper[4700]: I0227 17:04:50.517502 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 27 17:04:50 crc kubenswrapper[4700]: I0227 17:04:50.557543 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4d02b56f-e679-48b5-a2ce-eb3299cb67a8-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"4d02b56f-e679-48b5-a2ce-eb3299cb67a8\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 17:04:50 crc kubenswrapper[4700]: I0227 17:04:50.557769 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4d02b56f-e679-48b5-a2ce-eb3299cb67a8-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"4d02b56f-e679-48b5-a2ce-eb3299cb67a8\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 17:04:50 crc kubenswrapper[4700]: I0227 17:04:50.658958 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4d02b56f-e679-48b5-a2ce-eb3299cb67a8-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"4d02b56f-e679-48b5-a2ce-eb3299cb67a8\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 17:04:50 crc kubenswrapper[4700]: I0227 17:04:50.659037 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4d02b56f-e679-48b5-a2ce-eb3299cb67a8-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"4d02b56f-e679-48b5-a2ce-eb3299cb67a8\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 17:04:50 crc kubenswrapper[4700]: I0227 17:04:50.659380 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4d02b56f-e679-48b5-a2ce-eb3299cb67a8-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"4d02b56f-e679-48b5-a2ce-eb3299cb67a8\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 17:04:50 crc kubenswrapper[4700]: I0227 17:04:50.676479 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4d02b56f-e679-48b5-a2ce-eb3299cb67a8-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"4d02b56f-e679-48b5-a2ce-eb3299cb67a8\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 17:04:50 crc kubenswrapper[4700]: I0227 17:04:50.848546 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 17:04:51 crc kubenswrapper[4700]: I0227 17:04:51.287187 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-bfcdcc75b-rw22n"] Feb 27 17:04:51 crc kubenswrapper[4700]: I0227 17:04:51.287429 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-bfcdcc75b-rw22n" podUID="3707ad5e-da74-4541-9b24-da2763ac81a0" containerName="controller-manager" containerID="cri-o://4bc4f220b01758f44932a743fe9f79351b73253f802ec53240bdbb4452696787" gracePeriod=30 Feb 27 17:04:51 crc kubenswrapper[4700]: I0227 17:04:51.371880 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b557968-hzjth"] Feb 27 17:04:51 crc kubenswrapper[4700]: I0227 17:04:51.372712 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6b557968-hzjth" podUID="19761039-8fd5-41cc-9897-0fd64234cc55" containerName="route-controller-manager" containerID="cri-o://20807e027c8ae00862950c840f6166200a2e20b3c3f1e7f373434e82cac81d77" gracePeriod=30 Feb 27 17:04:51 crc kubenswrapper[4700]: I0227 17:04:51.894906 4700 generic.go:334] "Generic (PLEG): container finished" podID="19761039-8fd5-41cc-9897-0fd64234cc55" containerID="20807e027c8ae00862950c840f6166200a2e20b3c3f1e7f373434e82cac81d77" exitCode=0 Feb 27 17:04:51 crc kubenswrapper[4700]: I0227 17:04:51.895021 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6b557968-hzjth" event={"ID":"19761039-8fd5-41cc-9897-0fd64234cc55","Type":"ContainerDied","Data":"20807e027c8ae00862950c840f6166200a2e20b3c3f1e7f373434e82cac81d77"} Feb 27 17:04:51 crc kubenswrapper[4700]: I0227 17:04:51.897140 4700 generic.go:334] "Generic (PLEG): container finished" podID="3707ad5e-da74-4541-9b24-da2763ac81a0" containerID="4bc4f220b01758f44932a743fe9f79351b73253f802ec53240bdbb4452696787" exitCode=0 Feb 27 17:04:51 crc kubenswrapper[4700]: I0227 17:04:51.897172 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-bfcdcc75b-rw22n" event={"ID":"3707ad5e-da74-4541-9b24-da2763ac81a0","Type":"ContainerDied","Data":"4bc4f220b01758f44932a743fe9f79351b73253f802ec53240bdbb4452696787"} Feb 27 17:04:52 crc kubenswrapper[4700]: I0227 17:04:52.105715 4700 patch_prober.go:28] interesting pod/route-controller-manager-6b557968-hzjth container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.57:8443/healthz\": dial tcp 10.217.0.57:8443: connect: connection refused" start-of-body= Feb 27 17:04:52 crc kubenswrapper[4700]: I0227 17:04:52.105774 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6b557968-hzjth" podUID="19761039-8fd5-41cc-9897-0fd64234cc55" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.57:8443/healthz\": dial tcp 10.217.0.57:8443: connect: connection refused" Feb 27 17:04:52 crc kubenswrapper[4700]: I0227 17:04:52.831089 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 27 17:04:52 crc kubenswrapper[4700]: W0227 17:04:52.976329 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod4d02b56f_e679_48b5_a2ce_eb3299cb67a8.slice/crio-b06da517c59293d3bb51ee14bef94b7ae1e805499c5500630fab1db3c52d62e6 WatchSource:0}: Error finding container b06da517c59293d3bb51ee14bef94b7ae1e805499c5500630fab1db3c52d62e6: Status 404 returned error can't find the container with id b06da517c59293d3bb51ee14bef94b7ae1e805499c5500630fab1db3c52d62e6 Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.006388 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6b557968-hzjth" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.011328 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-bfcdcc75b-rw22n" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.051026 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-566f554dd9-8tfvw"] Feb 27 17:04:53 crc kubenswrapper[4700]: E0227 17:04:53.051256 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19761039-8fd5-41cc-9897-0fd64234cc55" containerName="route-controller-manager" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.051267 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="19761039-8fd5-41cc-9897-0fd64234cc55" containerName="route-controller-manager" Feb 27 17:04:53 crc kubenswrapper[4700]: E0227 17:04:53.051286 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3707ad5e-da74-4541-9b24-da2763ac81a0" containerName="controller-manager" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.051292 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="3707ad5e-da74-4541-9b24-da2763ac81a0" containerName="controller-manager" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.053693 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="3707ad5e-da74-4541-9b24-da2763ac81a0" containerName="controller-manager" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.053718 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="19761039-8fd5-41cc-9897-0fd64234cc55" containerName="route-controller-manager" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.054331 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-566f554dd9-8tfvw" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.080782 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-566f554dd9-8tfvw"] Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.094807 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2h4r\" (UniqueName: \"kubernetes.io/projected/19761039-8fd5-41cc-9897-0fd64234cc55-kube-api-access-b2h4r\") pod \"19761039-8fd5-41cc-9897-0fd64234cc55\" (UID: \"19761039-8fd5-41cc-9897-0fd64234cc55\") " Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.094934 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19761039-8fd5-41cc-9897-0fd64234cc55-serving-cert\") pod \"19761039-8fd5-41cc-9897-0fd64234cc55\" (UID: \"19761039-8fd5-41cc-9897-0fd64234cc55\") " Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.094982 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3707ad5e-da74-4541-9b24-da2763ac81a0-client-ca\") pod \"3707ad5e-da74-4541-9b24-da2763ac81a0\" (UID: \"3707ad5e-da74-4541-9b24-da2763ac81a0\") " Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.095025 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3707ad5e-da74-4541-9b24-da2763ac81a0-serving-cert\") pod \"3707ad5e-da74-4541-9b24-da2763ac81a0\" (UID: \"3707ad5e-da74-4541-9b24-da2763ac81a0\") " Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.095085 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19761039-8fd5-41cc-9897-0fd64234cc55-config\") pod \"19761039-8fd5-41cc-9897-0fd64234cc55\" (UID: \"19761039-8fd5-41cc-9897-0fd64234cc55\") " Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.095192 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3707ad5e-da74-4541-9b24-da2763ac81a0-proxy-ca-bundles\") pod \"3707ad5e-da74-4541-9b24-da2763ac81a0\" (UID: \"3707ad5e-da74-4541-9b24-da2763ac81a0\") " Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.095949 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3707ad5e-da74-4541-9b24-da2763ac81a0-config\") pod \"3707ad5e-da74-4541-9b24-da2763ac81a0\" (UID: \"3707ad5e-da74-4541-9b24-da2763ac81a0\") " Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.095981 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksv2v\" (UniqueName: \"kubernetes.io/projected/3707ad5e-da74-4541-9b24-da2763ac81a0-kube-api-access-ksv2v\") pod \"3707ad5e-da74-4541-9b24-da2763ac81a0\" (UID: \"3707ad5e-da74-4541-9b24-da2763ac81a0\") " Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.096028 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/19761039-8fd5-41cc-9897-0fd64234cc55-client-ca\") pod \"19761039-8fd5-41cc-9897-0fd64234cc55\" (UID: \"19761039-8fd5-41cc-9897-0fd64234cc55\") " Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.096205 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/45f2f427-ea74-4141-9b15-6079f565a1ba-client-ca\") pod \"route-controller-manager-566f554dd9-8tfvw\" (UID: \"45f2f427-ea74-4141-9b15-6079f565a1ba\") " pod="openshift-route-controller-manager/route-controller-manager-566f554dd9-8tfvw" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.096252 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45f2f427-ea74-4141-9b15-6079f565a1ba-serving-cert\") pod \"route-controller-manager-566f554dd9-8tfvw\" (UID: \"45f2f427-ea74-4141-9b15-6079f565a1ba\") " pod="openshift-route-controller-manager/route-controller-manager-566f554dd9-8tfvw" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.096286 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hhbw\" (UniqueName: \"kubernetes.io/projected/45f2f427-ea74-4141-9b15-6079f565a1ba-kube-api-access-9hhbw\") pod \"route-controller-manager-566f554dd9-8tfvw\" (UID: \"45f2f427-ea74-4141-9b15-6079f565a1ba\") " pod="openshift-route-controller-manager/route-controller-manager-566f554dd9-8tfvw" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.096321 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45f2f427-ea74-4141-9b15-6079f565a1ba-config\") pod \"route-controller-manager-566f554dd9-8tfvw\" (UID: \"45f2f427-ea74-4141-9b15-6079f565a1ba\") " pod="openshift-route-controller-manager/route-controller-manager-566f554dd9-8tfvw" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.097290 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19761039-8fd5-41cc-9897-0fd64234cc55-config" (OuterVolumeSpecName: "config") pod "19761039-8fd5-41cc-9897-0fd64234cc55" (UID: "19761039-8fd5-41cc-9897-0fd64234cc55"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.097535 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3707ad5e-da74-4541-9b24-da2763ac81a0-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "3707ad5e-da74-4541-9b24-da2763ac81a0" (UID: "3707ad5e-da74-4541-9b24-da2763ac81a0"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.097632 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3707ad5e-da74-4541-9b24-da2763ac81a0-client-ca" (OuterVolumeSpecName: "client-ca") pod "3707ad5e-da74-4541-9b24-da2763ac81a0" (UID: "3707ad5e-da74-4541-9b24-da2763ac81a0"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.097628 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3707ad5e-da74-4541-9b24-da2763ac81a0-config" (OuterVolumeSpecName: "config") pod "3707ad5e-da74-4541-9b24-da2763ac81a0" (UID: "3707ad5e-da74-4541-9b24-da2763ac81a0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.097813 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19761039-8fd5-41cc-9897-0fd64234cc55-client-ca" (OuterVolumeSpecName: "client-ca") pod "19761039-8fd5-41cc-9897-0fd64234cc55" (UID: "19761039-8fd5-41cc-9897-0fd64234cc55"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.102000 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19761039-8fd5-41cc-9897-0fd64234cc55-kube-api-access-b2h4r" (OuterVolumeSpecName: "kube-api-access-b2h4r") pod "19761039-8fd5-41cc-9897-0fd64234cc55" (UID: "19761039-8fd5-41cc-9897-0fd64234cc55"). InnerVolumeSpecName "kube-api-access-b2h4r". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.102548 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19761039-8fd5-41cc-9897-0fd64234cc55-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "19761039-8fd5-41cc-9897-0fd64234cc55" (UID: "19761039-8fd5-41cc-9897-0fd64234cc55"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.102559 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3707ad5e-da74-4541-9b24-da2763ac81a0-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "3707ad5e-da74-4541-9b24-da2763ac81a0" (UID: "3707ad5e-da74-4541-9b24-da2763ac81a0"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.108326 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3707ad5e-da74-4541-9b24-da2763ac81a0-kube-api-access-ksv2v" (OuterVolumeSpecName: "kube-api-access-ksv2v") pod "3707ad5e-da74-4541-9b24-da2763ac81a0" (UID: "3707ad5e-da74-4541-9b24-da2763ac81a0"). InnerVolumeSpecName "kube-api-access-ksv2v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.197779 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/45f2f427-ea74-4141-9b15-6079f565a1ba-client-ca\") pod \"route-controller-manager-566f554dd9-8tfvw\" (UID: \"45f2f427-ea74-4141-9b15-6079f565a1ba\") " pod="openshift-route-controller-manager/route-controller-manager-566f554dd9-8tfvw" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.197828 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45f2f427-ea74-4141-9b15-6079f565a1ba-serving-cert\") pod \"route-controller-manager-566f554dd9-8tfvw\" (UID: \"45f2f427-ea74-4141-9b15-6079f565a1ba\") " pod="openshift-route-controller-manager/route-controller-manager-566f554dd9-8tfvw" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.197858 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hhbw\" (UniqueName: \"kubernetes.io/projected/45f2f427-ea74-4141-9b15-6079f565a1ba-kube-api-access-9hhbw\") pod \"route-controller-manager-566f554dd9-8tfvw\" (UID: \"45f2f427-ea74-4141-9b15-6079f565a1ba\") " pod="openshift-route-controller-manager/route-controller-manager-566f554dd9-8tfvw" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.197878 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45f2f427-ea74-4141-9b15-6079f565a1ba-config\") pod \"route-controller-manager-566f554dd9-8tfvw\" (UID: \"45f2f427-ea74-4141-9b15-6079f565a1ba\") " pod="openshift-route-controller-manager/route-controller-manager-566f554dd9-8tfvw" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.197919 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3707ad5e-da74-4541-9b24-da2763ac81a0-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.197929 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ksv2v\" (UniqueName: \"kubernetes.io/projected/3707ad5e-da74-4541-9b24-da2763ac81a0-kube-api-access-ksv2v\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.197939 4700 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/19761039-8fd5-41cc-9897-0fd64234cc55-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.197948 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2h4r\" (UniqueName: \"kubernetes.io/projected/19761039-8fd5-41cc-9897-0fd64234cc55-kube-api-access-b2h4r\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.197956 4700 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19761039-8fd5-41cc-9897-0fd64234cc55-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.197964 4700 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3707ad5e-da74-4541-9b24-da2763ac81a0-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.197973 4700 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3707ad5e-da74-4541-9b24-da2763ac81a0-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.197981 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19761039-8fd5-41cc-9897-0fd64234cc55-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.197991 4700 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3707ad5e-da74-4541-9b24-da2763ac81a0-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.199007 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45f2f427-ea74-4141-9b15-6079f565a1ba-config\") pod \"route-controller-manager-566f554dd9-8tfvw\" (UID: \"45f2f427-ea74-4141-9b15-6079f565a1ba\") " pod="openshift-route-controller-manager/route-controller-manager-566f554dd9-8tfvw" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.199500 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/45f2f427-ea74-4141-9b15-6079f565a1ba-client-ca\") pod \"route-controller-manager-566f554dd9-8tfvw\" (UID: \"45f2f427-ea74-4141-9b15-6079f565a1ba\") " pod="openshift-route-controller-manager/route-controller-manager-566f554dd9-8tfvw" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.202054 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45f2f427-ea74-4141-9b15-6079f565a1ba-serving-cert\") pod \"route-controller-manager-566f554dd9-8tfvw\" (UID: \"45f2f427-ea74-4141-9b15-6079f565a1ba\") " pod="openshift-route-controller-manager/route-controller-manager-566f554dd9-8tfvw" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.214288 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hhbw\" (UniqueName: \"kubernetes.io/projected/45f2f427-ea74-4141-9b15-6079f565a1ba-kube-api-access-9hhbw\") pod \"route-controller-manager-566f554dd9-8tfvw\" (UID: \"45f2f427-ea74-4141-9b15-6079f565a1ba\") " pod="openshift-route-controller-manager/route-controller-manager-566f554dd9-8tfvw" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.381171 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-566f554dd9-8tfvw" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.872315 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-566f554dd9-8tfvw"] Feb 27 17:04:53 crc kubenswrapper[4700]: W0227 17:04:53.884645 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod45f2f427_ea74_4141_9b15_6079f565a1ba.slice/crio-61f0ae127ec3a7af7fe5fb118a62fbee4b502fdb3da3c52be1f768aafcec1af9 WatchSource:0}: Error finding container 61f0ae127ec3a7af7fe5fb118a62fbee4b502fdb3da3c52be1f768aafcec1af9: Status 404 returned error can't find the container with id 61f0ae127ec3a7af7fe5fb118a62fbee4b502fdb3da3c52be1f768aafcec1af9 Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.913475 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-566f554dd9-8tfvw" event={"ID":"45f2f427-ea74-4141-9b15-6079f565a1ba","Type":"ContainerStarted","Data":"61f0ae127ec3a7af7fe5fb118a62fbee4b502fdb3da3c52be1f768aafcec1af9"} Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.916144 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"4d02b56f-e679-48b5-a2ce-eb3299cb67a8","Type":"ContainerStarted","Data":"7cb0880c5839b50218376019d120f4c34f7d9e9cf6da55bd6374e926f68ae573"} Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.916188 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"4d02b56f-e679-48b5-a2ce-eb3299cb67a8","Type":"ContainerStarted","Data":"b06da517c59293d3bb51ee14bef94b7ae1e805499c5500630fab1db3c52d62e6"} Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.919230 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6b557968-hzjth" event={"ID":"19761039-8fd5-41cc-9897-0fd64234cc55","Type":"ContainerDied","Data":"7dce627c9011d2010d3dec362753570684ac5a7b504c5596e0a86d2a324cccb5"} Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.919276 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6b557968-hzjth" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.919307 4700 scope.go:117] "RemoveContainer" containerID="20807e027c8ae00862950c840f6166200a2e20b3c3f1e7f373434e82cac81d77" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.928258 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=3.928237284 podStartE2EDuration="3.928237284s" podCreationTimestamp="2026-02-27 17:04:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:53.927834543 +0000 UTC m=+253.913147280" watchObservedRunningTime="2026-02-27 17:04:53.928237284 +0000 UTC m=+253.913550031" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.933405 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f4s6x" event={"ID":"8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce","Type":"ContainerStarted","Data":"b44d4ea4d0907acb75324ec5ab7a15e205af7ff60f93171d32f0a22b0b6be541"} Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.943109 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ltq9s" event={"ID":"fe1b00ba-9936-45d5-9a6e-f4f93ce09b44","Type":"ContainerStarted","Data":"80839cb344fc98980a3f8b13fa2f1cd29349f2491d81ca794b0062991465053a"} Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.966491 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8bf7d" event={"ID":"1fd07395-2d50-418c-a75a-64f1f149b216","Type":"ContainerStarted","Data":"5a3c6e26e9b4752d0249aa241bd430951fca816f8c2e1b610e47a31cc3ac3681"} Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.973825 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-bfcdcc75b-rw22n" Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.974615 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-bfcdcc75b-rw22n" event={"ID":"3707ad5e-da74-4541-9b24-da2763ac81a0","Type":"ContainerDied","Data":"8ad842720502545d2fa44a8860dbb96b216044d6f02a6165a14ae13b4af30f1e"} Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.977749 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gpb6h" event={"ID":"8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2","Type":"ContainerStarted","Data":"ecda766de52f66adfbef2adbf1f8b9e59f9348d240a1d0cdf36242fb57e9d755"} Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.987587 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b557968-hzjth"] Feb 27 17:04:53 crc kubenswrapper[4700]: I0227 17:04:53.991310 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b557968-hzjth"] Feb 27 17:04:54 crc kubenswrapper[4700]: I0227 17:04:54.001285 4700 scope.go:117] "RemoveContainer" containerID="4bc4f220b01758f44932a743fe9f79351b73253f802ec53240bdbb4452696787" Feb 27 17:04:54 crc kubenswrapper[4700]: I0227 17:04:54.048231 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8bf7d" podStartSLOduration=26.82350958 podStartE2EDuration="37.048210203s" podCreationTimestamp="2026-02-27 17:04:17 +0000 UTC" firstStartedPulling="2026-02-27 17:04:42.869835169 +0000 UTC m=+242.855147926" lastFinishedPulling="2026-02-27 17:04:53.094535792 +0000 UTC m=+253.079848549" observedRunningTime="2026-02-27 17:04:54.030053006 +0000 UTC m=+254.015365753" watchObservedRunningTime="2026-02-27 17:04:54.048210203 +0000 UTC m=+254.033522950" Feb 27 17:04:54 crc kubenswrapper[4700]: I0227 17:04:54.055273 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-bfcdcc75b-rw22n"] Feb 27 17:04:54 crc kubenswrapper[4700]: I0227 17:04:54.065072 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-bfcdcc75b-rw22n"] Feb 27 17:04:54 crc kubenswrapper[4700]: I0227 17:04:54.077554 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gpb6h" podStartSLOduration=27.853160828 podStartE2EDuration="38.077533992s" podCreationTimestamp="2026-02-27 17:04:16 +0000 UTC" firstStartedPulling="2026-02-27 17:04:42.871443191 +0000 UTC m=+242.856755948" lastFinishedPulling="2026-02-27 17:04:53.095816365 +0000 UTC m=+253.081129112" observedRunningTime="2026-02-27 17:04:54.074184284 +0000 UTC m=+254.059497041" watchObservedRunningTime="2026-02-27 17:04:54.077533992 +0000 UTC m=+254.062846739" Feb 27 17:04:54 crc kubenswrapper[4700]: I0227 17:04:54.298094 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 17:04:54 crc kubenswrapper[4700]: I0227 17:04:54.987729 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19761039-8fd5-41cc-9897-0fd64234cc55" path="/var/lib/kubelet/pods/19761039-8fd5-41cc-9897-0fd64234cc55/volumes" Feb 27 17:04:54 crc kubenswrapper[4700]: I0227 17:04:54.988888 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3707ad5e-da74-4541-9b24-da2763ac81a0" path="/var/lib/kubelet/pods/3707ad5e-da74-4541-9b24-da2763ac81a0/volumes" Feb 27 17:04:54 crc kubenswrapper[4700]: I0227 17:04:54.991439 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5n6qh" event={"ID":"a72bdcb3-42e2-4a45-8c98-7b6e97afabab","Type":"ContainerStarted","Data":"b3893d88bbfde9f99743b4604b171f418e7a77f1939d38d528cb0a8c6f03f1de"} Feb 27 17:04:54 crc kubenswrapper[4700]: I0227 17:04:54.994097 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-566f554dd9-8tfvw" event={"ID":"45f2f427-ea74-4141-9b15-6079f565a1ba","Type":"ContainerStarted","Data":"ac829b1c6463c3a5ee1423a4800ea90a27fe65b08e64813a059f3de3f4a74b6a"} Feb 27 17:04:54 crc kubenswrapper[4700]: I0227 17:04:54.996681 4700 generic.go:334] "Generic (PLEG): container finished" podID="4d02b56f-e679-48b5-a2ce-eb3299cb67a8" containerID="7cb0880c5839b50218376019d120f4c34f7d9e9cf6da55bd6374e926f68ae573" exitCode=0 Feb 27 17:04:54 crc kubenswrapper[4700]: I0227 17:04:54.996743 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"4d02b56f-e679-48b5-a2ce-eb3299cb67a8","Type":"ContainerDied","Data":"7cb0880c5839b50218376019d120f4c34f7d9e9cf6da55bd6374e926f68ae573"} Feb 27 17:04:54 crc kubenswrapper[4700]: I0227 17:04:54.998734 4700 generic.go:334] "Generic (PLEG): container finished" podID="bc1137d0-2659-430b-8d76-d08017e7f133" containerID="2990c00b77ba4917913964a3cc3214736412a8da0a477191ac3f1952be9f98cc" exitCode=0 Feb 27 17:04:54 crc kubenswrapper[4700]: I0227 17:04:54.998774 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fltgg" event={"ID":"bc1137d0-2659-430b-8d76-d08017e7f133","Type":"ContainerDied","Data":"2990c00b77ba4917913964a3cc3214736412a8da0a477191ac3f1952be9f98cc"} Feb 27 17:04:55 crc kubenswrapper[4700]: I0227 17:04:55.008786 4700 generic.go:334] "Generic (PLEG): container finished" podID="8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce" containerID="b44d4ea4d0907acb75324ec5ab7a15e205af7ff60f93171d32f0a22b0b6be541" exitCode=0 Feb 27 17:04:55 crc kubenswrapper[4700]: I0227 17:04:55.008881 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f4s6x" event={"ID":"8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce","Type":"ContainerDied","Data":"b44d4ea4d0907acb75324ec5ab7a15e205af7ff60f93171d32f0a22b0b6be541"} Feb 27 17:04:55 crc kubenswrapper[4700]: I0227 17:04:55.011868 4700 generic.go:334] "Generic (PLEG): container finished" podID="fe1b00ba-9936-45d5-9a6e-f4f93ce09b44" containerID="80839cb344fc98980a3f8b13fa2f1cd29349f2491d81ca794b0062991465053a" exitCode=0 Feb 27 17:04:55 crc kubenswrapper[4700]: I0227 17:04:55.011979 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ltq9s" event={"ID":"fe1b00ba-9936-45d5-9a6e-f4f93ce09b44","Type":"ContainerDied","Data":"80839cb344fc98980a3f8b13fa2f1cd29349f2491d81ca794b0062991465053a"} Feb 27 17:04:55 crc kubenswrapper[4700]: I0227 17:04:55.725851 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7695c4f755-xt7w8"] Feb 27 17:04:55 crc kubenswrapper[4700]: I0227 17:04:55.726535 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7695c4f755-xt7w8" Feb 27 17:04:55 crc kubenswrapper[4700]: I0227 17:04:55.727836 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 27 17:04:55 crc kubenswrapper[4700]: I0227 17:04:55.729090 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 27 17:04:55 crc kubenswrapper[4700]: I0227 17:04:55.729110 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 27 17:04:55 crc kubenswrapper[4700]: I0227 17:04:55.729245 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 27 17:04:55 crc kubenswrapper[4700]: I0227 17:04:55.729574 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 27 17:04:55 crc kubenswrapper[4700]: I0227 17:04:55.741306 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 27 17:04:55 crc kubenswrapper[4700]: I0227 17:04:55.741483 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 27 17:04:55 crc kubenswrapper[4700]: I0227 17:04:55.744811 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7695c4f755-xt7w8"] Feb 27 17:04:55 crc kubenswrapper[4700]: I0227 17:04:55.828943 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b63de308-633e-4e7a-8428-9d5a24a6458d-serving-cert\") pod \"controller-manager-7695c4f755-xt7w8\" (UID: \"b63de308-633e-4e7a-8428-9d5a24a6458d\") " pod="openshift-controller-manager/controller-manager-7695c4f755-xt7w8" Feb 27 17:04:55 crc kubenswrapper[4700]: I0227 17:04:55.829008 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b63de308-633e-4e7a-8428-9d5a24a6458d-config\") pod \"controller-manager-7695c4f755-xt7w8\" (UID: \"b63de308-633e-4e7a-8428-9d5a24a6458d\") " pod="openshift-controller-manager/controller-manager-7695c4f755-xt7w8" Feb 27 17:04:55 crc kubenswrapper[4700]: I0227 17:04:55.829032 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvc5q\" (UniqueName: \"kubernetes.io/projected/b63de308-633e-4e7a-8428-9d5a24a6458d-kube-api-access-qvc5q\") pod \"controller-manager-7695c4f755-xt7w8\" (UID: \"b63de308-633e-4e7a-8428-9d5a24a6458d\") " pod="openshift-controller-manager/controller-manager-7695c4f755-xt7w8" Feb 27 17:04:55 crc kubenswrapper[4700]: I0227 17:04:55.829224 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b63de308-633e-4e7a-8428-9d5a24a6458d-proxy-ca-bundles\") pod \"controller-manager-7695c4f755-xt7w8\" (UID: \"b63de308-633e-4e7a-8428-9d5a24a6458d\") " pod="openshift-controller-manager/controller-manager-7695c4f755-xt7w8" Feb 27 17:04:55 crc kubenswrapper[4700]: I0227 17:04:55.829312 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b63de308-633e-4e7a-8428-9d5a24a6458d-client-ca\") pod \"controller-manager-7695c4f755-xt7w8\" (UID: \"b63de308-633e-4e7a-8428-9d5a24a6458d\") " pod="openshift-controller-manager/controller-manager-7695c4f755-xt7w8" Feb 27 17:04:55 crc kubenswrapper[4700]: I0227 17:04:55.930477 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b63de308-633e-4e7a-8428-9d5a24a6458d-proxy-ca-bundles\") pod \"controller-manager-7695c4f755-xt7w8\" (UID: \"b63de308-633e-4e7a-8428-9d5a24a6458d\") " pod="openshift-controller-manager/controller-manager-7695c4f755-xt7w8" Feb 27 17:04:55 crc kubenswrapper[4700]: I0227 17:04:55.930549 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b63de308-633e-4e7a-8428-9d5a24a6458d-client-ca\") pod \"controller-manager-7695c4f755-xt7w8\" (UID: \"b63de308-633e-4e7a-8428-9d5a24a6458d\") " pod="openshift-controller-manager/controller-manager-7695c4f755-xt7w8" Feb 27 17:04:55 crc kubenswrapper[4700]: I0227 17:04:55.930609 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b63de308-633e-4e7a-8428-9d5a24a6458d-serving-cert\") pod \"controller-manager-7695c4f755-xt7w8\" (UID: \"b63de308-633e-4e7a-8428-9d5a24a6458d\") " pod="openshift-controller-manager/controller-manager-7695c4f755-xt7w8" Feb 27 17:04:55 crc kubenswrapper[4700]: I0227 17:04:55.930655 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b63de308-633e-4e7a-8428-9d5a24a6458d-config\") pod \"controller-manager-7695c4f755-xt7w8\" (UID: \"b63de308-633e-4e7a-8428-9d5a24a6458d\") " pod="openshift-controller-manager/controller-manager-7695c4f755-xt7w8" Feb 27 17:04:55 crc kubenswrapper[4700]: I0227 17:04:55.930678 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvc5q\" (UniqueName: \"kubernetes.io/projected/b63de308-633e-4e7a-8428-9d5a24a6458d-kube-api-access-qvc5q\") pod \"controller-manager-7695c4f755-xt7w8\" (UID: \"b63de308-633e-4e7a-8428-9d5a24a6458d\") " pod="openshift-controller-manager/controller-manager-7695c4f755-xt7w8" Feb 27 17:04:55 crc kubenswrapper[4700]: I0227 17:04:55.931587 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b63de308-633e-4e7a-8428-9d5a24a6458d-client-ca\") pod \"controller-manager-7695c4f755-xt7w8\" (UID: \"b63de308-633e-4e7a-8428-9d5a24a6458d\") " pod="openshift-controller-manager/controller-manager-7695c4f755-xt7w8" Feb 27 17:04:55 crc kubenswrapper[4700]: I0227 17:04:55.932189 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b63de308-633e-4e7a-8428-9d5a24a6458d-config\") pod \"controller-manager-7695c4f755-xt7w8\" (UID: \"b63de308-633e-4e7a-8428-9d5a24a6458d\") " pod="openshift-controller-manager/controller-manager-7695c4f755-xt7w8" Feb 27 17:04:55 crc kubenswrapper[4700]: I0227 17:04:55.932197 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b63de308-633e-4e7a-8428-9d5a24a6458d-proxy-ca-bundles\") pod \"controller-manager-7695c4f755-xt7w8\" (UID: \"b63de308-633e-4e7a-8428-9d5a24a6458d\") " pod="openshift-controller-manager/controller-manager-7695c4f755-xt7w8" Feb 27 17:04:55 crc kubenswrapper[4700]: I0227 17:04:55.937070 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b63de308-633e-4e7a-8428-9d5a24a6458d-serving-cert\") pod \"controller-manager-7695c4f755-xt7w8\" (UID: \"b63de308-633e-4e7a-8428-9d5a24a6458d\") " pod="openshift-controller-manager/controller-manager-7695c4f755-xt7w8" Feb 27 17:04:55 crc kubenswrapper[4700]: I0227 17:04:55.952478 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvc5q\" (UniqueName: \"kubernetes.io/projected/b63de308-633e-4e7a-8428-9d5a24a6458d-kube-api-access-qvc5q\") pod \"controller-manager-7695c4f755-xt7w8\" (UID: \"b63de308-633e-4e7a-8428-9d5a24a6458d\") " pod="openshift-controller-manager/controller-manager-7695c4f755-xt7w8" Feb 27 17:04:56 crc kubenswrapper[4700]: I0227 17:04:56.023865 4700 generic.go:334] "Generic (PLEG): container finished" podID="a72bdcb3-42e2-4a45-8c98-7b6e97afabab" containerID="b3893d88bbfde9f99743b4604b171f418e7a77f1939d38d528cb0a8c6f03f1de" exitCode=0 Feb 27 17:04:56 crc kubenswrapper[4700]: I0227 17:04:56.023912 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5n6qh" event={"ID":"a72bdcb3-42e2-4a45-8c98-7b6e97afabab","Type":"ContainerDied","Data":"b3893d88bbfde9f99743b4604b171f418e7a77f1939d38d528cb0a8c6f03f1de"} Feb 27 17:04:56 crc kubenswrapper[4700]: I0227 17:04:56.024937 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-566f554dd9-8tfvw" Feb 27 17:04:56 crc kubenswrapper[4700]: I0227 17:04:56.036772 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-566f554dd9-8tfvw" Feb 27 17:04:56 crc kubenswrapper[4700]: I0227 17:04:56.049199 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7695c4f755-xt7w8" Feb 27 17:04:56 crc kubenswrapper[4700]: I0227 17:04:56.050218 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-566f554dd9-8tfvw" podStartSLOduration=5.050197907 podStartE2EDuration="5.050197907s" podCreationTimestamp="2026-02-27 17:04:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:04:56.048859002 +0000 UTC m=+256.034171759" watchObservedRunningTime="2026-02-27 17:04:56.050197907 +0000 UTC m=+256.035510664" Feb 27 17:04:56 crc kubenswrapper[4700]: I0227 17:04:56.359544 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 17:04:56 crc kubenswrapper[4700]: I0227 17:04:56.452270 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4d02b56f-e679-48b5-a2ce-eb3299cb67a8-kubelet-dir\") pod \"4d02b56f-e679-48b5-a2ce-eb3299cb67a8\" (UID: \"4d02b56f-e679-48b5-a2ce-eb3299cb67a8\") " Feb 27 17:04:56 crc kubenswrapper[4700]: I0227 17:04:56.452606 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4d02b56f-e679-48b5-a2ce-eb3299cb67a8-kube-api-access\") pod \"4d02b56f-e679-48b5-a2ce-eb3299cb67a8\" (UID: \"4d02b56f-e679-48b5-a2ce-eb3299cb67a8\") " Feb 27 17:04:56 crc kubenswrapper[4700]: I0227 17:04:56.452620 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d02b56f-e679-48b5-a2ce-eb3299cb67a8-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "4d02b56f-e679-48b5-a2ce-eb3299cb67a8" (UID: "4d02b56f-e679-48b5-a2ce-eb3299cb67a8"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:04:56 crc kubenswrapper[4700]: I0227 17:04:56.452979 4700 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4d02b56f-e679-48b5-a2ce-eb3299cb67a8-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:56 crc kubenswrapper[4700]: I0227 17:04:56.455795 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d02b56f-e679-48b5-a2ce-eb3299cb67a8-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "4d02b56f-e679-48b5-a2ce-eb3299cb67a8" (UID: "4d02b56f-e679-48b5-a2ce-eb3299cb67a8"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:04:56 crc kubenswrapper[4700]: I0227 17:04:56.509365 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 27 17:04:56 crc kubenswrapper[4700]: E0227 17:04:56.509653 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d02b56f-e679-48b5-a2ce-eb3299cb67a8" containerName="pruner" Feb 27 17:04:56 crc kubenswrapper[4700]: I0227 17:04:56.509678 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d02b56f-e679-48b5-a2ce-eb3299cb67a8" containerName="pruner" Feb 27 17:04:56 crc kubenswrapper[4700]: I0227 17:04:56.509829 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d02b56f-e679-48b5-a2ce-eb3299cb67a8" containerName="pruner" Feb 27 17:04:56 crc kubenswrapper[4700]: I0227 17:04:56.510401 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 27 17:04:56 crc kubenswrapper[4700]: I0227 17:04:56.519988 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 27 17:04:56 crc kubenswrapper[4700]: I0227 17:04:56.554244 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/eb3329cc-cc82-4f3c-87c0-b2baf19069eb-kubelet-dir\") pod \"installer-9-crc\" (UID: \"eb3329cc-cc82-4f3c-87c0-b2baf19069eb\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 27 17:04:56 crc kubenswrapper[4700]: I0227 17:04:56.554410 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/eb3329cc-cc82-4f3c-87c0-b2baf19069eb-var-lock\") pod \"installer-9-crc\" (UID: \"eb3329cc-cc82-4f3c-87c0-b2baf19069eb\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 27 17:04:56 crc kubenswrapper[4700]: I0227 17:04:56.554619 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eb3329cc-cc82-4f3c-87c0-b2baf19069eb-kube-api-access\") pod \"installer-9-crc\" (UID: \"eb3329cc-cc82-4f3c-87c0-b2baf19069eb\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 27 17:04:56 crc kubenswrapper[4700]: I0227 17:04:56.554707 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4d02b56f-e679-48b5-a2ce-eb3299cb67a8-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 27 17:04:56 crc kubenswrapper[4700]: I0227 17:04:56.565858 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-zxt8n"] Feb 27 17:04:56 crc kubenswrapper[4700]: I0227 17:04:56.655969 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eb3329cc-cc82-4f3c-87c0-b2baf19069eb-kube-api-access\") pod \"installer-9-crc\" (UID: \"eb3329cc-cc82-4f3c-87c0-b2baf19069eb\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 27 17:04:56 crc kubenswrapper[4700]: I0227 17:04:56.656025 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/eb3329cc-cc82-4f3c-87c0-b2baf19069eb-kubelet-dir\") pod \"installer-9-crc\" (UID: \"eb3329cc-cc82-4f3c-87c0-b2baf19069eb\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 27 17:04:56 crc kubenswrapper[4700]: I0227 17:04:56.656055 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/eb3329cc-cc82-4f3c-87c0-b2baf19069eb-var-lock\") pod \"installer-9-crc\" (UID: \"eb3329cc-cc82-4f3c-87c0-b2baf19069eb\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 27 17:04:56 crc kubenswrapper[4700]: I0227 17:04:56.656141 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/eb3329cc-cc82-4f3c-87c0-b2baf19069eb-var-lock\") pod \"installer-9-crc\" (UID: \"eb3329cc-cc82-4f3c-87c0-b2baf19069eb\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 27 17:04:56 crc kubenswrapper[4700]: I0227 17:04:56.656146 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/eb3329cc-cc82-4f3c-87c0-b2baf19069eb-kubelet-dir\") pod \"installer-9-crc\" (UID: \"eb3329cc-cc82-4f3c-87c0-b2baf19069eb\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 27 17:04:56 crc kubenswrapper[4700]: I0227 17:04:56.670769 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eb3329cc-cc82-4f3c-87c0-b2baf19069eb-kube-api-access\") pod \"installer-9-crc\" (UID: \"eb3329cc-cc82-4f3c-87c0-b2baf19069eb\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 27 17:04:56 crc kubenswrapper[4700]: I0227 17:04:56.826512 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 27 17:04:57 crc kubenswrapper[4700]: I0227 17:04:57.029371 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 17:04:57 crc kubenswrapper[4700]: I0227 17:04:57.029359 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"4d02b56f-e679-48b5-a2ce-eb3299cb67a8","Type":"ContainerDied","Data":"b06da517c59293d3bb51ee14bef94b7ae1e805499c5500630fab1db3c52d62e6"} Feb 27 17:04:57 crc kubenswrapper[4700]: I0227 17:04:57.029424 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b06da517c59293d3bb51ee14bef94b7ae1e805499c5500630fab1db3c52d62e6" Feb 27 17:04:57 crc kubenswrapper[4700]: I0227 17:04:57.097639 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gpb6h" Feb 27 17:04:57 crc kubenswrapper[4700]: I0227 17:04:57.097689 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gpb6h" Feb 27 17:04:57 crc kubenswrapper[4700]: I0227 17:04:57.510805 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8bf7d" Feb 27 17:04:57 crc kubenswrapper[4700]: I0227 17:04:57.511171 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8bf7d" Feb 27 17:04:57 crc kubenswrapper[4700]: I0227 17:04:57.864273 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8bf7d" Feb 27 17:04:57 crc kubenswrapper[4700]: I0227 17:04:57.866284 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gpb6h" Feb 27 17:04:57 crc kubenswrapper[4700]: I0227 17:04:57.897073 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7695c4f755-xt7w8"] Feb 27 17:04:57 crc kubenswrapper[4700]: W0227 17:04:57.915047 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb63de308_633e_4e7a_8428_9d5a24a6458d.slice/crio-13c3f854b4af9b1b969b7ceffd7a4237785a18fd776ff3b2c9bc51e184f3b7d4 WatchSource:0}: Error finding container 13c3f854b4af9b1b969b7ceffd7a4237785a18fd776ff3b2c9bc51e184f3b7d4: Status 404 returned error can't find the container with id 13c3f854b4af9b1b969b7ceffd7a4237785a18fd776ff3b2c9bc51e184f3b7d4 Feb 27 17:04:58 crc kubenswrapper[4700]: I0227 17:04:58.037721 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7695c4f755-xt7w8" event={"ID":"b63de308-633e-4e7a-8428-9d5a24a6458d","Type":"ContainerStarted","Data":"13c3f854b4af9b1b969b7ceffd7a4237785a18fd776ff3b2c9bc51e184f3b7d4"} Feb 27 17:04:58 crc kubenswrapper[4700]: I0227 17:04:58.080885 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8bf7d" Feb 27 17:04:58 crc kubenswrapper[4700]: I0227 17:04:58.086409 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gpb6h" Feb 27 17:04:59 crc kubenswrapper[4700]: I0227 17:04:59.607603 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 27 17:04:59 crc kubenswrapper[4700]: W0227 17:04:59.614108 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podeb3329cc_cc82_4f3c_87c0_b2baf19069eb.slice/crio-5656a62fdcaaeb77a82f27f4b3b1314e3d0390dc9677c0c171d54edcbbf4444f WatchSource:0}: Error finding container 5656a62fdcaaeb77a82f27f4b3b1314e3d0390dc9677c0c171d54edcbbf4444f: Status 404 returned error can't find the container with id 5656a62fdcaaeb77a82f27f4b3b1314e3d0390dc9677c0c171d54edcbbf4444f Feb 27 17:04:59 crc kubenswrapper[4700]: I0227 17:04:59.820281 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8bf7d"] Feb 27 17:05:00 crc kubenswrapper[4700]: I0227 17:05:00.048654 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"eb3329cc-cc82-4f3c-87c0-b2baf19069eb","Type":"ContainerStarted","Data":"5656a62fdcaaeb77a82f27f4b3b1314e3d0390dc9677c0c171d54edcbbf4444f"} Feb 27 17:05:00 crc kubenswrapper[4700]: I0227 17:05:00.048723 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8bf7d" podUID="1fd07395-2d50-418c-a75a-64f1f149b216" containerName="registry-server" containerID="cri-o://5a3c6e26e9b4752d0249aa241bd430951fca816f8c2e1b610e47a31cc3ac3681" gracePeriod=2 Feb 27 17:05:01 crc kubenswrapper[4700]: I0227 17:05:01.078045 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f4s6x" event={"ID":"8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce","Type":"ContainerStarted","Data":"92940ff183c7ddcf11b4208e82e0c12bdc7f2bc25fec01613de65a1d13892a3a"} Feb 27 17:05:01 crc kubenswrapper[4700]: I0227 17:05:01.080124 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7695c4f755-xt7w8" event={"ID":"b63de308-633e-4e7a-8428-9d5a24a6458d","Type":"ContainerStarted","Data":"bba3b8f51ee422cf402a58ab7fa9c4508221abb9df63d94fd77979f794917fdb"} Feb 27 17:05:01 crc kubenswrapper[4700]: I0227 17:05:01.080366 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7695c4f755-xt7w8" Feb 27 17:05:01 crc kubenswrapper[4700]: I0227 17:05:01.083049 4700 generic.go:334] "Generic (PLEG): container finished" podID="1fd07395-2d50-418c-a75a-64f1f149b216" containerID="5a3c6e26e9b4752d0249aa241bd430951fca816f8c2e1b610e47a31cc3ac3681" exitCode=0 Feb 27 17:05:01 crc kubenswrapper[4700]: I0227 17:05:01.083087 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8bf7d" event={"ID":"1fd07395-2d50-418c-a75a-64f1f149b216","Type":"ContainerDied","Data":"5a3c6e26e9b4752d0249aa241bd430951fca816f8c2e1b610e47a31cc3ac3681"} Feb 27 17:05:01 crc kubenswrapper[4700]: I0227 17:05:01.088584 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7695c4f755-xt7w8" Feb 27 17:05:01 crc kubenswrapper[4700]: I0227 17:05:01.145712 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-f4s6x" podStartSLOduration=27.215205374 podStartE2EDuration="43.145690344s" podCreationTimestamp="2026-02-27 17:04:18 +0000 UTC" firstStartedPulling="2026-02-27 17:04:42.872481978 +0000 UTC m=+242.857794735" lastFinishedPulling="2026-02-27 17:04:58.802966948 +0000 UTC m=+258.788279705" observedRunningTime="2026-02-27 17:05:01.120684848 +0000 UTC m=+261.105997595" watchObservedRunningTime="2026-02-27 17:05:01.145690344 +0000 UTC m=+261.131003091" Feb 27 17:05:01 crc kubenswrapper[4700]: I0227 17:05:01.150062 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7695c4f755-xt7w8" podStartSLOduration=10.15005015 podStartE2EDuration="10.15005015s" podCreationTimestamp="2026-02-27 17:04:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:05:01.136582961 +0000 UTC m=+261.121895708" watchObservedRunningTime="2026-02-27 17:05:01.15005015 +0000 UTC m=+261.135362917" Feb 27 17:05:01 crc kubenswrapper[4700]: I0227 17:05:01.371720 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8bf7d" Feb 27 17:05:01 crc kubenswrapper[4700]: I0227 17:05:01.432260 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fd07395-2d50-418c-a75a-64f1f149b216-catalog-content\") pod \"1fd07395-2d50-418c-a75a-64f1f149b216\" (UID: \"1fd07395-2d50-418c-a75a-64f1f149b216\") " Feb 27 17:05:01 crc kubenswrapper[4700]: I0227 17:05:01.432304 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fd07395-2d50-418c-a75a-64f1f149b216-utilities\") pod \"1fd07395-2d50-418c-a75a-64f1f149b216\" (UID: \"1fd07395-2d50-418c-a75a-64f1f149b216\") " Feb 27 17:05:01 crc kubenswrapper[4700]: I0227 17:05:01.432417 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwgkj\" (UniqueName: \"kubernetes.io/projected/1fd07395-2d50-418c-a75a-64f1f149b216-kube-api-access-rwgkj\") pod \"1fd07395-2d50-418c-a75a-64f1f149b216\" (UID: \"1fd07395-2d50-418c-a75a-64f1f149b216\") " Feb 27 17:05:01 crc kubenswrapper[4700]: I0227 17:05:01.433229 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1fd07395-2d50-418c-a75a-64f1f149b216-utilities" (OuterVolumeSpecName: "utilities") pod "1fd07395-2d50-418c-a75a-64f1f149b216" (UID: "1fd07395-2d50-418c-a75a-64f1f149b216"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:05:01 crc kubenswrapper[4700]: I0227 17:05:01.444602 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fd07395-2d50-418c-a75a-64f1f149b216-kube-api-access-rwgkj" (OuterVolumeSpecName: "kube-api-access-rwgkj") pod "1fd07395-2d50-418c-a75a-64f1f149b216" (UID: "1fd07395-2d50-418c-a75a-64f1f149b216"). InnerVolumeSpecName "kube-api-access-rwgkj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:05:01 crc kubenswrapper[4700]: I0227 17:05:01.462365 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1fd07395-2d50-418c-a75a-64f1f149b216-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1fd07395-2d50-418c-a75a-64f1f149b216" (UID: "1fd07395-2d50-418c-a75a-64f1f149b216"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:05:01 crc kubenswrapper[4700]: I0227 17:05:01.533412 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwgkj\" (UniqueName: \"kubernetes.io/projected/1fd07395-2d50-418c-a75a-64f1f149b216-kube-api-access-rwgkj\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:01 crc kubenswrapper[4700]: I0227 17:05:01.533471 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fd07395-2d50-418c-a75a-64f1f149b216-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:01 crc kubenswrapper[4700]: I0227 17:05:01.533486 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fd07395-2d50-418c-a75a-64f1f149b216-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:02 crc kubenswrapper[4700]: I0227 17:05:02.092362 4700 generic.go:334] "Generic (PLEG): container finished" podID="d1c12d33-6674-4c8a-b871-4d95efca7c23" containerID="69caf0b406ec72e1bfe6e57a12df73436af7f4bfb3cf613a8d40f70fd813393e" exitCode=0 Feb 27 17:05:02 crc kubenswrapper[4700]: I0227 17:05:02.092446 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j8wlz" event={"ID":"d1c12d33-6674-4c8a-b871-4d95efca7c23","Type":"ContainerDied","Data":"69caf0b406ec72e1bfe6e57a12df73436af7f4bfb3cf613a8d40f70fd813393e"} Feb 27 17:05:02 crc kubenswrapper[4700]: I0227 17:05:02.095786 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5n6qh" event={"ID":"a72bdcb3-42e2-4a45-8c98-7b6e97afabab","Type":"ContainerStarted","Data":"3ce0b02ad05b57bf02c6d6f8289417e613941767abac9195e2d52912363ab7e3"} Feb 27 17:05:02 crc kubenswrapper[4700]: I0227 17:05:02.098801 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"eb3329cc-cc82-4f3c-87c0-b2baf19069eb","Type":"ContainerStarted","Data":"1383f6808d490ad6f1816be8c0d30fbe36237e84bd01f85dc894dc2ad1e12db7"} Feb 27 17:05:02 crc kubenswrapper[4700]: I0227 17:05:02.102795 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fltgg" event={"ID":"bc1137d0-2659-430b-8d76-d08017e7f133","Type":"ContainerStarted","Data":"1b3935e941a6d1b5664d559bb603e5aeb21fa780c268b235a7dc91820e59b219"} Feb 27 17:05:02 crc kubenswrapper[4700]: I0227 17:05:02.106083 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ltq9s" event={"ID":"fe1b00ba-9936-45d5-9a6e-f4f93ce09b44","Type":"ContainerStarted","Data":"1906e32028c1280a1bcf8c6e2eb3953f22b5cbeac74a3f7e0dbd3515469080b3"} Feb 27 17:05:02 crc kubenswrapper[4700]: I0227 17:05:02.108573 4700 generic.go:334] "Generic (PLEG): container finished" podID="89845e2b-1804-445b-8462-36b2350ae663" containerID="ec5bde0329f6c2c6c159ece770d807e1c316d222e5db57f90e3841fbf5deb8a1" exitCode=0 Feb 27 17:05:02 crc kubenswrapper[4700]: I0227 17:05:02.108669 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9t75w" event={"ID":"89845e2b-1804-445b-8462-36b2350ae663","Type":"ContainerDied","Data":"ec5bde0329f6c2c6c159ece770d807e1c316d222e5db57f90e3841fbf5deb8a1"} Feb 27 17:05:02 crc kubenswrapper[4700]: I0227 17:05:02.111808 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8bf7d" event={"ID":"1fd07395-2d50-418c-a75a-64f1f149b216","Type":"ContainerDied","Data":"f4147df4bf7530581fb8acda7003127674d124861f8bc5aa3477842db094d73e"} Feb 27 17:05:02 crc kubenswrapper[4700]: I0227 17:05:02.111846 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8bf7d" Feb 27 17:05:02 crc kubenswrapper[4700]: I0227 17:05:02.111861 4700 scope.go:117] "RemoveContainer" containerID="5a3c6e26e9b4752d0249aa241bd430951fca816f8c2e1b610e47a31cc3ac3681" Feb 27 17:05:02 crc kubenswrapper[4700]: I0227 17:05:02.136056 4700 scope.go:117] "RemoveContainer" containerID="13aac29224c0ea229469303fa0edd0e95d593e718293b729d5aab6c3dc7c7a1c" Feb 27 17:05:02 crc kubenswrapper[4700]: I0227 17:05:02.150785 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=6.150765795 podStartE2EDuration="6.150765795s" podCreationTimestamp="2026-02-27 17:04:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:05:02.150006145 +0000 UTC m=+262.135318902" watchObservedRunningTime="2026-02-27 17:05:02.150765795 +0000 UTC m=+262.136078542" Feb 27 17:05:02 crc kubenswrapper[4700]: I0227 17:05:02.157538 4700 scope.go:117] "RemoveContainer" containerID="ffc4f32ded5ffef823909e33af29dc3cbb744b87b3f9ce208943276adfbd5614" Feb 27 17:05:02 crc kubenswrapper[4700]: I0227 17:05:02.190555 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fltgg" podStartSLOduration=5.563795167 podStartE2EDuration="48.190537193s" podCreationTimestamp="2026-02-27 17:04:14 +0000 UTC" firstStartedPulling="2026-02-27 17:04:16.562062619 +0000 UTC m=+216.547375366" lastFinishedPulling="2026-02-27 17:04:59.188804645 +0000 UTC m=+259.174117392" observedRunningTime="2026-02-27 17:05:02.172525374 +0000 UTC m=+262.157838121" watchObservedRunningTime="2026-02-27 17:05:02.190537193 +0000 UTC m=+262.175849950" Feb 27 17:05:02 crc kubenswrapper[4700]: I0227 17:05:02.191164 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ltq9s" podStartSLOduration=27.086423779 podStartE2EDuration="45.191158419s" podCreationTimestamp="2026-02-27 17:04:17 +0000 UTC" firstStartedPulling="2026-02-27 17:04:42.870213739 +0000 UTC m=+242.855526496" lastFinishedPulling="2026-02-27 17:05:00.974948389 +0000 UTC m=+260.960261136" observedRunningTime="2026-02-27 17:05:02.189761742 +0000 UTC m=+262.175074489" watchObservedRunningTime="2026-02-27 17:05:02.191158419 +0000 UTC m=+262.176471176" Feb 27 17:05:02 crc kubenswrapper[4700]: I0227 17:05:02.216581 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5n6qh" podStartSLOduration=3.537454436 podStartE2EDuration="48.216558215s" podCreationTimestamp="2026-02-27 17:04:14 +0000 UTC" firstStartedPulling="2026-02-27 17:04:16.543347788 +0000 UTC m=+216.528660535" lastFinishedPulling="2026-02-27 17:05:01.222451567 +0000 UTC m=+261.207764314" observedRunningTime="2026-02-27 17:05:02.212548969 +0000 UTC m=+262.197861706" watchObservedRunningTime="2026-02-27 17:05:02.216558215 +0000 UTC m=+262.201870952" Feb 27 17:05:02 crc kubenswrapper[4700]: I0227 17:05:02.226964 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8bf7d"] Feb 27 17:05:02 crc kubenswrapper[4700]: I0227 17:05:02.232903 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8bf7d"] Feb 27 17:05:02 crc kubenswrapper[4700]: I0227 17:05:02.997339 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fd07395-2d50-418c-a75a-64f1f149b216" path="/var/lib/kubelet/pods/1fd07395-2d50-418c-a75a-64f1f149b216/volumes" Feb 27 17:05:05 crc kubenswrapper[4700]: I0227 17:05:05.122343 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5n6qh" Feb 27 17:05:05 crc kubenswrapper[4700]: I0227 17:05:05.123164 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5n6qh" Feb 27 17:05:05 crc kubenswrapper[4700]: I0227 17:05:05.174368 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5n6qh" Feb 27 17:05:05 crc kubenswrapper[4700]: I0227 17:05:05.295007 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fltgg" Feb 27 17:05:05 crc kubenswrapper[4700]: I0227 17:05:05.295111 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fltgg" Feb 27 17:05:05 crc kubenswrapper[4700]: I0227 17:05:05.334832 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fltgg" Feb 27 17:05:06 crc kubenswrapper[4700]: I0227 17:05:06.182320 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fltgg" Feb 27 17:05:06 crc kubenswrapper[4700]: I0227 17:05:06.410225 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:05:06 crc kubenswrapper[4700]: I0227 17:05:06.410285 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:05:06 crc kubenswrapper[4700]: I0227 17:05:06.410334 4700 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" Feb 27 17:05:06 crc kubenswrapper[4700]: I0227 17:05:06.410908 4700 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2ef7c37b61de7a039756248f222bd1769b774703c406de5ad11abd9fb2f46dd2"} pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 17:05:06 crc kubenswrapper[4700]: I0227 17:05:06.410981 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" containerID="cri-o://2ef7c37b61de7a039756248f222bd1769b774703c406de5ad11abd9fb2f46dd2" gracePeriod=600 Feb 27 17:05:07 crc kubenswrapper[4700]: I0227 17:05:07.146000 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9t75w" event={"ID":"89845e2b-1804-445b-8462-36b2350ae663","Type":"ContainerStarted","Data":"d3dd09bfd09d07207b3bb46e8ad9c0e295fa697c60f2c5f65b86bcd9a9e755c8"} Feb 27 17:05:07 crc kubenswrapper[4700]: I0227 17:05:07.148142 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j8wlz" event={"ID":"d1c12d33-6674-4c8a-b871-4d95efca7c23","Type":"ContainerStarted","Data":"2ca68297fff8db72acd0c00f90e1e44a57f3393fafa068967d78d6f8e74d540a"} Feb 27 17:05:07 crc kubenswrapper[4700]: I0227 17:05:07.151995 4700 generic.go:334] "Generic (PLEG): container finished" podID="84b45511-b94c-479f-98db-fd2c4eceec46" containerID="2ef7c37b61de7a039756248f222bd1769b774703c406de5ad11abd9fb2f46dd2" exitCode=0 Feb 27 17:05:07 crc kubenswrapper[4700]: I0227 17:05:07.152086 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerDied","Data":"2ef7c37b61de7a039756248f222bd1769b774703c406de5ad11abd9fb2f46dd2"} Feb 27 17:05:07 crc kubenswrapper[4700]: I0227 17:05:07.173888 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9t75w" podStartSLOduration=3.548297559 podStartE2EDuration="53.173871672s" podCreationTimestamp="2026-02-27 17:04:14 +0000 UTC" firstStartedPulling="2026-02-27 17:04:16.509985012 +0000 UTC m=+216.495297759" lastFinishedPulling="2026-02-27 17:05:06.135559135 +0000 UTC m=+266.120871872" observedRunningTime="2026-02-27 17:05:07.170023189 +0000 UTC m=+267.155335926" watchObservedRunningTime="2026-02-27 17:05:07.173871672 +0000 UTC m=+267.159184419" Feb 27 17:05:07 crc kubenswrapper[4700]: I0227 17:05:07.193489 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-j8wlz" podStartSLOduration=2.677070617 podStartE2EDuration="52.193474054s" podCreationTimestamp="2026-02-27 17:04:15 +0000 UTC" firstStartedPulling="2026-02-27 17:04:16.535945304 +0000 UTC m=+216.521258051" lastFinishedPulling="2026-02-27 17:05:06.052348731 +0000 UTC m=+266.037661488" observedRunningTime="2026-02-27 17:05:07.190832193 +0000 UTC m=+267.176144940" watchObservedRunningTime="2026-02-27 17:05:07.193474054 +0000 UTC m=+267.178786801" Feb 27 17:05:08 crc kubenswrapper[4700]: I0227 17:05:08.093717 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ltq9s" Feb 27 17:05:08 crc kubenswrapper[4700]: I0227 17:05:08.094080 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ltq9s" Feb 27 17:05:08 crc kubenswrapper[4700]: I0227 17:05:08.145059 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ltq9s" Feb 27 17:05:08 crc kubenswrapper[4700]: I0227 17:05:08.161305 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerStarted","Data":"8776130d3c26488575d3ae9c1c22338916b362ebc2dc54951e165061dd752d23"} Feb 27 17:05:08 crc kubenswrapper[4700]: I0227 17:05:08.226780 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ltq9s" Feb 27 17:05:08 crc kubenswrapper[4700]: I0227 17:05:08.786102 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-f4s6x" Feb 27 17:05:08 crc kubenswrapper[4700]: I0227 17:05:08.786399 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-f4s6x" Feb 27 17:05:08 crc kubenswrapper[4700]: I0227 17:05:08.845973 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-f4s6x" Feb 27 17:05:09 crc kubenswrapper[4700]: I0227 17:05:09.225849 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-f4s6x" Feb 27 17:05:09 crc kubenswrapper[4700]: I0227 17:05:09.628038 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fltgg"] Feb 27 17:05:09 crc kubenswrapper[4700]: I0227 17:05:09.628421 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fltgg" podUID="bc1137d0-2659-430b-8d76-d08017e7f133" containerName="registry-server" containerID="cri-o://1b3935e941a6d1b5664d559bb603e5aeb21fa780c268b235a7dc91820e59b219" gracePeriod=2 Feb 27 17:05:10 crc kubenswrapper[4700]: I0227 17:05:10.179905 4700 generic.go:334] "Generic (PLEG): container finished" podID="bc1137d0-2659-430b-8d76-d08017e7f133" containerID="1b3935e941a6d1b5664d559bb603e5aeb21fa780c268b235a7dc91820e59b219" exitCode=0 Feb 27 17:05:10 crc kubenswrapper[4700]: I0227 17:05:10.180024 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fltgg" event={"ID":"bc1137d0-2659-430b-8d76-d08017e7f133","Type":"ContainerDied","Data":"1b3935e941a6d1b5664d559bb603e5aeb21fa780c268b235a7dc91820e59b219"} Feb 27 17:05:10 crc kubenswrapper[4700]: I0227 17:05:10.299452 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fltgg" Feb 27 17:05:10 crc kubenswrapper[4700]: I0227 17:05:10.347160 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc1137d0-2659-430b-8d76-d08017e7f133-catalog-content\") pod \"bc1137d0-2659-430b-8d76-d08017e7f133\" (UID: \"bc1137d0-2659-430b-8d76-d08017e7f133\") " Feb 27 17:05:10 crc kubenswrapper[4700]: I0227 17:05:10.347257 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqkjd\" (UniqueName: \"kubernetes.io/projected/bc1137d0-2659-430b-8d76-d08017e7f133-kube-api-access-lqkjd\") pod \"bc1137d0-2659-430b-8d76-d08017e7f133\" (UID: \"bc1137d0-2659-430b-8d76-d08017e7f133\") " Feb 27 17:05:10 crc kubenswrapper[4700]: I0227 17:05:10.347307 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc1137d0-2659-430b-8d76-d08017e7f133-utilities\") pod \"bc1137d0-2659-430b-8d76-d08017e7f133\" (UID: \"bc1137d0-2659-430b-8d76-d08017e7f133\") " Feb 27 17:05:10 crc kubenswrapper[4700]: I0227 17:05:10.348843 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc1137d0-2659-430b-8d76-d08017e7f133-utilities" (OuterVolumeSpecName: "utilities") pod "bc1137d0-2659-430b-8d76-d08017e7f133" (UID: "bc1137d0-2659-430b-8d76-d08017e7f133"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:05:10 crc kubenswrapper[4700]: I0227 17:05:10.356662 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc1137d0-2659-430b-8d76-d08017e7f133-kube-api-access-lqkjd" (OuterVolumeSpecName: "kube-api-access-lqkjd") pod "bc1137d0-2659-430b-8d76-d08017e7f133" (UID: "bc1137d0-2659-430b-8d76-d08017e7f133"). InnerVolumeSpecName "kube-api-access-lqkjd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:05:10 crc kubenswrapper[4700]: I0227 17:05:10.411245 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc1137d0-2659-430b-8d76-d08017e7f133-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bc1137d0-2659-430b-8d76-d08017e7f133" (UID: "bc1137d0-2659-430b-8d76-d08017e7f133"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:05:10 crc kubenswrapper[4700]: I0227 17:05:10.449159 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc1137d0-2659-430b-8d76-d08017e7f133-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:10 crc kubenswrapper[4700]: I0227 17:05:10.449206 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqkjd\" (UniqueName: \"kubernetes.io/projected/bc1137d0-2659-430b-8d76-d08017e7f133-kube-api-access-lqkjd\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:10 crc kubenswrapper[4700]: I0227 17:05:10.449228 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc1137d0-2659-430b-8d76-d08017e7f133-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.187730 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fltgg" event={"ID":"bc1137d0-2659-430b-8d76-d08017e7f133","Type":"ContainerDied","Data":"4239dc6fe2b6b1c90b3c6c4f883793b5d34c72d842095d1b21780c9c04dc6646"} Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.188154 4700 scope.go:117] "RemoveContainer" containerID="1b3935e941a6d1b5664d559bb603e5aeb21fa780c268b235a7dc91820e59b219" Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.187981 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fltgg" Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.216278 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fltgg"] Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.218980 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fltgg"] Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.239705 4700 scope.go:117] "RemoveContainer" containerID="2990c00b77ba4917913964a3cc3214736412a8da0a477191ac3f1952be9f98cc" Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.272649 4700 scope.go:117] "RemoveContainer" containerID="038784b5566875c18e91eea7fd9f0fe0438c8f7d149a8a3a40fb10426e904368" Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.310248 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7695c4f755-xt7w8"] Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.310513 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7695c4f755-xt7w8" podUID="b63de308-633e-4e7a-8428-9d5a24a6458d" containerName="controller-manager" containerID="cri-o://bba3b8f51ee422cf402a58ab7fa9c4508221abb9df63d94fd77979f794917fdb" gracePeriod=30 Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.325542 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-566f554dd9-8tfvw"] Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.325741 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-566f554dd9-8tfvw" podUID="45f2f427-ea74-4141-9b15-6079f565a1ba" containerName="route-controller-manager" containerID="cri-o://ac829b1c6463c3a5ee1423a4800ea90a27fe65b08e64813a059f3de3f4a74b6a" gracePeriod=30 Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.422489 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f4s6x"] Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.422832 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-f4s6x" podUID="8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce" containerName="registry-server" containerID="cri-o://92940ff183c7ddcf11b4208e82e0c12bdc7f2bc25fec01613de65a1d13892a3a" gracePeriod=2 Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.844550 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f4s6x" Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.873158 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-566f554dd9-8tfvw" Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.917841 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7695c4f755-xt7w8" Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.972759 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce-catalog-content\") pod \"8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce\" (UID: \"8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce\") " Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.972835 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce-utilities\") pod \"8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce\" (UID: \"8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce\") " Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.972868 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vpbh9\" (UniqueName: \"kubernetes.io/projected/8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce-kube-api-access-vpbh9\") pod \"8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce\" (UID: \"8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce\") " Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.972897 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b63de308-633e-4e7a-8428-9d5a24a6458d-config\") pod \"b63de308-633e-4e7a-8428-9d5a24a6458d\" (UID: \"b63de308-633e-4e7a-8428-9d5a24a6458d\") " Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.972943 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45f2f427-ea74-4141-9b15-6079f565a1ba-config\") pod \"45f2f427-ea74-4141-9b15-6079f565a1ba\" (UID: \"45f2f427-ea74-4141-9b15-6079f565a1ba\") " Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.972969 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45f2f427-ea74-4141-9b15-6079f565a1ba-serving-cert\") pod \"45f2f427-ea74-4141-9b15-6079f565a1ba\" (UID: \"45f2f427-ea74-4141-9b15-6079f565a1ba\") " Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.972993 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b63de308-633e-4e7a-8428-9d5a24a6458d-proxy-ca-bundles\") pod \"b63de308-633e-4e7a-8428-9d5a24a6458d\" (UID: \"b63de308-633e-4e7a-8428-9d5a24a6458d\") " Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.973023 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b63de308-633e-4e7a-8428-9d5a24a6458d-serving-cert\") pod \"b63de308-633e-4e7a-8428-9d5a24a6458d\" (UID: \"b63de308-633e-4e7a-8428-9d5a24a6458d\") " Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.973092 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvc5q\" (UniqueName: \"kubernetes.io/projected/b63de308-633e-4e7a-8428-9d5a24a6458d-kube-api-access-qvc5q\") pod \"b63de308-633e-4e7a-8428-9d5a24a6458d\" (UID: \"b63de308-633e-4e7a-8428-9d5a24a6458d\") " Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.973120 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hhbw\" (UniqueName: \"kubernetes.io/projected/45f2f427-ea74-4141-9b15-6079f565a1ba-kube-api-access-9hhbw\") pod \"45f2f427-ea74-4141-9b15-6079f565a1ba\" (UID: \"45f2f427-ea74-4141-9b15-6079f565a1ba\") " Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.973150 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/45f2f427-ea74-4141-9b15-6079f565a1ba-client-ca\") pod \"45f2f427-ea74-4141-9b15-6079f565a1ba\" (UID: \"45f2f427-ea74-4141-9b15-6079f565a1ba\") " Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.973170 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b63de308-633e-4e7a-8428-9d5a24a6458d-client-ca\") pod \"b63de308-633e-4e7a-8428-9d5a24a6458d\" (UID: \"b63de308-633e-4e7a-8428-9d5a24a6458d\") " Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.973595 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce-utilities" (OuterVolumeSpecName: "utilities") pod "8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce" (UID: "8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.973846 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b63de308-633e-4e7a-8428-9d5a24a6458d-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "b63de308-633e-4e7a-8428-9d5a24a6458d" (UID: "b63de308-633e-4e7a-8428-9d5a24a6458d"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.973878 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b63de308-633e-4e7a-8428-9d5a24a6458d-client-ca" (OuterVolumeSpecName: "client-ca") pod "b63de308-633e-4e7a-8428-9d5a24a6458d" (UID: "b63de308-633e-4e7a-8428-9d5a24a6458d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.974028 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45f2f427-ea74-4141-9b15-6079f565a1ba-client-ca" (OuterVolumeSpecName: "client-ca") pod "45f2f427-ea74-4141-9b15-6079f565a1ba" (UID: "45f2f427-ea74-4141-9b15-6079f565a1ba"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.974186 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b63de308-633e-4e7a-8428-9d5a24a6458d-config" (OuterVolumeSpecName: "config") pod "b63de308-633e-4e7a-8428-9d5a24a6458d" (UID: "b63de308-633e-4e7a-8428-9d5a24a6458d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.974547 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45f2f427-ea74-4141-9b15-6079f565a1ba-config" (OuterVolumeSpecName: "config") pod "45f2f427-ea74-4141-9b15-6079f565a1ba" (UID: "45f2f427-ea74-4141-9b15-6079f565a1ba"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.977143 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b63de308-633e-4e7a-8428-9d5a24a6458d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b63de308-633e-4e7a-8428-9d5a24a6458d" (UID: "b63de308-633e-4e7a-8428-9d5a24a6458d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.977958 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce-kube-api-access-vpbh9" (OuterVolumeSpecName: "kube-api-access-vpbh9") pod "8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce" (UID: "8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce"). InnerVolumeSpecName "kube-api-access-vpbh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.978735 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45f2f427-ea74-4141-9b15-6079f565a1ba-kube-api-access-9hhbw" (OuterVolumeSpecName: "kube-api-access-9hhbw") pod "45f2f427-ea74-4141-9b15-6079f565a1ba" (UID: "45f2f427-ea74-4141-9b15-6079f565a1ba"). InnerVolumeSpecName "kube-api-access-9hhbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.978786 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b63de308-633e-4e7a-8428-9d5a24a6458d-kube-api-access-qvc5q" (OuterVolumeSpecName: "kube-api-access-qvc5q") pod "b63de308-633e-4e7a-8428-9d5a24a6458d" (UID: "b63de308-633e-4e7a-8428-9d5a24a6458d"). InnerVolumeSpecName "kube-api-access-qvc5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:05:11 crc kubenswrapper[4700]: I0227 17:05:11.978895 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45f2f427-ea74-4141-9b15-6079f565a1ba-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "45f2f427-ea74-4141-9b15-6079f565a1ba" (UID: "45f2f427-ea74-4141-9b15-6079f565a1ba"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.075026 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45f2f427-ea74-4141-9b15-6079f565a1ba-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.075067 4700 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45f2f427-ea74-4141-9b15-6079f565a1ba-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.075086 4700 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b63de308-633e-4e7a-8428-9d5a24a6458d-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.075108 4700 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b63de308-633e-4e7a-8428-9d5a24a6458d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.075126 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvc5q\" (UniqueName: \"kubernetes.io/projected/b63de308-633e-4e7a-8428-9d5a24a6458d-kube-api-access-qvc5q\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.075144 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hhbw\" (UniqueName: \"kubernetes.io/projected/45f2f427-ea74-4141-9b15-6079f565a1ba-kube-api-access-9hhbw\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.075160 4700 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/45f2f427-ea74-4141-9b15-6079f565a1ba-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.075176 4700 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b63de308-633e-4e7a-8428-9d5a24a6458d-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.075232 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.075249 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vpbh9\" (UniqueName: \"kubernetes.io/projected/8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce-kube-api-access-vpbh9\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.075265 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b63de308-633e-4e7a-8428-9d5a24a6458d-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.130862 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce" (UID: "8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.176245 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.197328 4700 generic.go:334] "Generic (PLEG): container finished" podID="8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce" containerID="92940ff183c7ddcf11b4208e82e0c12bdc7f2bc25fec01613de65a1d13892a3a" exitCode=0 Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.197411 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f4s6x" event={"ID":"8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce","Type":"ContainerDied","Data":"92940ff183c7ddcf11b4208e82e0c12bdc7f2bc25fec01613de65a1d13892a3a"} Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.198604 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f4s6x" event={"ID":"8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce","Type":"ContainerDied","Data":"225e070473cae716ff3eec8053e05d138c0c93db36fed520e374b603853a93ab"} Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.198649 4700 scope.go:117] "RemoveContainer" containerID="92940ff183c7ddcf11b4208e82e0c12bdc7f2bc25fec01613de65a1d13892a3a" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.197438 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f4s6x" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.203913 4700 generic.go:334] "Generic (PLEG): container finished" podID="b63de308-633e-4e7a-8428-9d5a24a6458d" containerID="bba3b8f51ee422cf402a58ab7fa9c4508221abb9df63d94fd77979f794917fdb" exitCode=0 Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.204041 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7695c4f755-xt7w8" event={"ID":"b63de308-633e-4e7a-8428-9d5a24a6458d","Type":"ContainerDied","Data":"bba3b8f51ee422cf402a58ab7fa9c4508221abb9df63d94fd77979f794917fdb"} Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.204072 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7695c4f755-xt7w8" event={"ID":"b63de308-633e-4e7a-8428-9d5a24a6458d","Type":"ContainerDied","Data":"13c3f854b4af9b1b969b7ceffd7a4237785a18fd776ff3b2c9bc51e184f3b7d4"} Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.204130 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7695c4f755-xt7w8" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.208829 4700 generic.go:334] "Generic (PLEG): container finished" podID="45f2f427-ea74-4141-9b15-6079f565a1ba" containerID="ac829b1c6463c3a5ee1423a4800ea90a27fe65b08e64813a059f3de3f4a74b6a" exitCode=0 Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.208925 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-566f554dd9-8tfvw" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.208971 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-566f554dd9-8tfvw" event={"ID":"45f2f427-ea74-4141-9b15-6079f565a1ba","Type":"ContainerDied","Data":"ac829b1c6463c3a5ee1423a4800ea90a27fe65b08e64813a059f3de3f4a74b6a"} Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.209024 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-566f554dd9-8tfvw" event={"ID":"45f2f427-ea74-4141-9b15-6079f565a1ba","Type":"ContainerDied","Data":"61f0ae127ec3a7af7fe5fb118a62fbee4b502fdb3da3c52be1f768aafcec1af9"} Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.232871 4700 scope.go:117] "RemoveContainer" containerID="b44d4ea4d0907acb75324ec5ab7a15e205af7ff60f93171d32f0a22b0b6be541" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.250993 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f4s6x"] Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.256196 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-f4s6x"] Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.277128 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-566f554dd9-8tfvw"] Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.286218 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-566f554dd9-8tfvw"] Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.290335 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7695c4f755-xt7w8"] Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.294391 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7695c4f755-xt7w8"] Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.302803 4700 scope.go:117] "RemoveContainer" containerID="e15d3e5ae320d4ee0be90673b5bd412d4649023e638968b2f1d880715bb3499c" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.327582 4700 scope.go:117] "RemoveContainer" containerID="92940ff183c7ddcf11b4208e82e0c12bdc7f2bc25fec01613de65a1d13892a3a" Feb 27 17:05:12 crc kubenswrapper[4700]: E0227 17:05:12.328062 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92940ff183c7ddcf11b4208e82e0c12bdc7f2bc25fec01613de65a1d13892a3a\": container with ID starting with 92940ff183c7ddcf11b4208e82e0c12bdc7f2bc25fec01613de65a1d13892a3a not found: ID does not exist" containerID="92940ff183c7ddcf11b4208e82e0c12bdc7f2bc25fec01613de65a1d13892a3a" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.328121 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92940ff183c7ddcf11b4208e82e0c12bdc7f2bc25fec01613de65a1d13892a3a"} err="failed to get container status \"92940ff183c7ddcf11b4208e82e0c12bdc7f2bc25fec01613de65a1d13892a3a\": rpc error: code = NotFound desc = could not find container \"92940ff183c7ddcf11b4208e82e0c12bdc7f2bc25fec01613de65a1d13892a3a\": container with ID starting with 92940ff183c7ddcf11b4208e82e0c12bdc7f2bc25fec01613de65a1d13892a3a not found: ID does not exist" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.328164 4700 scope.go:117] "RemoveContainer" containerID="b44d4ea4d0907acb75324ec5ab7a15e205af7ff60f93171d32f0a22b0b6be541" Feb 27 17:05:12 crc kubenswrapper[4700]: E0227 17:05:12.328850 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b44d4ea4d0907acb75324ec5ab7a15e205af7ff60f93171d32f0a22b0b6be541\": container with ID starting with b44d4ea4d0907acb75324ec5ab7a15e205af7ff60f93171d32f0a22b0b6be541 not found: ID does not exist" containerID="b44d4ea4d0907acb75324ec5ab7a15e205af7ff60f93171d32f0a22b0b6be541" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.328903 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b44d4ea4d0907acb75324ec5ab7a15e205af7ff60f93171d32f0a22b0b6be541"} err="failed to get container status \"b44d4ea4d0907acb75324ec5ab7a15e205af7ff60f93171d32f0a22b0b6be541\": rpc error: code = NotFound desc = could not find container \"b44d4ea4d0907acb75324ec5ab7a15e205af7ff60f93171d32f0a22b0b6be541\": container with ID starting with b44d4ea4d0907acb75324ec5ab7a15e205af7ff60f93171d32f0a22b0b6be541 not found: ID does not exist" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.328940 4700 scope.go:117] "RemoveContainer" containerID="e15d3e5ae320d4ee0be90673b5bd412d4649023e638968b2f1d880715bb3499c" Feb 27 17:05:12 crc kubenswrapper[4700]: E0227 17:05:12.329306 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e15d3e5ae320d4ee0be90673b5bd412d4649023e638968b2f1d880715bb3499c\": container with ID starting with e15d3e5ae320d4ee0be90673b5bd412d4649023e638968b2f1d880715bb3499c not found: ID does not exist" containerID="e15d3e5ae320d4ee0be90673b5bd412d4649023e638968b2f1d880715bb3499c" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.329351 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e15d3e5ae320d4ee0be90673b5bd412d4649023e638968b2f1d880715bb3499c"} err="failed to get container status \"e15d3e5ae320d4ee0be90673b5bd412d4649023e638968b2f1d880715bb3499c\": rpc error: code = NotFound desc = could not find container \"e15d3e5ae320d4ee0be90673b5bd412d4649023e638968b2f1d880715bb3499c\": container with ID starting with e15d3e5ae320d4ee0be90673b5bd412d4649023e638968b2f1d880715bb3499c not found: ID does not exist" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.329381 4700 scope.go:117] "RemoveContainer" containerID="bba3b8f51ee422cf402a58ab7fa9c4508221abb9df63d94fd77979f794917fdb" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.351755 4700 scope.go:117] "RemoveContainer" containerID="bba3b8f51ee422cf402a58ab7fa9c4508221abb9df63d94fd77979f794917fdb" Feb 27 17:05:12 crc kubenswrapper[4700]: E0227 17:05:12.352450 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bba3b8f51ee422cf402a58ab7fa9c4508221abb9df63d94fd77979f794917fdb\": container with ID starting with bba3b8f51ee422cf402a58ab7fa9c4508221abb9df63d94fd77979f794917fdb not found: ID does not exist" containerID="bba3b8f51ee422cf402a58ab7fa9c4508221abb9df63d94fd77979f794917fdb" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.352549 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bba3b8f51ee422cf402a58ab7fa9c4508221abb9df63d94fd77979f794917fdb"} err="failed to get container status \"bba3b8f51ee422cf402a58ab7fa9c4508221abb9df63d94fd77979f794917fdb\": rpc error: code = NotFound desc = could not find container \"bba3b8f51ee422cf402a58ab7fa9c4508221abb9df63d94fd77979f794917fdb\": container with ID starting with bba3b8f51ee422cf402a58ab7fa9c4508221abb9df63d94fd77979f794917fdb not found: ID does not exist" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.352603 4700 scope.go:117] "RemoveContainer" containerID="ac829b1c6463c3a5ee1423a4800ea90a27fe65b08e64813a059f3de3f4a74b6a" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.370403 4700 scope.go:117] "RemoveContainer" containerID="ac829b1c6463c3a5ee1423a4800ea90a27fe65b08e64813a059f3de3f4a74b6a" Feb 27 17:05:12 crc kubenswrapper[4700]: E0227 17:05:12.370914 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac829b1c6463c3a5ee1423a4800ea90a27fe65b08e64813a059f3de3f4a74b6a\": container with ID starting with ac829b1c6463c3a5ee1423a4800ea90a27fe65b08e64813a059f3de3f4a74b6a not found: ID does not exist" containerID="ac829b1c6463c3a5ee1423a4800ea90a27fe65b08e64813a059f3de3f4a74b6a" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.371117 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac829b1c6463c3a5ee1423a4800ea90a27fe65b08e64813a059f3de3f4a74b6a"} err="failed to get container status \"ac829b1c6463c3a5ee1423a4800ea90a27fe65b08e64813a059f3de3f4a74b6a\": rpc error: code = NotFound desc = could not find container \"ac829b1c6463c3a5ee1423a4800ea90a27fe65b08e64813a059f3de3f4a74b6a\": container with ID starting with ac829b1c6463c3a5ee1423a4800ea90a27fe65b08e64813a059f3de3f4a74b6a not found: ID does not exist" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.754479 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-685c5585dd-mknx2"] Feb 27 17:05:12 crc kubenswrapper[4700]: E0227 17:05:12.755316 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fd07395-2d50-418c-a75a-64f1f149b216" containerName="extract-utilities" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.755506 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fd07395-2d50-418c-a75a-64f1f149b216" containerName="extract-utilities" Feb 27 17:05:12 crc kubenswrapper[4700]: E0227 17:05:12.755640 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fd07395-2d50-418c-a75a-64f1f149b216" containerName="registry-server" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.755755 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fd07395-2d50-418c-a75a-64f1f149b216" containerName="registry-server" Feb 27 17:05:12 crc kubenswrapper[4700]: E0227 17:05:12.755869 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce" containerName="extract-content" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.755977 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce" containerName="extract-content" Feb 27 17:05:12 crc kubenswrapper[4700]: E0227 17:05:12.756091 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45f2f427-ea74-4141-9b15-6079f565a1ba" containerName="route-controller-manager" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.756220 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="45f2f427-ea74-4141-9b15-6079f565a1ba" containerName="route-controller-manager" Feb 27 17:05:12 crc kubenswrapper[4700]: E0227 17:05:12.756341 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce" containerName="extract-utilities" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.756491 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce" containerName="extract-utilities" Feb 27 17:05:12 crc kubenswrapper[4700]: E0227 17:05:12.756625 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fd07395-2d50-418c-a75a-64f1f149b216" containerName="extract-content" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.756735 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fd07395-2d50-418c-a75a-64f1f149b216" containerName="extract-content" Feb 27 17:05:12 crc kubenswrapper[4700]: E0227 17:05:12.756876 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc1137d0-2659-430b-8d76-d08017e7f133" containerName="extract-utilities" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.757009 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc1137d0-2659-430b-8d76-d08017e7f133" containerName="extract-utilities" Feb 27 17:05:12 crc kubenswrapper[4700]: E0227 17:05:12.757153 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc1137d0-2659-430b-8d76-d08017e7f133" containerName="registry-server" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.757284 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc1137d0-2659-430b-8d76-d08017e7f133" containerName="registry-server" Feb 27 17:05:12 crc kubenswrapper[4700]: E0227 17:05:12.757401 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc1137d0-2659-430b-8d76-d08017e7f133" containerName="extract-content" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.757540 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc1137d0-2659-430b-8d76-d08017e7f133" containerName="extract-content" Feb 27 17:05:12 crc kubenswrapper[4700]: E0227 17:05:12.757669 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b63de308-633e-4e7a-8428-9d5a24a6458d" containerName="controller-manager" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.757805 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="b63de308-633e-4e7a-8428-9d5a24a6458d" containerName="controller-manager" Feb 27 17:05:12 crc kubenswrapper[4700]: E0227 17:05:12.757923 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce" containerName="registry-server" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.758033 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce" containerName="registry-server" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.758310 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce" containerName="registry-server" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.758443 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="b63de308-633e-4e7a-8428-9d5a24a6458d" containerName="controller-manager" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.758609 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc1137d0-2659-430b-8d76-d08017e7f133" containerName="registry-server" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.758758 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fd07395-2d50-418c-a75a-64f1f149b216" containerName="registry-server" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.758886 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="45f2f427-ea74-4141-9b15-6079f565a1ba" containerName="route-controller-manager" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.759616 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-685c5585dd-mknx2" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.784009 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.784176 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.784720 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.785159 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.786310 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.786750 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.787896 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6cd66f6876-tgwlc"] Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.791827 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6cd66f6876-tgwlc" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.797234 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.802036 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.802201 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.803001 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.803220 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.803247 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.803631 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.808768 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6cd66f6876-tgwlc"] Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.821108 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-685c5585dd-mknx2"] Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.910925 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03-config\") pod \"controller-manager-685c5585dd-mknx2\" (UID: \"59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03\") " pod="openshift-controller-manager/controller-manager-685c5585dd-mknx2" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.910987 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03-proxy-ca-bundles\") pod \"controller-manager-685c5585dd-mknx2\" (UID: \"59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03\") " pod="openshift-controller-manager/controller-manager-685c5585dd-mknx2" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.911015 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03-serving-cert\") pod \"controller-manager-685c5585dd-mknx2\" (UID: \"59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03\") " pod="openshift-controller-manager/controller-manager-685c5585dd-mknx2" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.911031 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqlxq\" (UniqueName: \"kubernetes.io/projected/59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03-kube-api-access-qqlxq\") pod \"controller-manager-685c5585dd-mknx2\" (UID: \"59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03\") " pod="openshift-controller-manager/controller-manager-685c5585dd-mknx2" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.911145 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f1cb2d6-8127-4a75-bdbd-751f69ef0613-serving-cert\") pod \"route-controller-manager-6cd66f6876-tgwlc\" (UID: \"8f1cb2d6-8127-4a75-bdbd-751f69ef0613\") " pod="openshift-route-controller-manager/route-controller-manager-6cd66f6876-tgwlc" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.911185 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xc5vp\" (UniqueName: \"kubernetes.io/projected/8f1cb2d6-8127-4a75-bdbd-751f69ef0613-kube-api-access-xc5vp\") pod \"route-controller-manager-6cd66f6876-tgwlc\" (UID: \"8f1cb2d6-8127-4a75-bdbd-751f69ef0613\") " pod="openshift-route-controller-manager/route-controller-manager-6cd66f6876-tgwlc" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.911227 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8f1cb2d6-8127-4a75-bdbd-751f69ef0613-client-ca\") pod \"route-controller-manager-6cd66f6876-tgwlc\" (UID: \"8f1cb2d6-8127-4a75-bdbd-751f69ef0613\") " pod="openshift-route-controller-manager/route-controller-manager-6cd66f6876-tgwlc" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.911251 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03-client-ca\") pod \"controller-manager-685c5585dd-mknx2\" (UID: \"59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03\") " pod="openshift-controller-manager/controller-manager-685c5585dd-mknx2" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.911265 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f1cb2d6-8127-4a75-bdbd-751f69ef0613-config\") pod \"route-controller-manager-6cd66f6876-tgwlc\" (UID: \"8f1cb2d6-8127-4a75-bdbd-751f69ef0613\") " pod="openshift-route-controller-manager/route-controller-manager-6cd66f6876-tgwlc" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.990992 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45f2f427-ea74-4141-9b15-6079f565a1ba" path="/var/lib/kubelet/pods/45f2f427-ea74-4141-9b15-6079f565a1ba/volumes" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.992621 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce" path="/var/lib/kubelet/pods/8fd5ee7b-1468-4a27-8bb9-e0f81d7f0cce/volumes" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.994583 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b63de308-633e-4e7a-8428-9d5a24a6458d" path="/var/lib/kubelet/pods/b63de308-633e-4e7a-8428-9d5a24a6458d/volumes" Feb 27 17:05:12 crc kubenswrapper[4700]: I0227 17:05:12.997230 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc1137d0-2659-430b-8d76-d08017e7f133" path="/var/lib/kubelet/pods/bc1137d0-2659-430b-8d76-d08017e7f133/volumes" Feb 27 17:05:13 crc kubenswrapper[4700]: I0227 17:05:13.012333 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03-config\") pod \"controller-manager-685c5585dd-mknx2\" (UID: \"59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03\") " pod="openshift-controller-manager/controller-manager-685c5585dd-mknx2" Feb 27 17:05:13 crc kubenswrapper[4700]: I0227 17:05:13.012523 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03-proxy-ca-bundles\") pod \"controller-manager-685c5585dd-mknx2\" (UID: \"59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03\") " pod="openshift-controller-manager/controller-manager-685c5585dd-mknx2" Feb 27 17:05:13 crc kubenswrapper[4700]: I0227 17:05:13.012679 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03-serving-cert\") pod \"controller-manager-685c5585dd-mknx2\" (UID: \"59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03\") " pod="openshift-controller-manager/controller-manager-685c5585dd-mknx2" Feb 27 17:05:13 crc kubenswrapper[4700]: I0227 17:05:13.012898 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqlxq\" (UniqueName: \"kubernetes.io/projected/59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03-kube-api-access-qqlxq\") pod \"controller-manager-685c5585dd-mknx2\" (UID: \"59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03\") " pod="openshift-controller-manager/controller-manager-685c5585dd-mknx2" Feb 27 17:05:13 crc kubenswrapper[4700]: I0227 17:05:13.013069 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f1cb2d6-8127-4a75-bdbd-751f69ef0613-serving-cert\") pod \"route-controller-manager-6cd66f6876-tgwlc\" (UID: \"8f1cb2d6-8127-4a75-bdbd-751f69ef0613\") " pod="openshift-route-controller-manager/route-controller-manager-6cd66f6876-tgwlc" Feb 27 17:05:13 crc kubenswrapper[4700]: I0227 17:05:13.013183 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xc5vp\" (UniqueName: \"kubernetes.io/projected/8f1cb2d6-8127-4a75-bdbd-751f69ef0613-kube-api-access-xc5vp\") pod \"route-controller-manager-6cd66f6876-tgwlc\" (UID: \"8f1cb2d6-8127-4a75-bdbd-751f69ef0613\") " pod="openshift-route-controller-manager/route-controller-manager-6cd66f6876-tgwlc" Feb 27 17:05:13 crc kubenswrapper[4700]: I0227 17:05:13.013328 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8f1cb2d6-8127-4a75-bdbd-751f69ef0613-client-ca\") pod \"route-controller-manager-6cd66f6876-tgwlc\" (UID: \"8f1cb2d6-8127-4a75-bdbd-751f69ef0613\") " pod="openshift-route-controller-manager/route-controller-manager-6cd66f6876-tgwlc" Feb 27 17:05:13 crc kubenswrapper[4700]: I0227 17:05:13.013443 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03-client-ca\") pod \"controller-manager-685c5585dd-mknx2\" (UID: \"59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03\") " pod="openshift-controller-manager/controller-manager-685c5585dd-mknx2" Feb 27 17:05:13 crc kubenswrapper[4700]: I0227 17:05:13.013575 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f1cb2d6-8127-4a75-bdbd-751f69ef0613-config\") pod \"route-controller-manager-6cd66f6876-tgwlc\" (UID: \"8f1cb2d6-8127-4a75-bdbd-751f69ef0613\") " pod="openshift-route-controller-manager/route-controller-manager-6cd66f6876-tgwlc" Feb 27 17:05:13 crc kubenswrapper[4700]: I0227 17:05:13.013791 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03-config\") pod \"controller-manager-685c5585dd-mknx2\" (UID: \"59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03\") " pod="openshift-controller-manager/controller-manager-685c5585dd-mknx2" Feb 27 17:05:13 crc kubenswrapper[4700]: I0227 17:05:13.014451 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03-proxy-ca-bundles\") pod \"controller-manager-685c5585dd-mknx2\" (UID: \"59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03\") " pod="openshift-controller-manager/controller-manager-685c5585dd-mknx2" Feb 27 17:05:13 crc kubenswrapper[4700]: I0227 17:05:13.014821 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03-client-ca\") pod \"controller-manager-685c5585dd-mknx2\" (UID: \"59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03\") " pod="openshift-controller-manager/controller-manager-685c5585dd-mknx2" Feb 27 17:05:13 crc kubenswrapper[4700]: I0227 17:05:13.015188 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8f1cb2d6-8127-4a75-bdbd-751f69ef0613-client-ca\") pod \"route-controller-manager-6cd66f6876-tgwlc\" (UID: \"8f1cb2d6-8127-4a75-bdbd-751f69ef0613\") " pod="openshift-route-controller-manager/route-controller-manager-6cd66f6876-tgwlc" Feb 27 17:05:13 crc kubenswrapper[4700]: I0227 17:05:13.016542 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f1cb2d6-8127-4a75-bdbd-751f69ef0613-config\") pod \"route-controller-manager-6cd66f6876-tgwlc\" (UID: \"8f1cb2d6-8127-4a75-bdbd-751f69ef0613\") " pod="openshift-route-controller-manager/route-controller-manager-6cd66f6876-tgwlc" Feb 27 17:05:13 crc kubenswrapper[4700]: I0227 17:05:13.023413 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03-serving-cert\") pod \"controller-manager-685c5585dd-mknx2\" (UID: \"59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03\") " pod="openshift-controller-manager/controller-manager-685c5585dd-mknx2" Feb 27 17:05:13 crc kubenswrapper[4700]: I0227 17:05:13.027810 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f1cb2d6-8127-4a75-bdbd-751f69ef0613-serving-cert\") pod \"route-controller-manager-6cd66f6876-tgwlc\" (UID: \"8f1cb2d6-8127-4a75-bdbd-751f69ef0613\") " pod="openshift-route-controller-manager/route-controller-manager-6cd66f6876-tgwlc" Feb 27 17:05:13 crc kubenswrapper[4700]: I0227 17:05:13.035684 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xc5vp\" (UniqueName: \"kubernetes.io/projected/8f1cb2d6-8127-4a75-bdbd-751f69ef0613-kube-api-access-xc5vp\") pod \"route-controller-manager-6cd66f6876-tgwlc\" (UID: \"8f1cb2d6-8127-4a75-bdbd-751f69ef0613\") " pod="openshift-route-controller-manager/route-controller-manager-6cd66f6876-tgwlc" Feb 27 17:05:13 crc kubenswrapper[4700]: I0227 17:05:13.037575 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqlxq\" (UniqueName: \"kubernetes.io/projected/59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03-kube-api-access-qqlxq\") pod \"controller-manager-685c5585dd-mknx2\" (UID: \"59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03\") " pod="openshift-controller-manager/controller-manager-685c5585dd-mknx2" Feb 27 17:05:13 crc kubenswrapper[4700]: I0227 17:05:13.102788 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-685c5585dd-mknx2" Feb 27 17:05:13 crc kubenswrapper[4700]: I0227 17:05:13.136710 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6cd66f6876-tgwlc" Feb 27 17:05:13 crc kubenswrapper[4700]: I0227 17:05:13.381633 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-685c5585dd-mknx2"] Feb 27 17:05:13 crc kubenswrapper[4700]: W0227 17:05:13.393750 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59722ff3_7fcc_4b0b_bea3_4abf3a2d3a03.slice/crio-021fbacddef86943aa7b150a0a781e7b94a82d098f95e4fa80e06f3ee47a5e9b WatchSource:0}: Error finding container 021fbacddef86943aa7b150a0a781e7b94a82d098f95e4fa80e06f3ee47a5e9b: Status 404 returned error can't find the container with id 021fbacddef86943aa7b150a0a781e7b94a82d098f95e4fa80e06f3ee47a5e9b Feb 27 17:05:13 crc kubenswrapper[4700]: I0227 17:05:13.455919 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6cd66f6876-tgwlc"] Feb 27 17:05:13 crc kubenswrapper[4700]: W0227 17:05:13.461937 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f1cb2d6_8127_4a75_bdbd_751f69ef0613.slice/crio-992e550f31f30277b3b90903ade030f8c733504eca2e7d903f6b31a129a6bec0 WatchSource:0}: Error finding container 992e550f31f30277b3b90903ade030f8c733504eca2e7d903f6b31a129a6bec0: Status 404 returned error can't find the container with id 992e550f31f30277b3b90903ade030f8c733504eca2e7d903f6b31a129a6bec0 Feb 27 17:05:14 crc kubenswrapper[4700]: I0227 17:05:14.254698 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-685c5585dd-mknx2" event={"ID":"59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03","Type":"ContainerStarted","Data":"0fbc4956eeca2f4ef308af038a9f4cd9a99b6cfc1d22b3b30d4e0bb6e708d687"} Feb 27 17:05:14 crc kubenswrapper[4700]: I0227 17:05:14.254750 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-685c5585dd-mknx2" event={"ID":"59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03","Type":"ContainerStarted","Data":"021fbacddef86943aa7b150a0a781e7b94a82d098f95e4fa80e06f3ee47a5e9b"} Feb 27 17:05:14 crc kubenswrapper[4700]: I0227 17:05:14.254913 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-685c5585dd-mknx2" Feb 27 17:05:14 crc kubenswrapper[4700]: I0227 17:05:14.257361 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6cd66f6876-tgwlc" event={"ID":"8f1cb2d6-8127-4a75-bdbd-751f69ef0613","Type":"ContainerStarted","Data":"19b111c34447df31b93ccabe652c10ac18aac74aacbd74268c519c2e1590d2ff"} Feb 27 17:05:14 crc kubenswrapper[4700]: I0227 17:05:14.257394 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6cd66f6876-tgwlc" event={"ID":"8f1cb2d6-8127-4a75-bdbd-751f69ef0613","Type":"ContainerStarted","Data":"992e550f31f30277b3b90903ade030f8c733504eca2e7d903f6b31a129a6bec0"} Feb 27 17:05:14 crc kubenswrapper[4700]: I0227 17:05:14.257723 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6cd66f6876-tgwlc" Feb 27 17:05:14 crc kubenswrapper[4700]: I0227 17:05:14.261792 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-685c5585dd-mknx2" Feb 27 17:05:14 crc kubenswrapper[4700]: I0227 17:05:14.284632 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-685c5585dd-mknx2" podStartSLOduration=3.284611464 podStartE2EDuration="3.284611464s" podCreationTimestamp="2026-02-27 17:05:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:05:14.282505478 +0000 UTC m=+274.267818235" watchObservedRunningTime="2026-02-27 17:05:14.284611464 +0000 UTC m=+274.269924211" Feb 27 17:05:14 crc kubenswrapper[4700]: I0227 17:05:14.327090 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6cd66f6876-tgwlc" podStartSLOduration=3.327070164 podStartE2EDuration="3.327070164s" podCreationTimestamp="2026-02-27 17:05:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:05:14.322387419 +0000 UTC m=+274.307700176" watchObservedRunningTime="2026-02-27 17:05:14.327070164 +0000 UTC m=+274.312382921" Feb 27 17:05:14 crc kubenswrapper[4700]: I0227 17:05:14.713688 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6cd66f6876-tgwlc" Feb 27 17:05:14 crc kubenswrapper[4700]: I0227 17:05:14.895173 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9t75w" Feb 27 17:05:14 crc kubenswrapper[4700]: I0227 17:05:14.895226 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9t75w" Feb 27 17:05:14 crc kubenswrapper[4700]: I0227 17:05:14.940317 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9t75w" Feb 27 17:05:15 crc kubenswrapper[4700]: I0227 17:05:15.175163 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5n6qh" Feb 27 17:05:15 crc kubenswrapper[4700]: I0227 17:05:15.327741 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9t75w" Feb 27 17:05:15 crc kubenswrapper[4700]: I0227 17:05:15.542647 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-j8wlz" Feb 27 17:05:15 crc kubenswrapper[4700]: I0227 17:05:15.542702 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-j8wlz" Feb 27 17:05:15 crc kubenswrapper[4700]: I0227 17:05:15.609208 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-j8wlz" Feb 27 17:05:16 crc kubenswrapper[4700]: I0227 17:05:16.353784 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-j8wlz" Feb 27 17:05:17 crc kubenswrapper[4700]: I0227 17:05:17.421904 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j8wlz"] Feb 27 17:05:18 crc kubenswrapper[4700]: I0227 17:05:18.288962 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-j8wlz" podUID="d1c12d33-6674-4c8a-b871-4d95efca7c23" containerName="registry-server" containerID="cri-o://2ca68297fff8db72acd0c00f90e1e44a57f3393fafa068967d78d6f8e74d540a" gracePeriod=2 Feb 27 17:05:18 crc kubenswrapper[4700]: I0227 17:05:18.821570 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j8wlz" Feb 27 17:05:18 crc kubenswrapper[4700]: I0227 17:05:18.913061 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d67j9\" (UniqueName: \"kubernetes.io/projected/d1c12d33-6674-4c8a-b871-4d95efca7c23-kube-api-access-d67j9\") pod \"d1c12d33-6674-4c8a-b871-4d95efca7c23\" (UID: \"d1c12d33-6674-4c8a-b871-4d95efca7c23\") " Feb 27 17:05:18 crc kubenswrapper[4700]: I0227 17:05:18.913159 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1c12d33-6674-4c8a-b871-4d95efca7c23-utilities\") pod \"d1c12d33-6674-4c8a-b871-4d95efca7c23\" (UID: \"d1c12d33-6674-4c8a-b871-4d95efca7c23\") " Feb 27 17:05:18 crc kubenswrapper[4700]: I0227 17:05:18.913203 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1c12d33-6674-4c8a-b871-4d95efca7c23-catalog-content\") pod \"d1c12d33-6674-4c8a-b871-4d95efca7c23\" (UID: \"d1c12d33-6674-4c8a-b871-4d95efca7c23\") " Feb 27 17:05:18 crc kubenswrapper[4700]: I0227 17:05:18.914916 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1c12d33-6674-4c8a-b871-4d95efca7c23-utilities" (OuterVolumeSpecName: "utilities") pod "d1c12d33-6674-4c8a-b871-4d95efca7c23" (UID: "d1c12d33-6674-4c8a-b871-4d95efca7c23"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:05:18 crc kubenswrapper[4700]: I0227 17:05:18.919854 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1c12d33-6674-4c8a-b871-4d95efca7c23-kube-api-access-d67j9" (OuterVolumeSpecName: "kube-api-access-d67j9") pod "d1c12d33-6674-4c8a-b871-4d95efca7c23" (UID: "d1c12d33-6674-4c8a-b871-4d95efca7c23"). InnerVolumeSpecName "kube-api-access-d67j9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:05:18 crc kubenswrapper[4700]: I0227 17:05:18.998174 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1c12d33-6674-4c8a-b871-4d95efca7c23-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d1c12d33-6674-4c8a-b871-4d95efca7c23" (UID: "d1c12d33-6674-4c8a-b871-4d95efca7c23"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:05:19 crc kubenswrapper[4700]: I0227 17:05:19.014718 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d67j9\" (UniqueName: \"kubernetes.io/projected/d1c12d33-6674-4c8a-b871-4d95efca7c23-kube-api-access-d67j9\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:19 crc kubenswrapper[4700]: I0227 17:05:19.014763 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1c12d33-6674-4c8a-b871-4d95efca7c23-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:19 crc kubenswrapper[4700]: I0227 17:05:19.014783 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1c12d33-6674-4c8a-b871-4d95efca7c23-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:19 crc kubenswrapper[4700]: I0227 17:05:19.300609 4700 generic.go:334] "Generic (PLEG): container finished" podID="d1c12d33-6674-4c8a-b871-4d95efca7c23" containerID="2ca68297fff8db72acd0c00f90e1e44a57f3393fafa068967d78d6f8e74d540a" exitCode=0 Feb 27 17:05:19 crc kubenswrapper[4700]: I0227 17:05:19.300688 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j8wlz" Feb 27 17:05:19 crc kubenswrapper[4700]: I0227 17:05:19.300698 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j8wlz" event={"ID":"d1c12d33-6674-4c8a-b871-4d95efca7c23","Type":"ContainerDied","Data":"2ca68297fff8db72acd0c00f90e1e44a57f3393fafa068967d78d6f8e74d540a"} Feb 27 17:05:19 crc kubenswrapper[4700]: I0227 17:05:19.301116 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j8wlz" event={"ID":"d1c12d33-6674-4c8a-b871-4d95efca7c23","Type":"ContainerDied","Data":"0940a658aa8c9b190adf4d90142fd96c7e440f4bdfaf73f142c5e54198590b5f"} Feb 27 17:05:19 crc kubenswrapper[4700]: I0227 17:05:19.301139 4700 scope.go:117] "RemoveContainer" containerID="2ca68297fff8db72acd0c00f90e1e44a57f3393fafa068967d78d6f8e74d540a" Feb 27 17:05:19 crc kubenswrapper[4700]: I0227 17:05:19.329688 4700 scope.go:117] "RemoveContainer" containerID="69caf0b406ec72e1bfe6e57a12df73436af7f4bfb3cf613a8d40f70fd813393e" Feb 27 17:05:19 crc kubenswrapper[4700]: I0227 17:05:19.362207 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j8wlz"] Feb 27 17:05:19 crc kubenswrapper[4700]: I0227 17:05:19.371282 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-j8wlz"] Feb 27 17:05:19 crc kubenswrapper[4700]: I0227 17:05:19.373762 4700 scope.go:117] "RemoveContainer" containerID="8f69a80ab7dcdcb9827cb4ec7403a9a897071125d1ea02b6f73a714cc2d75bec" Feb 27 17:05:19 crc kubenswrapper[4700]: I0227 17:05:19.401235 4700 scope.go:117] "RemoveContainer" containerID="2ca68297fff8db72acd0c00f90e1e44a57f3393fafa068967d78d6f8e74d540a" Feb 27 17:05:19 crc kubenswrapper[4700]: E0227 17:05:19.401778 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ca68297fff8db72acd0c00f90e1e44a57f3393fafa068967d78d6f8e74d540a\": container with ID starting with 2ca68297fff8db72acd0c00f90e1e44a57f3393fafa068967d78d6f8e74d540a not found: ID does not exist" containerID="2ca68297fff8db72acd0c00f90e1e44a57f3393fafa068967d78d6f8e74d540a" Feb 27 17:05:19 crc kubenswrapper[4700]: I0227 17:05:19.401845 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ca68297fff8db72acd0c00f90e1e44a57f3393fafa068967d78d6f8e74d540a"} err="failed to get container status \"2ca68297fff8db72acd0c00f90e1e44a57f3393fafa068967d78d6f8e74d540a\": rpc error: code = NotFound desc = could not find container \"2ca68297fff8db72acd0c00f90e1e44a57f3393fafa068967d78d6f8e74d540a\": container with ID starting with 2ca68297fff8db72acd0c00f90e1e44a57f3393fafa068967d78d6f8e74d540a not found: ID does not exist" Feb 27 17:05:19 crc kubenswrapper[4700]: I0227 17:05:19.401887 4700 scope.go:117] "RemoveContainer" containerID="69caf0b406ec72e1bfe6e57a12df73436af7f4bfb3cf613a8d40f70fd813393e" Feb 27 17:05:19 crc kubenswrapper[4700]: E0227 17:05:19.402406 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69caf0b406ec72e1bfe6e57a12df73436af7f4bfb3cf613a8d40f70fd813393e\": container with ID starting with 69caf0b406ec72e1bfe6e57a12df73436af7f4bfb3cf613a8d40f70fd813393e not found: ID does not exist" containerID="69caf0b406ec72e1bfe6e57a12df73436af7f4bfb3cf613a8d40f70fd813393e" Feb 27 17:05:19 crc kubenswrapper[4700]: I0227 17:05:19.402443 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69caf0b406ec72e1bfe6e57a12df73436af7f4bfb3cf613a8d40f70fd813393e"} err="failed to get container status \"69caf0b406ec72e1bfe6e57a12df73436af7f4bfb3cf613a8d40f70fd813393e\": rpc error: code = NotFound desc = could not find container \"69caf0b406ec72e1bfe6e57a12df73436af7f4bfb3cf613a8d40f70fd813393e\": container with ID starting with 69caf0b406ec72e1bfe6e57a12df73436af7f4bfb3cf613a8d40f70fd813393e not found: ID does not exist" Feb 27 17:05:19 crc kubenswrapper[4700]: I0227 17:05:19.402547 4700 scope.go:117] "RemoveContainer" containerID="8f69a80ab7dcdcb9827cb4ec7403a9a897071125d1ea02b6f73a714cc2d75bec" Feb 27 17:05:19 crc kubenswrapper[4700]: E0227 17:05:19.403056 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f69a80ab7dcdcb9827cb4ec7403a9a897071125d1ea02b6f73a714cc2d75bec\": container with ID starting with 8f69a80ab7dcdcb9827cb4ec7403a9a897071125d1ea02b6f73a714cc2d75bec not found: ID does not exist" containerID="8f69a80ab7dcdcb9827cb4ec7403a9a897071125d1ea02b6f73a714cc2d75bec" Feb 27 17:05:19 crc kubenswrapper[4700]: I0227 17:05:19.403106 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f69a80ab7dcdcb9827cb4ec7403a9a897071125d1ea02b6f73a714cc2d75bec"} err="failed to get container status \"8f69a80ab7dcdcb9827cb4ec7403a9a897071125d1ea02b6f73a714cc2d75bec\": rpc error: code = NotFound desc = could not find container \"8f69a80ab7dcdcb9827cb4ec7403a9a897071125d1ea02b6f73a714cc2d75bec\": container with ID starting with 8f69a80ab7dcdcb9827cb4ec7403a9a897071125d1ea02b6f73a714cc2d75bec not found: ID does not exist" Feb 27 17:05:21 crc kubenswrapper[4700]: I0227 17:05:21.011026 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1c12d33-6674-4c8a-b871-4d95efca7c23" path="/var/lib/kubelet/pods/d1c12d33-6674-4c8a-b871-4d95efca7c23/volumes" Feb 27 17:05:21 crc kubenswrapper[4700]: I0227 17:05:21.591059 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" podUID="649e4fdc-b0b6-4231-8038-d4fa805db298" containerName="oauth-openshift" containerID="cri-o://67075e64d69ab0d1b77c77d9b6745e62b45b9a116ec8ae12ea0d941fa0990695" gracePeriod=15 Feb 27 17:05:22 crc kubenswrapper[4700]: I0227 17:05:22.323918 4700 generic.go:334] "Generic (PLEG): container finished" podID="649e4fdc-b0b6-4231-8038-d4fa805db298" containerID="67075e64d69ab0d1b77c77d9b6745e62b45b9a116ec8ae12ea0d941fa0990695" exitCode=0 Feb 27 17:05:22 crc kubenswrapper[4700]: I0227 17:05:22.323973 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" event={"ID":"649e4fdc-b0b6-4231-8038-d4fa805db298","Type":"ContainerDied","Data":"67075e64d69ab0d1b77c77d9b6745e62b45b9a116ec8ae12ea0d941fa0990695"} Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.191577 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.275895 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-service-ca\") pod \"649e4fdc-b0b6-4231-8038-d4fa805db298\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.275962 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-user-template-error\") pod \"649e4fdc-b0b6-4231-8038-d4fa805db298\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.275995 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-user-template-login\") pod \"649e4fdc-b0b6-4231-8038-d4fa805db298\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.276024 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-user-idp-0-file-data\") pod \"649e4fdc-b0b6-4231-8038-d4fa805db298\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.276055 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-ocp-branding-template\") pod \"649e4fdc-b0b6-4231-8038-d4fa805db298\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.276083 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-trusted-ca-bundle\") pod \"649e4fdc-b0b6-4231-8038-d4fa805db298\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.276109 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-user-template-provider-selection\") pod \"649e4fdc-b0b6-4231-8038-d4fa805db298\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.276145 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-session\") pod \"649e4fdc-b0b6-4231-8038-d4fa805db298\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.276173 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-router-certs\") pod \"649e4fdc-b0b6-4231-8038-d4fa805db298\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.276206 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-cliconfig\") pod \"649e4fdc-b0b6-4231-8038-d4fa805db298\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.276236 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/649e4fdc-b0b6-4231-8038-d4fa805db298-audit-dir\") pod \"649e4fdc-b0b6-4231-8038-d4fa805db298\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.276272 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-serving-cert\") pod \"649e4fdc-b0b6-4231-8038-d4fa805db298\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.276305 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-db9mr\" (UniqueName: \"kubernetes.io/projected/649e4fdc-b0b6-4231-8038-d4fa805db298-kube-api-access-db9mr\") pod \"649e4fdc-b0b6-4231-8038-d4fa805db298\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.276339 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/649e4fdc-b0b6-4231-8038-d4fa805db298-audit-policies\") pod \"649e4fdc-b0b6-4231-8038-d4fa805db298\" (UID: \"649e4fdc-b0b6-4231-8038-d4fa805db298\") " Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.276729 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/649e4fdc-b0b6-4231-8038-d4fa805db298-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "649e4fdc-b0b6-4231-8038-d4fa805db298" (UID: "649e4fdc-b0b6-4231-8038-d4fa805db298"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.276943 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/649e4fdc-b0b6-4231-8038-d4fa805db298-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "649e4fdc-b0b6-4231-8038-d4fa805db298" (UID: "649e4fdc-b0b6-4231-8038-d4fa805db298"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.277022 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "649e4fdc-b0b6-4231-8038-d4fa805db298" (UID: "649e4fdc-b0b6-4231-8038-d4fa805db298"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.278147 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "649e4fdc-b0b6-4231-8038-d4fa805db298" (UID: "649e4fdc-b0b6-4231-8038-d4fa805db298"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.278679 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "649e4fdc-b0b6-4231-8038-d4fa805db298" (UID: "649e4fdc-b0b6-4231-8038-d4fa805db298"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.282731 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "649e4fdc-b0b6-4231-8038-d4fa805db298" (UID: "649e4fdc-b0b6-4231-8038-d4fa805db298"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.282899 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/649e4fdc-b0b6-4231-8038-d4fa805db298-kube-api-access-db9mr" (OuterVolumeSpecName: "kube-api-access-db9mr") pod "649e4fdc-b0b6-4231-8038-d4fa805db298" (UID: "649e4fdc-b0b6-4231-8038-d4fa805db298"). InnerVolumeSpecName "kube-api-access-db9mr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.283196 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "649e4fdc-b0b6-4231-8038-d4fa805db298" (UID: "649e4fdc-b0b6-4231-8038-d4fa805db298"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.283560 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "649e4fdc-b0b6-4231-8038-d4fa805db298" (UID: "649e4fdc-b0b6-4231-8038-d4fa805db298"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.284265 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "649e4fdc-b0b6-4231-8038-d4fa805db298" (UID: "649e4fdc-b0b6-4231-8038-d4fa805db298"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.284451 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "649e4fdc-b0b6-4231-8038-d4fa805db298" (UID: "649e4fdc-b0b6-4231-8038-d4fa805db298"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.284906 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "649e4fdc-b0b6-4231-8038-d4fa805db298" (UID: "649e4fdc-b0b6-4231-8038-d4fa805db298"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.285119 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "649e4fdc-b0b6-4231-8038-d4fa805db298" (UID: "649e4fdc-b0b6-4231-8038-d4fa805db298"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.289364 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "649e4fdc-b0b6-4231-8038-d4fa805db298" (UID: "649e4fdc-b0b6-4231-8038-d4fa805db298"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.333420 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" event={"ID":"649e4fdc-b0b6-4231-8038-d4fa805db298","Type":"ContainerDied","Data":"4ebea6a0755943d73b4d40b4eab2dcb763761f099ed234a8e8891c07dbeb3bd9"} Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.333523 4700 scope.go:117] "RemoveContainer" containerID="67075e64d69ab0d1b77c77d9b6745e62b45b9a116ec8ae12ea0d941fa0990695" Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.333533 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-zxt8n" Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.378325 4700 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/649e4fdc-b0b6-4231-8038-d4fa805db298-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.378380 4700 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.378403 4700 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.378425 4700 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.378503 4700 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.378522 4700 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.378540 4700 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.378559 4700 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.378581 4700 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.378601 4700 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.378618 4700 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.378635 4700 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/649e4fdc-b0b6-4231-8038-d4fa805db298-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.378654 4700 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/649e4fdc-b0b6-4231-8038-d4fa805db298-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.378673 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-db9mr\" (UniqueName: \"kubernetes.io/projected/649e4fdc-b0b6-4231-8038-d4fa805db298-kube-api-access-db9mr\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.381640 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-zxt8n"] Feb 27 17:05:23 crc kubenswrapper[4700]: I0227 17:05:23.386077 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-zxt8n"] Feb 27 17:05:24 crc kubenswrapper[4700]: I0227 17:05:24.988906 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="649e4fdc-b0b6-4231-8038-d4fa805db298" path="/var/lib/kubelet/pods/649e4fdc-b0b6-4231-8038-d4fa805db298/volumes" Feb 27 17:05:31 crc kubenswrapper[4700]: I0227 17:05:31.292166 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-685c5585dd-mknx2"] Feb 27 17:05:31 crc kubenswrapper[4700]: I0227 17:05:31.294114 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-685c5585dd-mknx2" podUID="59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03" containerName="controller-manager" containerID="cri-o://0fbc4956eeca2f4ef308af038a9f4cd9a99b6cfc1d22b3b30d4e0bb6e708d687" gracePeriod=30 Feb 27 17:05:31 crc kubenswrapper[4700]: I0227 17:05:31.387701 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6cd66f6876-tgwlc"] Feb 27 17:05:31 crc kubenswrapper[4700]: I0227 17:05:31.387989 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6cd66f6876-tgwlc" podUID="8f1cb2d6-8127-4a75-bdbd-751f69ef0613" containerName="route-controller-manager" containerID="cri-o://19b111c34447df31b93ccabe652c10ac18aac74aacbd74268c519c2e1590d2ff" gracePeriod=30 Feb 27 17:05:31 crc kubenswrapper[4700]: I0227 17:05:31.886396 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6cd66f6876-tgwlc" Feb 27 17:05:31 crc kubenswrapper[4700]: I0227 17:05:31.891209 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-685c5585dd-mknx2" Feb 27 17:05:31 crc kubenswrapper[4700]: I0227 17:05:31.998436 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8f1cb2d6-8127-4a75-bdbd-751f69ef0613-client-ca\") pod \"8f1cb2d6-8127-4a75-bdbd-751f69ef0613\" (UID: \"8f1cb2d6-8127-4a75-bdbd-751f69ef0613\") " Feb 27 17:05:31 crc kubenswrapper[4700]: I0227 17:05:31.998556 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03-config\") pod \"59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03\" (UID: \"59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03\") " Feb 27 17:05:31 crc kubenswrapper[4700]: I0227 17:05:31.998590 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03-serving-cert\") pod \"59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03\" (UID: \"59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03\") " Feb 27 17:05:31 crc kubenswrapper[4700]: I0227 17:05:31.998614 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xc5vp\" (UniqueName: \"kubernetes.io/projected/8f1cb2d6-8127-4a75-bdbd-751f69ef0613-kube-api-access-xc5vp\") pod \"8f1cb2d6-8127-4a75-bdbd-751f69ef0613\" (UID: \"8f1cb2d6-8127-4a75-bdbd-751f69ef0613\") " Feb 27 17:05:31 crc kubenswrapper[4700]: I0227 17:05:31.998664 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f1cb2d6-8127-4a75-bdbd-751f69ef0613-config\") pod \"8f1cb2d6-8127-4a75-bdbd-751f69ef0613\" (UID: \"8f1cb2d6-8127-4a75-bdbd-751f69ef0613\") " Feb 27 17:05:31 crc kubenswrapper[4700]: I0227 17:05:31.998699 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03-proxy-ca-bundles\") pod \"59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03\" (UID: \"59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03\") " Feb 27 17:05:31 crc kubenswrapper[4700]: I0227 17:05:31.998723 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f1cb2d6-8127-4a75-bdbd-751f69ef0613-serving-cert\") pod \"8f1cb2d6-8127-4a75-bdbd-751f69ef0613\" (UID: \"8f1cb2d6-8127-4a75-bdbd-751f69ef0613\") " Feb 27 17:05:31 crc kubenswrapper[4700]: I0227 17:05:31.998744 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03-client-ca\") pod \"59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03\" (UID: \"59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03\") " Feb 27 17:05:31 crc kubenswrapper[4700]: I0227 17:05:31.998777 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqlxq\" (UniqueName: \"kubernetes.io/projected/59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03-kube-api-access-qqlxq\") pod \"59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03\" (UID: \"59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03\") " Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:31.999848 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f1cb2d6-8127-4a75-bdbd-751f69ef0613-config" (OuterVolumeSpecName: "config") pod "8f1cb2d6-8127-4a75-bdbd-751f69ef0613" (UID: "8f1cb2d6-8127-4a75-bdbd-751f69ef0613"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:31.999899 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03-config" (OuterVolumeSpecName: "config") pod "59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03" (UID: "59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:31.999927 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03-client-ca" (OuterVolumeSpecName: "client-ca") pod "59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03" (UID: "59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.000408 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03" (UID: "59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.001314 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f1cb2d6-8127-4a75-bdbd-751f69ef0613-client-ca" (OuterVolumeSpecName: "client-ca") pod "8f1cb2d6-8127-4a75-bdbd-751f69ef0613" (UID: "8f1cb2d6-8127-4a75-bdbd-751f69ef0613"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.005061 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03" (UID: "59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.005824 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f1cb2d6-8127-4a75-bdbd-751f69ef0613-kube-api-access-xc5vp" (OuterVolumeSpecName: "kube-api-access-xc5vp") pod "8f1cb2d6-8127-4a75-bdbd-751f69ef0613" (UID: "8f1cb2d6-8127-4a75-bdbd-751f69ef0613"). InnerVolumeSpecName "kube-api-access-xc5vp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.009573 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f1cb2d6-8127-4a75-bdbd-751f69ef0613-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8f1cb2d6-8127-4a75-bdbd-751f69ef0613" (UID: "8f1cb2d6-8127-4a75-bdbd-751f69ef0613"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.009676 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03-kube-api-access-qqlxq" (OuterVolumeSpecName: "kube-api-access-qqlxq") pod "59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03" (UID: "59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03"). InnerVolumeSpecName "kube-api-access-qqlxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.100322 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xc5vp\" (UniqueName: \"kubernetes.io/projected/8f1cb2d6-8127-4a75-bdbd-751f69ef0613-kube-api-access-xc5vp\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.100945 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f1cb2d6-8127-4a75-bdbd-751f69ef0613-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.101288 4700 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.101538 4700 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f1cb2d6-8127-4a75-bdbd-751f69ef0613-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.101836 4700 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.102691 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqlxq\" (UniqueName: \"kubernetes.io/projected/59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03-kube-api-access-qqlxq\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.102859 4700 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8f1cb2d6-8127-4a75-bdbd-751f69ef0613-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.103055 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.103228 4700 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.399550 4700 generic.go:334] "Generic (PLEG): container finished" podID="59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03" containerID="0fbc4956eeca2f4ef308af038a9f4cd9a99b6cfc1d22b3b30d4e0bb6e708d687" exitCode=0 Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.399652 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-685c5585dd-mknx2" event={"ID":"59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03","Type":"ContainerDied","Data":"0fbc4956eeca2f4ef308af038a9f4cd9a99b6cfc1d22b3b30d4e0bb6e708d687"} Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.399741 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-685c5585dd-mknx2" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.399837 4700 scope.go:117] "RemoveContainer" containerID="0fbc4956eeca2f4ef308af038a9f4cd9a99b6cfc1d22b3b30d4e0bb6e708d687" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.399780 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-685c5585dd-mknx2" event={"ID":"59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03","Type":"ContainerDied","Data":"021fbacddef86943aa7b150a0a781e7b94a82d098f95e4fa80e06f3ee47a5e9b"} Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.402040 4700 generic.go:334] "Generic (PLEG): container finished" podID="8f1cb2d6-8127-4a75-bdbd-751f69ef0613" containerID="19b111c34447df31b93ccabe652c10ac18aac74aacbd74268c519c2e1590d2ff" exitCode=0 Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.402087 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6cd66f6876-tgwlc" event={"ID":"8f1cb2d6-8127-4a75-bdbd-751f69ef0613","Type":"ContainerDied","Data":"19b111c34447df31b93ccabe652c10ac18aac74aacbd74268c519c2e1590d2ff"} Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.402118 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6cd66f6876-tgwlc" event={"ID":"8f1cb2d6-8127-4a75-bdbd-751f69ef0613","Type":"ContainerDied","Data":"992e550f31f30277b3b90903ade030f8c733504eca2e7d903f6b31a129a6bec0"} Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.402190 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6cd66f6876-tgwlc" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.449531 4700 scope.go:117] "RemoveContainer" containerID="0fbc4956eeca2f4ef308af038a9f4cd9a99b6cfc1d22b3b30d4e0bb6e708d687" Feb 27 17:05:32 crc kubenswrapper[4700]: E0227 17:05:32.451239 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fbc4956eeca2f4ef308af038a9f4cd9a99b6cfc1d22b3b30d4e0bb6e708d687\": container with ID starting with 0fbc4956eeca2f4ef308af038a9f4cd9a99b6cfc1d22b3b30d4e0bb6e708d687 not found: ID does not exist" containerID="0fbc4956eeca2f4ef308af038a9f4cd9a99b6cfc1d22b3b30d4e0bb6e708d687" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.451303 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fbc4956eeca2f4ef308af038a9f4cd9a99b6cfc1d22b3b30d4e0bb6e708d687"} err="failed to get container status \"0fbc4956eeca2f4ef308af038a9f4cd9a99b6cfc1d22b3b30d4e0bb6e708d687\": rpc error: code = NotFound desc = could not find container \"0fbc4956eeca2f4ef308af038a9f4cd9a99b6cfc1d22b3b30d4e0bb6e708d687\": container with ID starting with 0fbc4956eeca2f4ef308af038a9f4cd9a99b6cfc1d22b3b30d4e0bb6e708d687 not found: ID does not exist" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.451354 4700 scope.go:117] "RemoveContainer" containerID="19b111c34447df31b93ccabe652c10ac18aac74aacbd74268c519c2e1590d2ff" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.468143 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-685c5585dd-mknx2"] Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.479135 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-685c5585dd-mknx2"] Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.482649 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6cd66f6876-tgwlc"] Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.485804 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6cd66f6876-tgwlc"] Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.487849 4700 scope.go:117] "RemoveContainer" containerID="19b111c34447df31b93ccabe652c10ac18aac74aacbd74268c519c2e1590d2ff" Feb 27 17:05:32 crc kubenswrapper[4700]: E0227 17:05:32.488646 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19b111c34447df31b93ccabe652c10ac18aac74aacbd74268c519c2e1590d2ff\": container with ID starting with 19b111c34447df31b93ccabe652c10ac18aac74aacbd74268c519c2e1590d2ff not found: ID does not exist" containerID="19b111c34447df31b93ccabe652c10ac18aac74aacbd74268c519c2e1590d2ff" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.488732 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19b111c34447df31b93ccabe652c10ac18aac74aacbd74268c519c2e1590d2ff"} err="failed to get container status \"19b111c34447df31b93ccabe652c10ac18aac74aacbd74268c519c2e1590d2ff\": rpc error: code = NotFound desc = could not find container \"19b111c34447df31b93ccabe652c10ac18aac74aacbd74268c519c2e1590d2ff\": container with ID starting with 19b111c34447df31b93ccabe652c10ac18aac74aacbd74268c519c2e1590d2ff not found: ID does not exist" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.768949 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7454867c54-fxpk6"] Feb 27 17:05:32 crc kubenswrapper[4700]: E0227 17:05:32.769344 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f1cb2d6-8127-4a75-bdbd-751f69ef0613" containerName="route-controller-manager" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.769371 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f1cb2d6-8127-4a75-bdbd-751f69ef0613" containerName="route-controller-manager" Feb 27 17:05:32 crc kubenswrapper[4700]: E0227 17:05:32.769391 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1c12d33-6674-4c8a-b871-4d95efca7c23" containerName="registry-server" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.769403 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1c12d33-6674-4c8a-b871-4d95efca7c23" containerName="registry-server" Feb 27 17:05:32 crc kubenswrapper[4700]: E0227 17:05:32.769425 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="649e4fdc-b0b6-4231-8038-d4fa805db298" containerName="oauth-openshift" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.769441 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="649e4fdc-b0b6-4231-8038-d4fa805db298" containerName="oauth-openshift" Feb 27 17:05:32 crc kubenswrapper[4700]: E0227 17:05:32.769476 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1c12d33-6674-4c8a-b871-4d95efca7c23" containerName="extract-utilities" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.769534 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1c12d33-6674-4c8a-b871-4d95efca7c23" containerName="extract-utilities" Feb 27 17:05:32 crc kubenswrapper[4700]: E0227 17:05:32.769551 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1c12d33-6674-4c8a-b871-4d95efca7c23" containerName="extract-content" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.769563 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1c12d33-6674-4c8a-b871-4d95efca7c23" containerName="extract-content" Feb 27 17:05:32 crc kubenswrapper[4700]: E0227 17:05:32.769581 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03" containerName="controller-manager" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.769595 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03" containerName="controller-manager" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.769810 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="649e4fdc-b0b6-4231-8038-d4fa805db298" containerName="oauth-openshift" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.769831 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f1cb2d6-8127-4a75-bdbd-751f69ef0613" containerName="route-controller-manager" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.769858 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1c12d33-6674-4c8a-b871-4d95efca7c23" containerName="registry-server" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.769877 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03" containerName="controller-manager" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.770609 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7454867c54-fxpk6" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.773239 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.773416 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.773851 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.773920 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.774691 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.779198 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs"] Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.780479 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.784016 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.785347 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.789190 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.796530 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.796998 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.797630 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf"] Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.798626 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.799105 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.799628 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.799899 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.800091 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.800368 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.800637 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.800840 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.801668 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.804183 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.805532 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.818790 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf"] Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.824181 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.825680 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.825949 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.826094 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.826319 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.826448 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.827991 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.832393 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7454867c54-fxpk6"] Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.837981 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.843644 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs"] Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.921375 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-v4-0-config-system-service-ca\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.921433 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7f4a2ab-8949-4dc5-b0cf-721714a85905-serving-cert\") pod \"controller-manager-7454867c54-fxpk6\" (UID: \"d7f4a2ab-8949-4dc5-b0cf-721714a85905\") " pod="openshift-controller-manager/controller-manager-7454867c54-fxpk6" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.921489 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-v4-0-config-system-session\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.921521 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7r67w\" (UniqueName: \"kubernetes.io/projected/21dd241c-061a-4e48-b039-f5877cdc8aba-kube-api-access-7r67w\") pod \"route-controller-manager-77b96b4458-bm8hf\" (UID: \"21dd241c-061a-4e48-b039-f5877cdc8aba\") " pod="openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.921541 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.921606 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.921632 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.921661 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.921758 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/21dd241c-061a-4e48-b039-f5877cdc8aba-client-ca\") pod \"route-controller-manager-77b96b4458-bm8hf\" (UID: \"21dd241c-061a-4e48-b039-f5877cdc8aba\") " pod="openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.921808 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-v4-0-config-system-router-certs\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.921835 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-v4-0-config-user-template-error\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.921884 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7f4a2ab-8949-4dc5-b0cf-721714a85905-config\") pod \"controller-manager-7454867c54-fxpk6\" (UID: \"d7f4a2ab-8949-4dc5-b0cf-721714a85905\") " pod="openshift-controller-manager/controller-manager-7454867c54-fxpk6" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.921964 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.922008 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2d4l\" (UniqueName: \"kubernetes.io/projected/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-kube-api-access-h2d4l\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.922066 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21dd241c-061a-4e48-b039-f5877cdc8aba-serving-cert\") pod \"route-controller-manager-77b96b4458-bm8hf\" (UID: \"21dd241c-061a-4e48-b039-f5877cdc8aba\") " pod="openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.922118 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-audit-policies\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.922171 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-audit-dir\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.922254 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d7f4a2ab-8949-4dc5-b0cf-721714a85905-proxy-ca-bundles\") pod \"controller-manager-7454867c54-fxpk6\" (UID: \"d7f4a2ab-8949-4dc5-b0cf-721714a85905\") " pod="openshift-controller-manager/controller-manager-7454867c54-fxpk6" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.922305 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-v4-0-config-user-template-login\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.922354 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgxcv\" (UniqueName: \"kubernetes.io/projected/d7f4a2ab-8949-4dc5-b0cf-721714a85905-kube-api-access-lgxcv\") pod \"controller-manager-7454867c54-fxpk6\" (UID: \"d7f4a2ab-8949-4dc5-b0cf-721714a85905\") " pod="openshift-controller-manager/controller-manager-7454867c54-fxpk6" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.922389 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21dd241c-061a-4e48-b039-f5877cdc8aba-config\") pod \"route-controller-manager-77b96b4458-bm8hf\" (UID: \"21dd241c-061a-4e48-b039-f5877cdc8aba\") " pod="openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.922431 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.922535 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d7f4a2ab-8949-4dc5-b0cf-721714a85905-client-ca\") pod \"controller-manager-7454867c54-fxpk6\" (UID: \"d7f4a2ab-8949-4dc5-b0cf-721714a85905\") " pod="openshift-controller-manager/controller-manager-7454867c54-fxpk6" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.993273 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03" path="/var/lib/kubelet/pods/59722ff3-7fcc-4b0b-bea3-4abf3a2d3a03/volumes" Feb 27 17:05:32 crc kubenswrapper[4700]: I0227 17:05:32.993952 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f1cb2d6-8127-4a75-bdbd-751f69ef0613" path="/var/lib/kubelet/pods/8f1cb2d6-8127-4a75-bdbd-751f69ef0613/volumes" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.023790 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d7f4a2ab-8949-4dc5-b0cf-721714a85905-client-ca\") pod \"controller-manager-7454867c54-fxpk6\" (UID: \"d7f4a2ab-8949-4dc5-b0cf-721714a85905\") " pod="openshift-controller-manager/controller-manager-7454867c54-fxpk6" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.023831 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-v4-0-config-system-service-ca\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.023866 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7f4a2ab-8949-4dc5-b0cf-721714a85905-serving-cert\") pod \"controller-manager-7454867c54-fxpk6\" (UID: \"d7f4a2ab-8949-4dc5-b0cf-721714a85905\") " pod="openshift-controller-manager/controller-manager-7454867c54-fxpk6" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.023890 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-v4-0-config-system-session\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.023922 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7r67w\" (UniqueName: \"kubernetes.io/projected/21dd241c-061a-4e48-b039-f5877cdc8aba-kube-api-access-7r67w\") pod \"route-controller-manager-77b96b4458-bm8hf\" (UID: \"21dd241c-061a-4e48-b039-f5877cdc8aba\") " pod="openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.023949 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.023987 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.024023 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.024061 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.024098 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/21dd241c-061a-4e48-b039-f5877cdc8aba-client-ca\") pod \"route-controller-manager-77b96b4458-bm8hf\" (UID: \"21dd241c-061a-4e48-b039-f5877cdc8aba\") " pod="openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.024125 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-v4-0-config-system-router-certs\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.024153 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-v4-0-config-user-template-error\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.024183 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7f4a2ab-8949-4dc5-b0cf-721714a85905-config\") pod \"controller-manager-7454867c54-fxpk6\" (UID: \"d7f4a2ab-8949-4dc5-b0cf-721714a85905\") " pod="openshift-controller-manager/controller-manager-7454867c54-fxpk6" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.024216 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.024247 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2d4l\" (UniqueName: \"kubernetes.io/projected/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-kube-api-access-h2d4l\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.024279 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21dd241c-061a-4e48-b039-f5877cdc8aba-serving-cert\") pod \"route-controller-manager-77b96b4458-bm8hf\" (UID: \"21dd241c-061a-4e48-b039-f5877cdc8aba\") " pod="openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.024308 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-audit-policies\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.024346 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-audit-dir\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.024383 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d7f4a2ab-8949-4dc5-b0cf-721714a85905-proxy-ca-bundles\") pod \"controller-manager-7454867c54-fxpk6\" (UID: \"d7f4a2ab-8949-4dc5-b0cf-721714a85905\") " pod="openshift-controller-manager/controller-manager-7454867c54-fxpk6" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.024415 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-v4-0-config-user-template-login\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.024448 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgxcv\" (UniqueName: \"kubernetes.io/projected/d7f4a2ab-8949-4dc5-b0cf-721714a85905-kube-api-access-lgxcv\") pod \"controller-manager-7454867c54-fxpk6\" (UID: \"d7f4a2ab-8949-4dc5-b0cf-721714a85905\") " pod="openshift-controller-manager/controller-manager-7454867c54-fxpk6" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.024509 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21dd241c-061a-4e48-b039-f5877cdc8aba-config\") pod \"route-controller-manager-77b96b4458-bm8hf\" (UID: \"21dd241c-061a-4e48-b039-f5877cdc8aba\") " pod="openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.024543 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.024852 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-v4-0-config-system-service-ca\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.024899 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-audit-dir\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.025322 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.025853 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.026402 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7f4a2ab-8949-4dc5-b0cf-721714a85905-config\") pod \"controller-manager-7454867c54-fxpk6\" (UID: \"d7f4a2ab-8949-4dc5-b0cf-721714a85905\") " pod="openshift-controller-manager/controller-manager-7454867c54-fxpk6" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.028724 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21dd241c-061a-4e48-b039-f5877cdc8aba-config\") pod \"route-controller-manager-77b96b4458-bm8hf\" (UID: \"21dd241c-061a-4e48-b039-f5877cdc8aba\") " pod="openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.029204 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d7f4a2ab-8949-4dc5-b0cf-721714a85905-proxy-ca-bundles\") pod \"controller-manager-7454867c54-fxpk6\" (UID: \"d7f4a2ab-8949-4dc5-b0cf-721714a85905\") " pod="openshift-controller-manager/controller-manager-7454867c54-fxpk6" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.030188 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-v4-0-config-system-session\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.030208 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.024866 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d7f4a2ab-8949-4dc5-b0cf-721714a85905-client-ca\") pod \"controller-manager-7454867c54-fxpk6\" (UID: \"d7f4a2ab-8949-4dc5-b0cf-721714a85905\") " pod="openshift-controller-manager/controller-manager-7454867c54-fxpk6" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.030342 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-audit-policies\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.030896 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7f4a2ab-8949-4dc5-b0cf-721714a85905-serving-cert\") pod \"controller-manager-7454867c54-fxpk6\" (UID: \"d7f4a2ab-8949-4dc5-b0cf-721714a85905\") " pod="openshift-controller-manager/controller-manager-7454867c54-fxpk6" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.030964 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/21dd241c-061a-4e48-b039-f5877cdc8aba-client-ca\") pod \"route-controller-manager-77b96b4458-bm8hf\" (UID: \"21dd241c-061a-4e48-b039-f5877cdc8aba\") " pod="openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.032704 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-v4-0-config-system-router-certs\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.033587 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-v4-0-config-user-template-login\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.033627 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.035407 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.036043 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-v4-0-config-user-template-error\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.041437 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.042295 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21dd241c-061a-4e48-b039-f5877cdc8aba-serving-cert\") pod \"route-controller-manager-77b96b4458-bm8hf\" (UID: \"21dd241c-061a-4e48-b039-f5877cdc8aba\") " pod="openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.048683 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgxcv\" (UniqueName: \"kubernetes.io/projected/d7f4a2ab-8949-4dc5-b0cf-721714a85905-kube-api-access-lgxcv\") pod \"controller-manager-7454867c54-fxpk6\" (UID: \"d7f4a2ab-8949-4dc5-b0cf-721714a85905\") " pod="openshift-controller-manager/controller-manager-7454867c54-fxpk6" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.055746 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7r67w\" (UniqueName: \"kubernetes.io/projected/21dd241c-061a-4e48-b039-f5877cdc8aba-kube-api-access-7r67w\") pod \"route-controller-manager-77b96b4458-bm8hf\" (UID: \"21dd241c-061a-4e48-b039-f5877cdc8aba\") " pod="openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.056031 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2d4l\" (UniqueName: \"kubernetes.io/projected/04191b5b-0e24-4490-a6aa-b9cda6b6c8d3-kube-api-access-h2d4l\") pod \"oauth-openshift-7d9bbcf4d4-6g7cs\" (UID: \"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.123346 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7454867c54-fxpk6" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.142208 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.155309 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf" Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.496550 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf"] Feb 27 17:05:33 crc kubenswrapper[4700]: W0227 17:05:33.502082 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod21dd241c_061a_4e48_b039_f5877cdc8aba.slice/crio-3f24f76b3287bb76381603fa0c0a5fa5c1a19272833c22bf1f554443923d2880 WatchSource:0}: Error finding container 3f24f76b3287bb76381603fa0c0a5fa5c1a19272833c22bf1f554443923d2880: Status 404 returned error can't find the container with id 3f24f76b3287bb76381603fa0c0a5fa5c1a19272833c22bf1f554443923d2880 Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.594826 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7454867c54-fxpk6"] Feb 27 17:05:33 crc kubenswrapper[4700]: I0227 17:05:33.604496 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs"] Feb 27 17:05:33 crc kubenswrapper[4700]: W0227 17:05:33.608061 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd7f4a2ab_8949_4dc5_b0cf_721714a85905.slice/crio-ed37c15352910827bcc499c9c4053182c72982530b6e6447765741f845b27ddd WatchSource:0}: Error finding container ed37c15352910827bcc499c9c4053182c72982530b6e6447765741f845b27ddd: Status 404 returned error can't find the container with id ed37c15352910827bcc499c9c4053182c72982530b6e6447765741f845b27ddd Feb 27 17:05:33 crc kubenswrapper[4700]: W0227 17:05:33.621845 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod04191b5b_0e24_4490_a6aa_b9cda6b6c8d3.slice/crio-ff20ce38b4a627fd97870132ab1a2bd4363e8782342012871a17f88f2e049907 WatchSource:0}: Error finding container ff20ce38b4a627fd97870132ab1a2bd4363e8782342012871a17f88f2e049907: Status 404 returned error can't find the container with id ff20ce38b4a627fd97870132ab1a2bd4363e8782342012871a17f88f2e049907 Feb 27 17:05:34 crc kubenswrapper[4700]: I0227 17:05:34.421900 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7454867c54-fxpk6" event={"ID":"d7f4a2ab-8949-4dc5-b0cf-721714a85905","Type":"ContainerStarted","Data":"1d1d4c21b0036aa6c915bac9cf7cce1192b63e164883b2aa77e2a840347ab7a4"} Feb 27 17:05:34 crc kubenswrapper[4700]: I0227 17:05:34.422424 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7454867c54-fxpk6" Feb 27 17:05:34 crc kubenswrapper[4700]: I0227 17:05:34.422511 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7454867c54-fxpk6" event={"ID":"d7f4a2ab-8949-4dc5-b0cf-721714a85905","Type":"ContainerStarted","Data":"ed37c15352910827bcc499c9c4053182c72982530b6e6447765741f845b27ddd"} Feb 27 17:05:34 crc kubenswrapper[4700]: I0227 17:05:34.425611 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" event={"ID":"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3","Type":"ContainerStarted","Data":"1ed4c6e3d4795bee0392ec33e50b4ef2e632d06da5466937b6d7cf2707b096ae"} Feb 27 17:05:34 crc kubenswrapper[4700]: I0227 17:05:34.425657 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" event={"ID":"04191b5b-0e24-4490-a6aa-b9cda6b6c8d3","Type":"ContainerStarted","Data":"ff20ce38b4a627fd97870132ab1a2bd4363e8782342012871a17f88f2e049907"} Feb 27 17:05:34 crc kubenswrapper[4700]: I0227 17:05:34.425893 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:34 crc kubenswrapper[4700]: I0227 17:05:34.427812 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7454867c54-fxpk6" Feb 27 17:05:34 crc kubenswrapper[4700]: I0227 17:05:34.428723 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf" event={"ID":"21dd241c-061a-4e48-b039-f5877cdc8aba","Type":"ContainerStarted","Data":"57516f85c29d09c8565695064903ec1f31d6e6b6c6e8e8cf2ed636ca80fa1255"} Feb 27 17:05:34 crc kubenswrapper[4700]: I0227 17:05:34.428806 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf" event={"ID":"21dd241c-061a-4e48-b039-f5877cdc8aba","Type":"ContainerStarted","Data":"3f24f76b3287bb76381603fa0c0a5fa5c1a19272833c22bf1f554443923d2880"} Feb 27 17:05:34 crc kubenswrapper[4700]: I0227 17:05:34.428995 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf" Feb 27 17:05:34 crc kubenswrapper[4700]: I0227 17:05:34.436989 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf" Feb 27 17:05:34 crc kubenswrapper[4700]: I0227 17:05:34.446313 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7454867c54-fxpk6" podStartSLOduration=3.446279347 podStartE2EDuration="3.446279347s" podCreationTimestamp="2026-02-27 17:05:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:05:34.442586059 +0000 UTC m=+294.427898826" watchObservedRunningTime="2026-02-27 17:05:34.446279347 +0000 UTC m=+294.431592134" Feb 27 17:05:34 crc kubenswrapper[4700]: I0227 17:05:34.483882 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf" podStartSLOduration=3.483865207 podStartE2EDuration="3.483865207s" podCreationTimestamp="2026-02-27 17:05:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:05:34.482632214 +0000 UTC m=+294.467945051" watchObservedRunningTime="2026-02-27 17:05:34.483865207 +0000 UTC m=+294.469177954" Feb 27 17:05:34 crc kubenswrapper[4700]: I0227 17:05:34.515923 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" podStartSLOduration=38.51590795 podStartE2EDuration="38.51590795s" podCreationTimestamp="2026-02-27 17:04:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:05:34.51515071 +0000 UTC m=+294.500463477" watchObservedRunningTime="2026-02-27 17:05:34.51590795 +0000 UTC m=+294.501220697" Feb 27 17:05:35 crc kubenswrapper[4700]: I0227 17:05:35.042268 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.098792 4700 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.099652 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.101261 4700 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.101711 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb" gracePeriod=15 Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.101743 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951" gracePeriod=15 Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.101808 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35" gracePeriod=15 Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.101749 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://4d2d2c9065b008ff0936c19bcce2ce6a574580911a8188084c5f978e08177b69" gracePeriod=15 Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.101909 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5" gracePeriod=15 Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.103029 4700 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 27 17:05:39 crc kubenswrapper[4700]: E0227 17:05:39.103318 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.103337 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 27 17:05:39 crc kubenswrapper[4700]: E0227 17:05:39.103357 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.103368 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 17:05:39 crc kubenswrapper[4700]: E0227 17:05:39.103382 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.103392 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 27 17:05:39 crc kubenswrapper[4700]: E0227 17:05:39.103406 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.103417 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 17:05:39 crc kubenswrapper[4700]: E0227 17:05:39.103430 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.103440 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 17:05:39 crc kubenswrapper[4700]: E0227 17:05:39.103453 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.103490 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 27 17:05:39 crc kubenswrapper[4700]: E0227 17:05:39.103510 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.103521 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 27 17:05:39 crc kubenswrapper[4700]: E0227 17:05:39.103534 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.103543 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.103717 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.103740 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.103756 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.103769 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.103782 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.103796 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.103810 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.103822 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.103838 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 27 17:05:39 crc kubenswrapper[4700]: E0227 17:05:39.104014 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.104029 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 17:05:39 crc kubenswrapper[4700]: E0227 17:05:39.104045 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.104055 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.153629 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.217994 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.218078 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.218113 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.218145 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.218230 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.218328 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.218379 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.218501 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.320025 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.320367 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.320284 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.320419 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.320514 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.320525 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.320921 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.320988 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.321016 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.321043 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.321071 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.321232 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.321252 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.321305 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.321331 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.321312 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.452174 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.465695 4700 generic.go:334] "Generic (PLEG): container finished" podID="eb3329cc-cc82-4f3c-87c0-b2baf19069eb" containerID="1383f6808d490ad6f1816be8c0d30fbe36237e84bd01f85dc894dc2ad1e12db7" exitCode=0 Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.465762 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"eb3329cc-cc82-4f3c-87c0-b2baf19069eb","Type":"ContainerDied","Data":"1383f6808d490ad6f1816be8c0d30fbe36237e84bd01f85dc894dc2ad1e12db7"} Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.466566 4700 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.467359 4700 status_manager.go:851] "Failed to get status for pod" podUID="eb3329cc-cc82-4f3c-87c0-b2baf19069eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.467906 4700 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.469050 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.471669 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.473428 4700 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4d2d2c9065b008ff0936c19bcce2ce6a574580911a8188084c5f978e08177b69" exitCode=0 Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.473540 4700 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35" exitCode=0 Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.473559 4700 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951" exitCode=0 Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.473570 4700 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5" exitCode=2 Feb 27 17:05:39 crc kubenswrapper[4700]: I0227 17:05:39.473627 4700 scope.go:117] "RemoveContainer" containerID="26edb85b456d4c81a864dcb412aeb000cd5d71c8fd2cf586dc3a2d395adf2a4d" Feb 27 17:05:39 crc kubenswrapper[4700]: E0227 17:05:39.491365 4700 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.222:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.1898295e900d8223 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:05:39.490759203 +0000 UTC m=+299.476071960,LastTimestamp:2026-02-27 17:05:39.490759203 +0000 UTC m=+299.476071960,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:05:40 crc kubenswrapper[4700]: I0227 17:05:40.304704 4700 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" start-of-body= Feb 27 17:05:40 crc kubenswrapper[4700]: I0227 17:05:40.305117 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" Feb 27 17:05:40 crc kubenswrapper[4700]: I0227 17:05:40.484796 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"7ef0dfdaac76cf9ec1a86a1b96773b7ccc0a4488abafebed86587f514efab52b"} Feb 27 17:05:40 crc kubenswrapper[4700]: I0227 17:05:40.485207 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"d7786374b14e70245fb82160cb77c68c0b9fe5c7cf7596839ed5d6f7889bbef1"} Feb 27 17:05:40 crc kubenswrapper[4700]: I0227 17:05:40.485761 4700 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:05:40 crc kubenswrapper[4700]: I0227 17:05:40.486227 4700 status_manager.go:851] "Failed to get status for pod" podUID="eb3329cc-cc82-4f3c-87c0-b2baf19069eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:05:40 crc kubenswrapper[4700]: I0227 17:05:40.486722 4700 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:05:40 crc kubenswrapper[4700]: I0227 17:05:40.489763 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 27 17:05:40 crc kubenswrapper[4700]: I0227 17:05:40.935350 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 27 17:05:40 crc kubenswrapper[4700]: I0227 17:05:40.935962 4700 status_manager.go:851] "Failed to get status for pod" podUID="eb3329cc-cc82-4f3c-87c0-b2baf19069eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:05:40 crc kubenswrapper[4700]: I0227 17:05:40.936380 4700 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:05:40 crc kubenswrapper[4700]: I0227 17:05:40.936961 4700 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:05:40 crc kubenswrapper[4700]: I0227 17:05:40.982502 4700 status_manager.go:851] "Failed to get status for pod" podUID="eb3329cc-cc82-4f3c-87c0-b2baf19069eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:05:40 crc kubenswrapper[4700]: I0227 17:05:40.982677 4700 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:05:40 crc kubenswrapper[4700]: I0227 17:05:40.982814 4700 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.060749 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eb3329cc-cc82-4f3c-87c0-b2baf19069eb-kube-api-access\") pod \"eb3329cc-cc82-4f3c-87c0-b2baf19069eb\" (UID: \"eb3329cc-cc82-4f3c-87c0-b2baf19069eb\") " Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.060868 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/eb3329cc-cc82-4f3c-87c0-b2baf19069eb-var-lock\") pod \"eb3329cc-cc82-4f3c-87c0-b2baf19069eb\" (UID: \"eb3329cc-cc82-4f3c-87c0-b2baf19069eb\") " Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.060965 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eb3329cc-cc82-4f3c-87c0-b2baf19069eb-var-lock" (OuterVolumeSpecName: "var-lock") pod "eb3329cc-cc82-4f3c-87c0-b2baf19069eb" (UID: "eb3329cc-cc82-4f3c-87c0-b2baf19069eb"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.061038 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/eb3329cc-cc82-4f3c-87c0-b2baf19069eb-kubelet-dir\") pod \"eb3329cc-cc82-4f3c-87c0-b2baf19069eb\" (UID: \"eb3329cc-cc82-4f3c-87c0-b2baf19069eb\") " Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.061111 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eb3329cc-cc82-4f3c-87c0-b2baf19069eb-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "eb3329cc-cc82-4f3c-87c0-b2baf19069eb" (UID: "eb3329cc-cc82-4f3c-87c0-b2baf19069eb"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.063936 4700 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/eb3329cc-cc82-4f3c-87c0-b2baf19069eb-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.063993 4700 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/eb3329cc-cc82-4f3c-87c0-b2baf19069eb-var-lock\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.071756 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb3329cc-cc82-4f3c-87c0-b2baf19069eb-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "eb3329cc-cc82-4f3c-87c0-b2baf19069eb" (UID: "eb3329cc-cc82-4f3c-87c0-b2baf19069eb"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.164767 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eb3329cc-cc82-4f3c-87c0-b2baf19069eb-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.479836 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.481406 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.482127 4700 status_manager.go:851] "Failed to get status for pod" podUID="eb3329cc-cc82-4f3c-87c0-b2baf19069eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.482904 4700 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.483245 4700 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.500200 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"eb3329cc-cc82-4f3c-87c0-b2baf19069eb","Type":"ContainerDied","Data":"5656a62fdcaaeb77a82f27f4b3b1314e3d0390dc9677c0c171d54edcbbf4444f"} Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.500240 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5656a62fdcaaeb77a82f27f4b3b1314e3d0390dc9677c0c171d54edcbbf4444f" Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.500274 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.503340 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.504103 4700 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb" exitCode=0 Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.504976 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.505202 4700 scope.go:117] "RemoveContainer" containerID="4d2d2c9065b008ff0936c19bcce2ce6a574580911a8188084c5f978e08177b69" Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.523853 4700 status_manager.go:851] "Failed to get status for pod" podUID="eb3329cc-cc82-4f3c-87c0-b2baf19069eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.524737 4700 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.525289 4700 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.528569 4700 scope.go:117] "RemoveContainer" containerID="f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35" Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.547648 4700 scope.go:117] "RemoveContainer" containerID="d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951" Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.566732 4700 scope.go:117] "RemoveContainer" containerID="d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5" Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.573400 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.573459 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.573486 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.573517 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.573583 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.573740 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.573865 4700 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.573901 4700 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.573910 4700 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.586221 4700 scope.go:117] "RemoveContainer" containerID="9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb" Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.605532 4700 scope.go:117] "RemoveContainer" containerID="3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c" Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.818056 4700 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.818226 4700 status_manager.go:851] "Failed to get status for pod" podUID="eb3329cc-cc82-4f3c-87c0-b2baf19069eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:05:41 crc kubenswrapper[4700]: I0227 17:05:41.818363 4700 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:05:42 crc kubenswrapper[4700]: E0227 17:05:42.077210 4700 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.222:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.1898295e900d8223 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:05:39.490759203 +0000 UTC m=+299.476071960,LastTimestamp:2026-02-27 17:05:39.490759203 +0000 UTC m=+299.476071960,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:05:42 crc kubenswrapper[4700]: I0227 17:05:42.220155 4700 scope.go:117] "RemoveContainer" containerID="4d2d2c9065b008ff0936c19bcce2ce6a574580911a8188084c5f978e08177b69" Feb 27 17:05:42 crc kubenswrapper[4700]: E0227 17:05:42.221041 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d2d2c9065b008ff0936c19bcce2ce6a574580911a8188084c5f978e08177b69\": container with ID starting with 4d2d2c9065b008ff0936c19bcce2ce6a574580911a8188084c5f978e08177b69 not found: ID does not exist" containerID="4d2d2c9065b008ff0936c19bcce2ce6a574580911a8188084c5f978e08177b69" Feb 27 17:05:42 crc kubenswrapper[4700]: I0227 17:05:42.221104 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d2d2c9065b008ff0936c19bcce2ce6a574580911a8188084c5f978e08177b69"} err="failed to get container status \"4d2d2c9065b008ff0936c19bcce2ce6a574580911a8188084c5f978e08177b69\": rpc error: code = NotFound desc = could not find container \"4d2d2c9065b008ff0936c19bcce2ce6a574580911a8188084c5f978e08177b69\": container with ID starting with 4d2d2c9065b008ff0936c19bcce2ce6a574580911a8188084c5f978e08177b69 not found: ID does not exist" Feb 27 17:05:42 crc kubenswrapper[4700]: I0227 17:05:42.221222 4700 scope.go:117] "RemoveContainer" containerID="f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35" Feb 27 17:05:42 crc kubenswrapper[4700]: E0227 17:05:42.222083 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35\": container with ID starting with f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35 not found: ID does not exist" containerID="f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35" Feb 27 17:05:42 crc kubenswrapper[4700]: I0227 17:05:42.222202 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35"} err="failed to get container status \"f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35\": rpc error: code = NotFound desc = could not find container \"f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35\": container with ID starting with f2a5a1c15dd1d6ba9b7fb6e3a4a397c04ba63cd92dc789b6f4eebf0cb01f4f35 not found: ID does not exist" Feb 27 17:05:42 crc kubenswrapper[4700]: I0227 17:05:42.222256 4700 scope.go:117] "RemoveContainer" containerID="d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951" Feb 27 17:05:42 crc kubenswrapper[4700]: E0227 17:05:42.224162 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951\": container with ID starting with d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951 not found: ID does not exist" containerID="d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951" Feb 27 17:05:42 crc kubenswrapper[4700]: I0227 17:05:42.224218 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951"} err="failed to get container status \"d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951\": rpc error: code = NotFound desc = could not find container \"d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951\": container with ID starting with d02f6cf454d123fa44168a74d295f0874ed991ed63294b007aff9056c0c3a951 not found: ID does not exist" Feb 27 17:05:42 crc kubenswrapper[4700]: I0227 17:05:42.224937 4700 scope.go:117] "RemoveContainer" containerID="d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5" Feb 27 17:05:42 crc kubenswrapper[4700]: E0227 17:05:42.225905 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5\": container with ID starting with d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5 not found: ID does not exist" containerID="d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5" Feb 27 17:05:42 crc kubenswrapper[4700]: I0227 17:05:42.225968 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5"} err="failed to get container status \"d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5\": rpc error: code = NotFound desc = could not find container \"d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5\": container with ID starting with d8ee74e414928706d7febeada99aea6a73c3b8014bfacc305bfc46f1a9306cf5 not found: ID does not exist" Feb 27 17:05:42 crc kubenswrapper[4700]: I0227 17:05:42.225997 4700 scope.go:117] "RemoveContainer" containerID="9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb" Feb 27 17:05:42 crc kubenswrapper[4700]: E0227 17:05:42.226885 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb\": container with ID starting with 9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb not found: ID does not exist" containerID="9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb" Feb 27 17:05:42 crc kubenswrapper[4700]: I0227 17:05:42.226930 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb"} err="failed to get container status \"9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb\": rpc error: code = NotFound desc = could not find container \"9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb\": container with ID starting with 9ed9172300d137684795b169a7d294baf48a982666024d0b369bc89a8fd898bb not found: ID does not exist" Feb 27 17:05:42 crc kubenswrapper[4700]: I0227 17:05:42.226959 4700 scope.go:117] "RemoveContainer" containerID="3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c" Feb 27 17:05:42 crc kubenswrapper[4700]: E0227 17:05:42.227656 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\": container with ID starting with 3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c not found: ID does not exist" containerID="3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c" Feb 27 17:05:42 crc kubenswrapper[4700]: I0227 17:05:42.227730 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c"} err="failed to get container status \"3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\": rpc error: code = NotFound desc = could not find container \"3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c\": container with ID starting with 3aa910d58898f9c1725a974b2b9eda7e0ca89c0926a594a6a8cb3412ba96387c not found: ID does not exist" Feb 27 17:05:42 crc kubenswrapper[4700]: I0227 17:05:42.992721 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Feb 27 17:05:45 crc kubenswrapper[4700]: E0227 17:05:45.992504 4700 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:05:45 crc kubenswrapper[4700]: E0227 17:05:45.992939 4700 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:05:45 crc kubenswrapper[4700]: E0227 17:05:45.993304 4700 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:05:45 crc kubenswrapper[4700]: E0227 17:05:45.993875 4700 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:05:45 crc kubenswrapper[4700]: E0227 17:05:45.994532 4700 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:05:45 crc kubenswrapper[4700]: I0227 17:05:45.994589 4700 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Feb 27 17:05:45 crc kubenswrapper[4700]: E0227 17:05:45.995001 4700 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.222:6443: connect: connection refused" interval="200ms" Feb 27 17:05:46 crc kubenswrapper[4700]: E0227 17:05:46.195903 4700 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.222:6443: connect: connection refused" interval="400ms" Feb 27 17:05:46 crc kubenswrapper[4700]: E0227 17:05:46.597399 4700 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.222:6443: connect: connection refused" interval="800ms" Feb 27 17:05:47 crc kubenswrapper[4700]: E0227 17:05:47.398949 4700 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.222:6443: connect: connection refused" interval="1.6s" Feb 27 17:05:49 crc kubenswrapper[4700]: E0227 17:05:49.000521 4700 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.222:6443: connect: connection refused" interval="3.2s" Feb 27 17:05:50 crc kubenswrapper[4700]: I0227 17:05:50.986272 4700 status_manager.go:851] "Failed to get status for pod" podUID="eb3329cc-cc82-4f3c-87c0-b2baf19069eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:05:50 crc kubenswrapper[4700]: I0227 17:05:50.986965 4700 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:05:52 crc kubenswrapper[4700]: E0227 17:05:52.079361 4700 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.222:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.1898295e900d8223 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:05:39.490759203 +0000 UTC m=+299.476071960,LastTimestamp:2026-02-27 17:05:39.490759203 +0000 UTC m=+299.476071960,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:05:52 crc kubenswrapper[4700]: E0227 17:05:52.202339 4700 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.222:6443: connect: connection refused" interval="6.4s" Feb 27 17:05:53 crc kubenswrapper[4700]: I0227 17:05:53.980817 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:05:53 crc kubenswrapper[4700]: I0227 17:05:53.981976 4700 status_manager.go:851] "Failed to get status for pod" podUID="eb3329cc-cc82-4f3c-87c0-b2baf19069eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:05:53 crc kubenswrapper[4700]: I0227 17:05:53.982541 4700 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:05:54 crc kubenswrapper[4700]: I0227 17:05:54.002656 4700 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1a2d1b57-7fc1-4614-aac5-7144cdd812d5" Feb 27 17:05:54 crc kubenswrapper[4700]: I0227 17:05:54.002695 4700 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1a2d1b57-7fc1-4614-aac5-7144cdd812d5" Feb 27 17:05:54 crc kubenswrapper[4700]: E0227 17:05:54.003170 4700 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:05:54 crc kubenswrapper[4700]: I0227 17:05:54.003758 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:05:54 crc kubenswrapper[4700]: I0227 17:05:54.606888 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"6f797f010211c9f062e01baf4fa972d7fe4ee8e7a745f25e0b40f12b48d68004"} Feb 27 17:05:57 crc kubenswrapper[4700]: I0227 17:05:57.160628 4700 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Readiness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Feb 27 17:05:57 crc kubenswrapper[4700]: I0227 17:05:57.160962 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Feb 27 17:05:58 crc kubenswrapper[4700]: E0227 17:05:58.604546 4700 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.222:6443: connect: connection refused" interval="7s" Feb 27 17:05:58 crc kubenswrapper[4700]: I0227 17:05:58.638084 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Feb 27 17:05:58 crc kubenswrapper[4700]: I0227 17:05:58.639282 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 27 17:05:58 crc kubenswrapper[4700]: I0227 17:05:58.639357 4700 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="555793b646e823bff19401b489c4ab8de44ee3cd8f82ebcd66b5b08e67450be7" exitCode=1 Feb 27 17:05:58 crc kubenswrapper[4700]: I0227 17:05:58.639400 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"555793b646e823bff19401b489c4ab8de44ee3cd8f82ebcd66b5b08e67450be7"} Feb 27 17:05:58 crc kubenswrapper[4700]: I0227 17:05:58.640109 4700 scope.go:117] "RemoveContainer" containerID="555793b646e823bff19401b489c4ab8de44ee3cd8f82ebcd66b5b08e67450be7" Feb 27 17:05:58 crc kubenswrapper[4700]: I0227 17:05:58.641001 4700 status_manager.go:851] "Failed to get status for pod" podUID="eb3329cc-cc82-4f3c-87c0-b2baf19069eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:05:58 crc kubenswrapper[4700]: I0227 17:05:58.641648 4700 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:05:58 crc kubenswrapper[4700]: I0227 17:05:58.642399 4700 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:05:59 crc kubenswrapper[4700]: I0227 17:05:59.649444 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"05c9f28f8b09d5779c90d220772b9fbe43ec6393c459e1610c8ecd706a9a451f"} Feb 27 17:06:00 crc kubenswrapper[4700]: I0227 17:06:00.991686 4700 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:06:00 crc kubenswrapper[4700]: I0227 17:06:00.992831 4700 status_manager.go:851] "Failed to get status for pod" podUID="eb3329cc-cc82-4f3c-87c0-b2baf19069eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:06:00 crc kubenswrapper[4700]: I0227 17:06:00.993390 4700 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:06:00 crc kubenswrapper[4700]: I0227 17:06:00.993848 4700 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:06:01 crc kubenswrapper[4700]: I0227 17:06:01.670099 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Feb 27 17:06:01 crc kubenswrapper[4700]: I0227 17:06:01.670909 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 27 17:06:02 crc kubenswrapper[4700]: E0227 17:06:02.081765 4700 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.222:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.1898295e900d8223 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 17:05:39.490759203 +0000 UTC m=+299.476071960,LastTimestamp:2026-02-27 17:05:39.490759203 +0000 UTC m=+299.476071960,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 17:06:02 crc kubenswrapper[4700]: I0227 17:06:02.189495 4700 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 17:06:02 crc kubenswrapper[4700]: I0227 17:06:02.680601 4700 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="05c9f28f8b09d5779c90d220772b9fbe43ec6393c459e1610c8ecd706a9a451f" exitCode=0 Feb 27 17:06:02 crc kubenswrapper[4700]: I0227 17:06:02.680701 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"05c9f28f8b09d5779c90d220772b9fbe43ec6393c459e1610c8ecd706a9a451f"} Feb 27 17:06:02 crc kubenswrapper[4700]: I0227 17:06:02.681212 4700 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1a2d1b57-7fc1-4614-aac5-7144cdd812d5" Feb 27 17:06:02 crc kubenswrapper[4700]: I0227 17:06:02.681279 4700 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1a2d1b57-7fc1-4614-aac5-7144cdd812d5" Feb 27 17:06:02 crc kubenswrapper[4700]: I0227 17:06:02.681644 4700 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:06:02 crc kubenswrapper[4700]: E0227 17:06:02.681869 4700 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:06:02 crc kubenswrapper[4700]: I0227 17:06:02.682211 4700 status_manager.go:851] "Failed to get status for pod" podUID="eb3329cc-cc82-4f3c-87c0-b2baf19069eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:06:02 crc kubenswrapper[4700]: I0227 17:06:02.682822 4700 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:06:02 crc kubenswrapper[4700]: I0227 17:06:02.683313 4700 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:06:02 crc kubenswrapper[4700]: I0227 17:06:02.686069 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Feb 27 17:06:02 crc kubenswrapper[4700]: I0227 17:06:02.687012 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 27 17:06:02 crc kubenswrapper[4700]: I0227 17:06:02.687090 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"39e08d9b219abc555675d205aad5313a653e65584ac6bdb77752db220c584b9d"} Feb 27 17:06:02 crc kubenswrapper[4700]: I0227 17:06:02.687876 4700 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:06:02 crc kubenswrapper[4700]: I0227 17:06:02.688623 4700 status_manager.go:851] "Failed to get status for pod" podUID="eb3329cc-cc82-4f3c-87c0-b2baf19069eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:06:02 crc kubenswrapper[4700]: I0227 17:06:02.689117 4700 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:06:02 crc kubenswrapper[4700]: I0227 17:06:02.689626 4700 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.222:6443: connect: connection refused" Feb 27 17:06:03 crc kubenswrapper[4700]: E0227 17:06:03.026438 4700 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.222:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" volumeName="registry-storage" Feb 27 17:06:03 crc kubenswrapper[4700]: I0227 17:06:03.706741 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"97999cf66a76ac903767db83709f3ce3ee5fcc9dc8592cfacffe2dbd85270cde"} Feb 27 17:06:03 crc kubenswrapper[4700]: I0227 17:06:03.706805 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"10c4636cd8969974bb843212300e4568724871a8cddb0f359ff267996ce3fbd7"} Feb 27 17:06:04 crc kubenswrapper[4700]: I0227 17:06:04.718177 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"9d813967cdb3bd182a8a26d1f6b676cd09dd2e00cc925e4c93922d5596d8eb85"} Feb 27 17:06:04 crc kubenswrapper[4700]: I0227 17:06:04.718539 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:06:04 crc kubenswrapper[4700]: I0227 17:06:04.718555 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"89a7e6118c561ce346e69bcdd1289ad10fcad2086344ad41459a74edc521e4db"} Feb 27 17:06:04 crc kubenswrapper[4700]: I0227 17:06:04.718568 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"7c2ea4dcbaec77f5065497a036fcc9d51554ccf9f49220a3c483bafe21c1f1ca"} Feb 27 17:06:04 crc kubenswrapper[4700]: I0227 17:06:04.718707 4700 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1a2d1b57-7fc1-4614-aac5-7144cdd812d5" Feb 27 17:06:04 crc kubenswrapper[4700]: I0227 17:06:04.718745 4700 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1a2d1b57-7fc1-4614-aac5-7144cdd812d5" Feb 27 17:06:06 crc kubenswrapper[4700]: I0227 17:06:06.296268 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 17:06:06 crc kubenswrapper[4700]: I0227 17:06:06.302594 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 17:06:06 crc kubenswrapper[4700]: I0227 17:06:06.730500 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 17:06:09 crc kubenswrapper[4700]: I0227 17:06:09.004626 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:06:09 crc kubenswrapper[4700]: I0227 17:06:09.004709 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:06:09 crc kubenswrapper[4700]: I0227 17:06:09.013387 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:06:09 crc kubenswrapper[4700]: I0227 17:06:09.821025 4700 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:06:09 crc kubenswrapper[4700]: I0227 17:06:09.875866 4700 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a2d1b57-7fc1-4614-aac5-7144cdd812d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:06:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:06:02Z\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T17:05:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10c4636cd8969974bb843212300e4568724871a8cddb0f359ff267996ce3fbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c2ea4dcbaec77f5065497a036fcc9d51554ccf9f49220a3c483bafe21c1f1ca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97999cf66a76ac903767db83709f3ce3ee5fcc9dc8592cfacffe2dbd85270cde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:06:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d813967cdb3bd182a8a26d1f6b676cd09dd2e00cc925e4c93922d5596d8eb85\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:06:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89a7e6118c561ce346e69bcdd1289ad10fcad2086344ad41459a74edc521e4db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T17:06:04Z\\\"}}}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05c9f28f8b09d5779c90d220772b9fbe43ec6393c459e1610c8ecd706a9a451f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05c9f28f8b09d5779c90d220772b9fbe43ec6393c459e1610c8ecd706a9a451f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-27T17:06:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-27T17:05:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}]}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Pod \"kube-apiserver-crc\" is invalid: metadata.uid: Invalid value: \"1a2d1b57-7fc1-4614-aac5-7144cdd812d5\": field is immutable" Feb 27 17:06:09 crc kubenswrapper[4700]: I0227 17:06:09.969112 4700 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="cd65e78d-38c0-4dbf-936a-20285d929487" Feb 27 17:06:10 crc kubenswrapper[4700]: I0227 17:06:10.756922 4700 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1a2d1b57-7fc1-4614-aac5-7144cdd812d5" Feb 27 17:06:10 crc kubenswrapper[4700]: I0227 17:06:10.756974 4700 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1a2d1b57-7fc1-4614-aac5-7144cdd812d5" Feb 27 17:06:10 crc kubenswrapper[4700]: I0227 17:06:10.761796 4700 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="cd65e78d-38c0-4dbf-936a-20285d929487" Feb 27 17:06:17 crc kubenswrapper[4700]: I0227 17:06:17.164256 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 17:06:24 crc kubenswrapper[4700]: I0227 17:06:24.067945 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 27 17:06:26 crc kubenswrapper[4700]: I0227 17:06:26.587942 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 27 17:06:27 crc kubenswrapper[4700]: I0227 17:06:27.938359 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 27 17:06:30 crc kubenswrapper[4700]: I0227 17:06:30.368272 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 27 17:06:30 crc kubenswrapper[4700]: I0227 17:06:30.586251 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 27 17:06:34 crc kubenswrapper[4700]: I0227 17:06:34.356004 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 27 17:06:34 crc kubenswrapper[4700]: I0227 17:06:34.560353 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 27 17:06:37 crc kubenswrapper[4700]: I0227 17:06:37.518039 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 27 17:06:37 crc kubenswrapper[4700]: I0227 17:06:37.846739 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 27 17:06:38 crc kubenswrapper[4700]: I0227 17:06:38.648171 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 27 17:06:38 crc kubenswrapper[4700]: I0227 17:06:38.732673 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 27 17:06:39 crc kubenswrapper[4700]: I0227 17:06:39.099803 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 27 17:06:39 crc kubenswrapper[4700]: I0227 17:06:39.294075 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 27 17:06:39 crc kubenswrapper[4700]: I0227 17:06:39.729925 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 27 17:06:40 crc kubenswrapper[4700]: I0227 17:06:40.245847 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 27 17:06:40 crc kubenswrapper[4700]: I0227 17:06:40.453297 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 27 17:06:40 crc kubenswrapper[4700]: I0227 17:06:40.657054 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 27 17:06:41 crc kubenswrapper[4700]: I0227 17:06:41.113765 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 27 17:06:41 crc kubenswrapper[4700]: I0227 17:06:41.343937 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 27 17:06:41 crc kubenswrapper[4700]: I0227 17:06:41.614631 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 27 17:06:41 crc kubenswrapper[4700]: I0227 17:06:41.692019 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 27 17:06:41 crc kubenswrapper[4700]: I0227 17:06:41.938697 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 27 17:06:42 crc kubenswrapper[4700]: I0227 17:06:42.361861 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 27 17:06:42 crc kubenswrapper[4700]: I0227 17:06:42.461175 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 27 17:06:42 crc kubenswrapper[4700]: I0227 17:06:42.644190 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 27 17:06:42 crc kubenswrapper[4700]: I0227 17:06:42.682842 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 27 17:06:42 crc kubenswrapper[4700]: I0227 17:06:42.897825 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 27 17:06:43 crc kubenswrapper[4700]: I0227 17:06:43.963686 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 27 17:06:44 crc kubenswrapper[4700]: I0227 17:06:44.380690 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 27 17:06:44 crc kubenswrapper[4700]: I0227 17:06:44.430602 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 27 17:06:44 crc kubenswrapper[4700]: I0227 17:06:44.435224 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 27 17:06:44 crc kubenswrapper[4700]: I0227 17:06:44.442437 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 27 17:06:44 crc kubenswrapper[4700]: I0227 17:06:44.722702 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 27 17:06:45 crc kubenswrapper[4700]: I0227 17:06:45.024952 4700 generic.go:334] "Generic (PLEG): container finished" podID="0cc33e9e-5984-454d-880c-5187145a2c36" containerID="d4afb43ef0068fad5a80a64fec6abe581f5e3d7507a435c30c279abc4eaff145" exitCode=0 Feb 27 17:06:45 crc kubenswrapper[4700]: I0227 17:06:45.025015 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hc86c" event={"ID":"0cc33e9e-5984-454d-880c-5187145a2c36","Type":"ContainerDied","Data":"d4afb43ef0068fad5a80a64fec6abe581f5e3d7507a435c30c279abc4eaff145"} Feb 27 17:06:45 crc kubenswrapper[4700]: I0227 17:06:45.025709 4700 scope.go:117] "RemoveContainer" containerID="d4afb43ef0068fad5a80a64fec6abe581f5e3d7507a435c30c279abc4eaff145" Feb 27 17:06:45 crc kubenswrapper[4700]: I0227 17:06:45.042167 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 27 17:06:45 crc kubenswrapper[4700]: I0227 17:06:45.046189 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 27 17:06:45 crc kubenswrapper[4700]: I0227 17:06:45.633161 4700 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 27 17:06:45 crc kubenswrapper[4700]: I0227 17:06:45.684284 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 27 17:06:45 crc kubenswrapper[4700]: I0227 17:06:45.698393 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 27 17:06:45 crc kubenswrapper[4700]: I0227 17:06:45.754712 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 27 17:06:46 crc kubenswrapper[4700]: I0227 17:06:46.035221 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-hc86c_0cc33e9e-5984-454d-880c-5187145a2c36/marketplace-operator/1.log" Feb 27 17:06:46 crc kubenswrapper[4700]: I0227 17:06:46.037063 4700 generic.go:334] "Generic (PLEG): container finished" podID="0cc33e9e-5984-454d-880c-5187145a2c36" containerID="3582c66af7d192ae89576057e17a16a2426a9994835101bbc236e3316945298e" exitCode=1 Feb 27 17:06:46 crc kubenswrapper[4700]: I0227 17:06:46.037125 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hc86c" event={"ID":"0cc33e9e-5984-454d-880c-5187145a2c36","Type":"ContainerDied","Data":"3582c66af7d192ae89576057e17a16a2426a9994835101bbc236e3316945298e"} Feb 27 17:06:46 crc kubenswrapper[4700]: I0227 17:06:46.037185 4700 scope.go:117] "RemoveContainer" containerID="d4afb43ef0068fad5a80a64fec6abe581f5e3d7507a435c30c279abc4eaff145" Feb 27 17:06:46 crc kubenswrapper[4700]: I0227 17:06:46.037864 4700 scope.go:117] "RemoveContainer" containerID="3582c66af7d192ae89576057e17a16a2426a9994835101bbc236e3316945298e" Feb 27 17:06:46 crc kubenswrapper[4700]: E0227 17:06:46.038238 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-hc86c_openshift-marketplace(0cc33e9e-5984-454d-880c-5187145a2c36)\"" pod="openshift-marketplace/marketplace-operator-79b997595-hc86c" podUID="0cc33e9e-5984-454d-880c-5187145a2c36" Feb 27 17:06:46 crc kubenswrapper[4700]: I0227 17:06:46.311976 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 27 17:06:46 crc kubenswrapper[4700]: I0227 17:06:46.332609 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 27 17:06:46 crc kubenswrapper[4700]: I0227 17:06:46.429092 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 27 17:06:46 crc kubenswrapper[4700]: I0227 17:06:46.872893 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 27 17:06:46 crc kubenswrapper[4700]: I0227 17:06:46.915959 4700 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 27 17:06:46 crc kubenswrapper[4700]: I0227 17:06:46.934154 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 27 17:06:47 crc kubenswrapper[4700]: I0227 17:06:47.044589 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-hc86c_0cc33e9e-5984-454d-880c-5187145a2c36/marketplace-operator/1.log" Feb 27 17:06:47 crc kubenswrapper[4700]: I0227 17:06:47.528814 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 27 17:06:47 crc kubenswrapper[4700]: I0227 17:06:47.602622 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 27 17:06:47 crc kubenswrapper[4700]: I0227 17:06:47.619175 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 27 17:06:47 crc kubenswrapper[4700]: I0227 17:06:47.693381 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 27 17:06:47 crc kubenswrapper[4700]: I0227 17:06:47.711765 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 27 17:06:47 crc kubenswrapper[4700]: I0227 17:06:47.758978 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 27 17:06:47 crc kubenswrapper[4700]: I0227 17:06:47.877180 4700 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-marketplace/marketplace-operator-79b997595-hc86c" Feb 27 17:06:47 crc kubenswrapper[4700]: I0227 17:06:47.877266 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-hc86c" Feb 27 17:06:47 crc kubenswrapper[4700]: I0227 17:06:47.878038 4700 scope.go:117] "RemoveContainer" containerID="3582c66af7d192ae89576057e17a16a2426a9994835101bbc236e3316945298e" Feb 27 17:06:47 crc kubenswrapper[4700]: E0227 17:06:47.878503 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-hc86c_openshift-marketplace(0cc33e9e-5984-454d-880c-5187145a2c36)\"" pod="openshift-marketplace/marketplace-operator-79b997595-hc86c" podUID="0cc33e9e-5984-454d-880c-5187145a2c36" Feb 27 17:06:47 crc kubenswrapper[4700]: I0227 17:06:47.907363 4700 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 27 17:06:48 crc kubenswrapper[4700]: I0227 17:06:48.009911 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 27 17:06:48 crc kubenswrapper[4700]: I0227 17:06:48.027267 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 27 17:06:48 crc kubenswrapper[4700]: I0227 17:06:48.056239 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 27 17:06:48 crc kubenswrapper[4700]: I0227 17:06:48.096971 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 27 17:06:48 crc kubenswrapper[4700]: I0227 17:06:48.355290 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 27 17:06:48 crc kubenswrapper[4700]: I0227 17:06:48.379733 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 27 17:06:48 crc kubenswrapper[4700]: I0227 17:06:48.453104 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 27 17:06:48 crc kubenswrapper[4700]: I0227 17:06:48.743326 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 27 17:06:49 crc kubenswrapper[4700]: I0227 17:06:49.003331 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 27 17:06:49 crc kubenswrapper[4700]: I0227 17:06:49.084374 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 27 17:06:49 crc kubenswrapper[4700]: I0227 17:06:49.170370 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 27 17:06:49 crc kubenswrapper[4700]: I0227 17:06:49.326634 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 27 17:06:49 crc kubenswrapper[4700]: I0227 17:06:49.327116 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 27 17:06:49 crc kubenswrapper[4700]: I0227 17:06:49.327433 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 27 17:06:49 crc kubenswrapper[4700]: I0227 17:06:49.407292 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 27 17:06:49 crc kubenswrapper[4700]: I0227 17:06:49.613732 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 27 17:06:49 crc kubenswrapper[4700]: I0227 17:06:49.640302 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 27 17:06:49 crc kubenswrapper[4700]: I0227 17:06:49.648802 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 27 17:06:49 crc kubenswrapper[4700]: I0227 17:06:49.724122 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 27 17:06:49 crc kubenswrapper[4700]: I0227 17:06:49.838537 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 27 17:06:49 crc kubenswrapper[4700]: I0227 17:06:49.965519 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 27 17:06:50 crc kubenswrapper[4700]: I0227 17:06:50.006843 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 27 17:06:50 crc kubenswrapper[4700]: I0227 17:06:50.023402 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 27 17:06:50 crc kubenswrapper[4700]: I0227 17:06:50.195695 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 27 17:06:50 crc kubenswrapper[4700]: I0227 17:06:50.467867 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 27 17:06:50 crc kubenswrapper[4700]: I0227 17:06:50.474074 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 27 17:06:50 crc kubenswrapper[4700]: I0227 17:06:50.481762 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 27 17:06:50 crc kubenswrapper[4700]: I0227 17:06:50.718628 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 27 17:06:50 crc kubenswrapper[4700]: I0227 17:06:50.966234 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 27 17:06:50 crc kubenswrapper[4700]: I0227 17:06:50.999521 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 27 17:06:51 crc kubenswrapper[4700]: I0227 17:06:51.101560 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 27 17:06:51 crc kubenswrapper[4700]: I0227 17:06:51.121333 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 27 17:06:51 crc kubenswrapper[4700]: I0227 17:06:51.134975 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 27 17:06:51 crc kubenswrapper[4700]: I0227 17:06:51.477581 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 27 17:06:51 crc kubenswrapper[4700]: I0227 17:06:51.512509 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 27 17:06:51 crc kubenswrapper[4700]: I0227 17:06:51.529040 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 27 17:06:51 crc kubenswrapper[4700]: I0227 17:06:51.541824 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 27 17:06:51 crc kubenswrapper[4700]: I0227 17:06:51.691018 4700 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 27 17:06:51 crc kubenswrapper[4700]: I0227 17:06:51.695915 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=72.695894001 podStartE2EDuration="1m12.695894001s" podCreationTimestamp="2026-02-27 17:05:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:06:09.926207458 +0000 UTC m=+329.911520215" watchObservedRunningTime="2026-02-27 17:06:51.695894001 +0000 UTC m=+371.681206778" Feb 27 17:06:51 crc kubenswrapper[4700]: I0227 17:06:51.698039 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 27 17:06:51 crc kubenswrapper[4700]: I0227 17:06:51.698096 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 27 17:06:51 crc kubenswrapper[4700]: I0227 17:06:51.705281 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:06:51 crc kubenswrapper[4700]: I0227 17:06:51.706972 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 17:06:51 crc kubenswrapper[4700]: I0227 17:06:51.732325 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 27 17:06:51 crc kubenswrapper[4700]: I0227 17:06:51.732800 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 27 17:06:51 crc kubenswrapper[4700]: I0227 17:06:51.765853 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=42.765828695 podStartE2EDuration="42.765828695s" podCreationTimestamp="2026-02-27 17:06:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:06:51.729133163 +0000 UTC m=+371.714445950" watchObservedRunningTime="2026-02-27 17:06:51.765828695 +0000 UTC m=+371.751141482" Feb 27 17:06:52 crc kubenswrapper[4700]: I0227 17:06:52.602580 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 27 17:06:52 crc kubenswrapper[4700]: I0227 17:06:52.714311 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 27 17:06:52 crc kubenswrapper[4700]: I0227 17:06:52.805089 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 27 17:06:52 crc kubenswrapper[4700]: I0227 17:06:52.810449 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 27 17:06:52 crc kubenswrapper[4700]: I0227 17:06:52.819498 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 27 17:06:53 crc kubenswrapper[4700]: I0227 17:06:53.104609 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 27 17:06:53 crc kubenswrapper[4700]: I0227 17:06:53.270885 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 27 17:06:53 crc kubenswrapper[4700]: I0227 17:06:53.543813 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 27 17:06:53 crc kubenswrapper[4700]: I0227 17:06:53.555560 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 27 17:06:53 crc kubenswrapper[4700]: I0227 17:06:53.676736 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 27 17:06:53 crc kubenswrapper[4700]: I0227 17:06:53.784876 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 27 17:06:53 crc kubenswrapper[4700]: I0227 17:06:53.941492 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 27 17:06:53 crc kubenswrapper[4700]: I0227 17:06:53.955902 4700 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 27 17:06:53 crc kubenswrapper[4700]: I0227 17:06:53.956141 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://7ef0dfdaac76cf9ec1a86a1b96773b7ccc0a4488abafebed86587f514efab52b" gracePeriod=5 Feb 27 17:06:53 crc kubenswrapper[4700]: I0227 17:06:53.973637 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 27 17:06:54 crc kubenswrapper[4700]: I0227 17:06:54.067004 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 27 17:06:54 crc kubenswrapper[4700]: I0227 17:06:54.075840 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 27 17:06:54 crc kubenswrapper[4700]: I0227 17:06:54.299122 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 27 17:06:54 crc kubenswrapper[4700]: I0227 17:06:54.351050 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 27 17:06:54 crc kubenswrapper[4700]: I0227 17:06:54.422952 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 27 17:06:54 crc kubenswrapper[4700]: I0227 17:06:54.587730 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 27 17:06:54 crc kubenswrapper[4700]: I0227 17:06:54.837497 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 27 17:06:54 crc kubenswrapper[4700]: I0227 17:06:54.870582 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 27 17:06:55 crc kubenswrapper[4700]: I0227 17:06:55.051490 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 27 17:06:55 crc kubenswrapper[4700]: I0227 17:06:55.140066 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 27 17:06:55 crc kubenswrapper[4700]: I0227 17:06:55.149648 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 27 17:06:55 crc kubenswrapper[4700]: I0227 17:06:55.357304 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 27 17:06:55 crc kubenswrapper[4700]: I0227 17:06:55.485196 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 27 17:06:55 crc kubenswrapper[4700]: I0227 17:06:55.579390 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 27 17:06:55 crc kubenswrapper[4700]: I0227 17:06:55.631342 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 27 17:06:55 crc kubenswrapper[4700]: I0227 17:06:55.672538 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 27 17:06:56 crc kubenswrapper[4700]: I0227 17:06:56.132232 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 27 17:06:56 crc kubenswrapper[4700]: I0227 17:06:56.155294 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 27 17:06:56 crc kubenswrapper[4700]: I0227 17:06:56.393713 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 27 17:06:56 crc kubenswrapper[4700]: I0227 17:06:56.425498 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 27 17:06:56 crc kubenswrapper[4700]: I0227 17:06:56.647381 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 27 17:06:56 crc kubenswrapper[4700]: I0227 17:06:56.797370 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 27 17:06:56 crc kubenswrapper[4700]: I0227 17:06:56.817481 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 27 17:06:56 crc kubenswrapper[4700]: I0227 17:06:56.987615 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 27 17:06:57 crc kubenswrapper[4700]: I0227 17:06:57.102023 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 27 17:06:57 crc kubenswrapper[4700]: I0227 17:06:57.160993 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 27 17:06:57 crc kubenswrapper[4700]: I0227 17:06:57.381090 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 27 17:06:57 crc kubenswrapper[4700]: I0227 17:06:57.445880 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 27 17:06:57 crc kubenswrapper[4700]: I0227 17:06:57.768874 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 27 17:06:57 crc kubenswrapper[4700]: I0227 17:06:57.771110 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 27 17:06:57 crc kubenswrapper[4700]: I0227 17:06:57.804326 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 27 17:06:58 crc kubenswrapper[4700]: I0227 17:06:58.025295 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 27 17:06:58 crc kubenswrapper[4700]: I0227 17:06:58.061897 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 27 17:06:58 crc kubenswrapper[4700]: I0227 17:06:58.148311 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 27 17:06:58 crc kubenswrapper[4700]: I0227 17:06:58.233562 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 27 17:06:58 crc kubenswrapper[4700]: I0227 17:06:58.409957 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 27 17:06:58 crc kubenswrapper[4700]: I0227 17:06:58.463419 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 27 17:06:58 crc kubenswrapper[4700]: I0227 17:06:58.616844 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536866-th9ft"] Feb 27 17:06:58 crc kubenswrapper[4700]: E0227 17:06:58.617179 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 27 17:06:58 crc kubenswrapper[4700]: I0227 17:06:58.617201 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 27 17:06:58 crc kubenswrapper[4700]: E0227 17:06:58.617219 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb3329cc-cc82-4f3c-87c0-b2baf19069eb" containerName="installer" Feb 27 17:06:58 crc kubenswrapper[4700]: I0227 17:06:58.617232 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb3329cc-cc82-4f3c-87c0-b2baf19069eb" containerName="installer" Feb 27 17:06:58 crc kubenswrapper[4700]: I0227 17:06:58.617414 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb3329cc-cc82-4f3c-87c0-b2baf19069eb" containerName="installer" Feb 27 17:06:58 crc kubenswrapper[4700]: I0227 17:06:58.617454 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 27 17:06:58 crc kubenswrapper[4700]: I0227 17:06:58.618048 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536866-th9ft" Feb 27 17:06:58 crc kubenswrapper[4700]: I0227 17:06:58.621628 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:06:58 crc kubenswrapper[4700]: I0227 17:06:58.622011 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:06:58 crc kubenswrapper[4700]: I0227 17:06:58.622359 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d62kc" Feb 27 17:06:58 crc kubenswrapper[4700]: I0227 17:06:58.623593 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 27 17:06:58 crc kubenswrapper[4700]: I0227 17:06:58.631976 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536866-th9ft"] Feb 27 17:06:58 crc kubenswrapper[4700]: I0227 17:06:58.656148 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 27 17:06:58 crc kubenswrapper[4700]: I0227 17:06:58.688761 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 27 17:06:58 crc kubenswrapper[4700]: I0227 17:06:58.747616 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 27 17:06:58 crc kubenswrapper[4700]: I0227 17:06:58.790786 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9pzp\" (UniqueName: \"kubernetes.io/projected/4f0dd4fc-ce75-4088-be4d-664e4b5a31ea-kube-api-access-f9pzp\") pod \"auto-csr-approver-29536866-th9ft\" (UID: \"4f0dd4fc-ce75-4088-be4d-664e4b5a31ea\") " pod="openshift-infra/auto-csr-approver-29536866-th9ft" Feb 27 17:06:58 crc kubenswrapper[4700]: I0227 17:06:58.892733 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9pzp\" (UniqueName: \"kubernetes.io/projected/4f0dd4fc-ce75-4088-be4d-664e4b5a31ea-kube-api-access-f9pzp\") pod \"auto-csr-approver-29536866-th9ft\" (UID: \"4f0dd4fc-ce75-4088-be4d-664e4b5a31ea\") " pod="openshift-infra/auto-csr-approver-29536866-th9ft" Feb 27 17:06:58 crc kubenswrapper[4700]: I0227 17:06:58.932394 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9pzp\" (UniqueName: \"kubernetes.io/projected/4f0dd4fc-ce75-4088-be4d-664e4b5a31ea-kube-api-access-f9pzp\") pod \"auto-csr-approver-29536866-th9ft\" (UID: \"4f0dd4fc-ce75-4088-be4d-664e4b5a31ea\") " pod="openshift-infra/auto-csr-approver-29536866-th9ft" Feb 27 17:06:58 crc kubenswrapper[4700]: I0227 17:06:58.950424 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536866-th9ft" Feb 27 17:06:58 crc kubenswrapper[4700]: I0227 17:06:58.954478 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 27 17:06:59 crc kubenswrapper[4700]: I0227 17:06:59.126984 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 27 17:06:59 crc kubenswrapper[4700]: I0227 17:06:59.127251 4700 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="7ef0dfdaac76cf9ec1a86a1b96773b7ccc0a4488abafebed86587f514efab52b" exitCode=137 Feb 27 17:06:59 crc kubenswrapper[4700]: I0227 17:06:59.128165 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 27 17:06:59 crc kubenswrapper[4700]: I0227 17:06:59.152119 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 27 17:06:59 crc kubenswrapper[4700]: I0227 17:06:59.207709 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 27 17:06:59 crc kubenswrapper[4700]: I0227 17:06:59.268267 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 27 17:06:59 crc kubenswrapper[4700]: I0227 17:06:59.479907 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 27 17:06:59 crc kubenswrapper[4700]: I0227 17:06:59.485987 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536866-th9ft"] Feb 27 17:06:59 crc kubenswrapper[4700]: I0227 17:06:59.566299 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 27 17:06:59 crc kubenswrapper[4700]: I0227 17:06:59.566704 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 17:06:59 crc kubenswrapper[4700]: I0227 17:06:59.629213 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 27 17:06:59 crc kubenswrapper[4700]: I0227 17:06:59.672911 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 27 17:06:59 crc kubenswrapper[4700]: I0227 17:06:59.705218 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 27 17:06:59 crc kubenswrapper[4700]: I0227 17:06:59.705283 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:06:59 crc kubenswrapper[4700]: I0227 17:06:59.705384 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 27 17:06:59 crc kubenswrapper[4700]: I0227 17:06:59.705432 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 27 17:06:59 crc kubenswrapper[4700]: I0227 17:06:59.705536 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 27 17:06:59 crc kubenswrapper[4700]: I0227 17:06:59.705568 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:06:59 crc kubenswrapper[4700]: I0227 17:06:59.705586 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 27 17:06:59 crc kubenswrapper[4700]: I0227 17:06:59.705635 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:06:59 crc kubenswrapper[4700]: I0227 17:06:59.705751 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:06:59 crc kubenswrapper[4700]: I0227 17:06:59.705918 4700 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 27 17:06:59 crc kubenswrapper[4700]: I0227 17:06:59.705939 4700 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Feb 27 17:06:59 crc kubenswrapper[4700]: I0227 17:06:59.705953 4700 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Feb 27 17:06:59 crc kubenswrapper[4700]: I0227 17:06:59.705966 4700 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Feb 27 17:06:59 crc kubenswrapper[4700]: I0227 17:06:59.717343 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:06:59 crc kubenswrapper[4700]: I0227 17:06:59.788014 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 27 17:06:59 crc kubenswrapper[4700]: I0227 17:06:59.797930 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 27 17:06:59 crc kubenswrapper[4700]: I0227 17:06:59.807048 4700 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 27 17:06:59 crc kubenswrapper[4700]: I0227 17:06:59.935194 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 27 17:07:00 crc kubenswrapper[4700]: I0227 17:07:00.111119 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 27 17:07:00 crc kubenswrapper[4700]: I0227 17:07:00.140297 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 27 17:07:00 crc kubenswrapper[4700]: I0227 17:07:00.140376 4700 scope.go:117] "RemoveContainer" containerID="7ef0dfdaac76cf9ec1a86a1b96773b7ccc0a4488abafebed86587f514efab52b" Feb 27 17:07:00 crc kubenswrapper[4700]: I0227 17:07:00.140566 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 17:07:00 crc kubenswrapper[4700]: I0227 17:07:00.155928 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536866-th9ft" event={"ID":"4f0dd4fc-ce75-4088-be4d-664e4b5a31ea","Type":"ContainerStarted","Data":"42b5f5d2b4e634c18bbec54d24ae3e3cac14fe4097399704344aa1a669aa56cf"} Feb 27 17:07:00 crc kubenswrapper[4700]: I0227 17:07:00.262559 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 27 17:07:00 crc kubenswrapper[4700]: I0227 17:07:00.331113 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 27 17:07:00 crc kubenswrapper[4700]: I0227 17:07:00.427307 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 27 17:07:00 crc kubenswrapper[4700]: I0227 17:07:00.661615 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 27 17:07:00 crc kubenswrapper[4700]: I0227 17:07:00.680683 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 27 17:07:00 crc kubenswrapper[4700]: I0227 17:07:00.994922 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Feb 27 17:07:00 crc kubenswrapper[4700]: I0227 17:07:00.995260 4700 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Feb 27 17:07:01 crc kubenswrapper[4700]: I0227 17:07:01.017053 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 27 17:07:01 crc kubenswrapper[4700]: I0227 17:07:01.017100 4700 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="025a7b24-bf92-4de9-8d4b-2f69e82084a1" Feb 27 17:07:01 crc kubenswrapper[4700]: I0227 17:07:01.023547 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 27 17:07:01 crc kubenswrapper[4700]: I0227 17:07:01.023611 4700 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="025a7b24-bf92-4de9-8d4b-2f69e82084a1" Feb 27 17:07:01 crc kubenswrapper[4700]: I0227 17:07:01.034813 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 27 17:07:01 crc kubenswrapper[4700]: I0227 17:07:01.121981 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 27 17:07:01 crc kubenswrapper[4700]: I0227 17:07:01.122641 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 27 17:07:01 crc kubenswrapper[4700]: I0227 17:07:01.162602 4700 generic.go:334] "Generic (PLEG): container finished" podID="4f0dd4fc-ce75-4088-be4d-664e4b5a31ea" containerID="b7d75a20f6b55ff385d8e95852faafad27de83d67600bb63df0f57aceea1a925" exitCode=0 Feb 27 17:07:01 crc kubenswrapper[4700]: I0227 17:07:01.162843 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536866-th9ft" event={"ID":"4f0dd4fc-ce75-4088-be4d-664e4b5a31ea","Type":"ContainerDied","Data":"b7d75a20f6b55ff385d8e95852faafad27de83d67600bb63df0f57aceea1a925"} Feb 27 17:07:01 crc kubenswrapper[4700]: I0227 17:07:01.337614 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 27 17:07:01 crc kubenswrapper[4700]: I0227 17:07:01.343351 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 27 17:07:01 crc kubenswrapper[4700]: I0227 17:07:01.365126 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 27 17:07:01 crc kubenswrapper[4700]: I0227 17:07:01.392507 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 27 17:07:01 crc kubenswrapper[4700]: I0227 17:07:01.432607 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 27 17:07:01 crc kubenswrapper[4700]: I0227 17:07:01.545083 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 27 17:07:01 crc kubenswrapper[4700]: I0227 17:07:01.786957 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 27 17:07:02 crc kubenswrapper[4700]: I0227 17:07:02.016140 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 27 17:07:02 crc kubenswrapper[4700]: I0227 17:07:02.096869 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 27 17:07:02 crc kubenswrapper[4700]: I0227 17:07:02.234236 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 27 17:07:02 crc kubenswrapper[4700]: I0227 17:07:02.277921 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 27 17:07:02 crc kubenswrapper[4700]: I0227 17:07:02.425563 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 27 17:07:02 crc kubenswrapper[4700]: I0227 17:07:02.528598 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 27 17:07:02 crc kubenswrapper[4700]: I0227 17:07:02.538041 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536866-th9ft" Feb 27 17:07:02 crc kubenswrapper[4700]: I0227 17:07:02.645419 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9pzp\" (UniqueName: \"kubernetes.io/projected/4f0dd4fc-ce75-4088-be4d-664e4b5a31ea-kube-api-access-f9pzp\") pod \"4f0dd4fc-ce75-4088-be4d-664e4b5a31ea\" (UID: \"4f0dd4fc-ce75-4088-be4d-664e4b5a31ea\") " Feb 27 17:07:02 crc kubenswrapper[4700]: I0227 17:07:02.655637 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f0dd4fc-ce75-4088-be4d-664e4b5a31ea-kube-api-access-f9pzp" (OuterVolumeSpecName: "kube-api-access-f9pzp") pod "4f0dd4fc-ce75-4088-be4d-664e4b5a31ea" (UID: "4f0dd4fc-ce75-4088-be4d-664e4b5a31ea"). InnerVolumeSpecName "kube-api-access-f9pzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:07:02 crc kubenswrapper[4700]: I0227 17:07:02.702600 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 27 17:07:02 crc kubenswrapper[4700]: I0227 17:07:02.713652 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 27 17:07:02 crc kubenswrapper[4700]: I0227 17:07:02.746139 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 27 17:07:02 crc kubenswrapper[4700]: I0227 17:07:02.747256 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9pzp\" (UniqueName: \"kubernetes.io/projected/4f0dd4fc-ce75-4088-be4d-664e4b5a31ea-kube-api-access-f9pzp\") on node \"crc\" DevicePath \"\"" Feb 27 17:07:02 crc kubenswrapper[4700]: I0227 17:07:02.845963 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 27 17:07:02 crc kubenswrapper[4700]: I0227 17:07:02.866245 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 27 17:07:02 crc kubenswrapper[4700]: I0227 17:07:02.981238 4700 scope.go:117] "RemoveContainer" containerID="3582c66af7d192ae89576057e17a16a2426a9994835101bbc236e3316945298e" Feb 27 17:07:02 crc kubenswrapper[4700]: I0227 17:07:02.982526 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 27 17:07:03 crc kubenswrapper[4700]: I0227 17:07:03.050046 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 27 17:07:03 crc kubenswrapper[4700]: I0227 17:07:03.180289 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536866-th9ft" event={"ID":"4f0dd4fc-ce75-4088-be4d-664e4b5a31ea","Type":"ContainerDied","Data":"42b5f5d2b4e634c18bbec54d24ae3e3cac14fe4097399704344aa1a669aa56cf"} Feb 27 17:07:03 crc kubenswrapper[4700]: I0227 17:07:03.180324 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42b5f5d2b4e634c18bbec54d24ae3e3cac14fe4097399704344aa1a669aa56cf" Feb 27 17:07:03 crc kubenswrapper[4700]: I0227 17:07:03.180408 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536866-th9ft" Feb 27 17:07:03 crc kubenswrapper[4700]: I0227 17:07:03.349910 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 27 17:07:03 crc kubenswrapper[4700]: I0227 17:07:03.560834 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 27 17:07:03 crc kubenswrapper[4700]: I0227 17:07:03.565141 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 27 17:07:03 crc kubenswrapper[4700]: I0227 17:07:03.936634 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 27 17:07:03 crc kubenswrapper[4700]: I0227 17:07:03.969398 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 27 17:07:04 crc kubenswrapper[4700]: I0227 17:07:04.044268 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 27 17:07:04 crc kubenswrapper[4700]: I0227 17:07:04.188944 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-hc86c_0cc33e9e-5984-454d-880c-5187145a2c36/marketplace-operator/1.log" Feb 27 17:07:04 crc kubenswrapper[4700]: I0227 17:07:04.188996 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hc86c" event={"ID":"0cc33e9e-5984-454d-880c-5187145a2c36","Type":"ContainerStarted","Data":"56bc59b0b2d3f5679646a78732d5bb245b651d31ba05be659b7825b6bfb8078b"} Feb 27 17:07:04 crc kubenswrapper[4700]: I0227 17:07:04.189595 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-hc86c" Feb 27 17:07:04 crc kubenswrapper[4700]: I0227 17:07:04.194497 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-hc86c" Feb 27 17:07:04 crc kubenswrapper[4700]: I0227 17:07:04.247627 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 27 17:07:04 crc kubenswrapper[4700]: I0227 17:07:04.259652 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 27 17:07:04 crc kubenswrapper[4700]: I0227 17:07:04.305629 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 27 17:07:04 crc kubenswrapper[4700]: I0227 17:07:04.327691 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 27 17:07:04 crc kubenswrapper[4700]: I0227 17:07:04.327845 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 27 17:07:04 crc kubenswrapper[4700]: I0227 17:07:04.354543 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 27 17:07:04 crc kubenswrapper[4700]: I0227 17:07:04.434799 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 27 17:07:04 crc kubenswrapper[4700]: I0227 17:07:04.531664 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 27 17:07:04 crc kubenswrapper[4700]: I0227 17:07:04.536214 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 27 17:07:04 crc kubenswrapper[4700]: I0227 17:07:04.804748 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 27 17:07:05 crc kubenswrapper[4700]: I0227 17:07:05.116874 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 27 17:07:05 crc kubenswrapper[4700]: I0227 17:07:05.121936 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 27 17:07:05 crc kubenswrapper[4700]: I0227 17:07:05.185754 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 27 17:07:05 crc kubenswrapper[4700]: I0227 17:07:05.398326 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 27 17:07:05 crc kubenswrapper[4700]: I0227 17:07:05.496091 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 27 17:07:05 crc kubenswrapper[4700]: I0227 17:07:05.810441 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 27 17:07:05 crc kubenswrapper[4700]: I0227 17:07:05.885853 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 27 17:07:06 crc kubenswrapper[4700]: I0227 17:07:06.010591 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 27 17:07:06 crc kubenswrapper[4700]: I0227 17:07:06.067765 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 27 17:07:06 crc kubenswrapper[4700]: I0227 17:07:06.143248 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 27 17:07:06 crc kubenswrapper[4700]: I0227 17:07:06.536129 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 27 17:07:06 crc kubenswrapper[4700]: I0227 17:07:06.648384 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 27 17:07:06 crc kubenswrapper[4700]: I0227 17:07:06.655634 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 27 17:07:06 crc kubenswrapper[4700]: I0227 17:07:06.736139 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 27 17:07:06 crc kubenswrapper[4700]: I0227 17:07:06.767005 4700 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 27 17:07:07 crc kubenswrapper[4700]: I0227 17:07:07.077003 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 27 17:07:07 crc kubenswrapper[4700]: I0227 17:07:07.607107 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 27 17:07:07 crc kubenswrapper[4700]: I0227 17:07:07.747498 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 27 17:07:07 crc kubenswrapper[4700]: I0227 17:07:07.877027 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 27 17:07:07 crc kubenswrapper[4700]: I0227 17:07:07.934242 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 27 17:07:08 crc kubenswrapper[4700]: I0227 17:07:08.129953 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 27 17:07:08 crc kubenswrapper[4700]: I0227 17:07:08.706742 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 27 17:07:09 crc kubenswrapper[4700]: I0227 17:07:09.064202 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 27 17:07:09 crc kubenswrapper[4700]: I0227 17:07:09.135520 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 27 17:07:09 crc kubenswrapper[4700]: I0227 17:07:09.621164 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 27 17:07:09 crc kubenswrapper[4700]: I0227 17:07:09.694108 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 27 17:07:09 crc kubenswrapper[4700]: I0227 17:07:09.728384 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 27 17:07:09 crc kubenswrapper[4700]: I0227 17:07:09.929103 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 27 17:07:10 crc kubenswrapper[4700]: I0227 17:07:10.650628 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 27 17:07:10 crc kubenswrapper[4700]: I0227 17:07:10.730202 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 27 17:07:10 crc kubenswrapper[4700]: I0227 17:07:10.907184 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 27 17:07:10 crc kubenswrapper[4700]: I0227 17:07:10.924017 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 27 17:07:10 crc kubenswrapper[4700]: I0227 17:07:10.998907 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 27 17:07:11 crc kubenswrapper[4700]: I0227 17:07:11.055568 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 27 17:07:11 crc kubenswrapper[4700]: I0227 17:07:11.097947 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 27 17:07:11 crc kubenswrapper[4700]: I0227 17:07:11.118936 4700 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 27 17:07:11 crc kubenswrapper[4700]: I0227 17:07:11.224339 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 27 17:07:11 crc kubenswrapper[4700]: I0227 17:07:11.421447 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 27 17:07:11 crc kubenswrapper[4700]: I0227 17:07:11.860344 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 27 17:07:12 crc kubenswrapper[4700]: I0227 17:07:12.101339 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 27 17:07:12 crc kubenswrapper[4700]: I0227 17:07:12.257336 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 27 17:07:13 crc kubenswrapper[4700]: I0227 17:07:13.845261 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 27 17:07:14 crc kubenswrapper[4700]: I0227 17:07:14.133560 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 27 17:07:14 crc kubenswrapper[4700]: I0227 17:07:14.246448 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 27 17:07:14 crc kubenswrapper[4700]: I0227 17:07:14.785651 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 27 17:07:15 crc kubenswrapper[4700]: I0227 17:07:15.328088 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 27 17:07:15 crc kubenswrapper[4700]: I0227 17:07:15.361776 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 27 17:07:16 crc kubenswrapper[4700]: I0227 17:07:16.023014 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 27 17:07:16 crc kubenswrapper[4700]: I0227 17:07:16.031705 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 27 17:07:16 crc kubenswrapper[4700]: I0227 17:07:16.581377 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 27 17:07:36 crc kubenswrapper[4700]: I0227 17:07:36.411343 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:07:36 crc kubenswrapper[4700]: I0227 17:07:36.412130 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:08:00 crc kubenswrapper[4700]: I0227 17:08:00.137618 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536868-vf4cp"] Feb 27 17:08:00 crc kubenswrapper[4700]: E0227 17:08:00.138647 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f0dd4fc-ce75-4088-be4d-664e4b5a31ea" containerName="oc" Feb 27 17:08:00 crc kubenswrapper[4700]: I0227 17:08:00.138670 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f0dd4fc-ce75-4088-be4d-664e4b5a31ea" containerName="oc" Feb 27 17:08:00 crc kubenswrapper[4700]: I0227 17:08:00.138868 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f0dd4fc-ce75-4088-be4d-664e4b5a31ea" containerName="oc" Feb 27 17:08:00 crc kubenswrapper[4700]: I0227 17:08:00.139426 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536868-vf4cp" Feb 27 17:08:00 crc kubenswrapper[4700]: I0227 17:08:00.141948 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:08:00 crc kubenswrapper[4700]: I0227 17:08:00.142191 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d62kc" Feb 27 17:08:00 crc kubenswrapper[4700]: I0227 17:08:00.144411 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:08:00 crc kubenswrapper[4700]: I0227 17:08:00.147008 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536868-vf4cp"] Feb 27 17:08:00 crc kubenswrapper[4700]: I0227 17:08:00.293540 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5tpp\" (UniqueName: \"kubernetes.io/projected/0f466047-1e51-43ae-9fa2-a29dc4ee0c40-kube-api-access-g5tpp\") pod \"auto-csr-approver-29536868-vf4cp\" (UID: \"0f466047-1e51-43ae-9fa2-a29dc4ee0c40\") " pod="openshift-infra/auto-csr-approver-29536868-vf4cp" Feb 27 17:08:00 crc kubenswrapper[4700]: I0227 17:08:00.395222 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5tpp\" (UniqueName: \"kubernetes.io/projected/0f466047-1e51-43ae-9fa2-a29dc4ee0c40-kube-api-access-g5tpp\") pod \"auto-csr-approver-29536868-vf4cp\" (UID: \"0f466047-1e51-43ae-9fa2-a29dc4ee0c40\") " pod="openshift-infra/auto-csr-approver-29536868-vf4cp" Feb 27 17:08:00 crc kubenswrapper[4700]: I0227 17:08:00.431184 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5tpp\" (UniqueName: \"kubernetes.io/projected/0f466047-1e51-43ae-9fa2-a29dc4ee0c40-kube-api-access-g5tpp\") pod \"auto-csr-approver-29536868-vf4cp\" (UID: \"0f466047-1e51-43ae-9fa2-a29dc4ee0c40\") " pod="openshift-infra/auto-csr-approver-29536868-vf4cp" Feb 27 17:08:00 crc kubenswrapper[4700]: I0227 17:08:00.469787 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536868-vf4cp" Feb 27 17:08:00 crc kubenswrapper[4700]: I0227 17:08:00.947728 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536868-vf4cp"] Feb 27 17:08:01 crc kubenswrapper[4700]: I0227 17:08:01.570755 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536868-vf4cp" event={"ID":"0f466047-1e51-43ae-9fa2-a29dc4ee0c40","Type":"ContainerStarted","Data":"9914b0e575f121abfde54c917bc4bfadbb5ea2f393797bf443ff5426474785ac"} Feb 27 17:08:02 crc kubenswrapper[4700]: I0227 17:08:02.578650 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536868-vf4cp" event={"ID":"0f466047-1e51-43ae-9fa2-a29dc4ee0c40","Type":"ContainerStarted","Data":"7f88d980ae85d9c36365402938b1ecfccd20e654a5be3321bb42802a5863fcea"} Feb 27 17:08:02 crc kubenswrapper[4700]: I0227 17:08:02.600046 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536868-vf4cp" podStartSLOduration=1.413315055 podStartE2EDuration="2.600019497s" podCreationTimestamp="2026-02-27 17:08:00 +0000 UTC" firstStartedPulling="2026-02-27 17:08:00.964821304 +0000 UTC m=+440.950134081" lastFinishedPulling="2026-02-27 17:08:02.151525756 +0000 UTC m=+442.136838523" observedRunningTime="2026-02-27 17:08:02.598136626 +0000 UTC m=+442.583449503" watchObservedRunningTime="2026-02-27 17:08:02.600019497 +0000 UTC m=+442.585332284" Feb 27 17:08:03 crc kubenswrapper[4700]: I0227 17:08:03.586746 4700 generic.go:334] "Generic (PLEG): container finished" podID="0f466047-1e51-43ae-9fa2-a29dc4ee0c40" containerID="7f88d980ae85d9c36365402938b1ecfccd20e654a5be3321bb42802a5863fcea" exitCode=0 Feb 27 17:08:03 crc kubenswrapper[4700]: I0227 17:08:03.586819 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536868-vf4cp" event={"ID":"0f466047-1e51-43ae-9fa2-a29dc4ee0c40","Type":"ContainerDied","Data":"7f88d980ae85d9c36365402938b1ecfccd20e654a5be3321bb42802a5863fcea"} Feb 27 17:08:04 crc kubenswrapper[4700]: I0227 17:08:04.883760 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536868-vf4cp" Feb 27 17:08:05 crc kubenswrapper[4700]: I0227 17:08:05.065345 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5tpp\" (UniqueName: \"kubernetes.io/projected/0f466047-1e51-43ae-9fa2-a29dc4ee0c40-kube-api-access-g5tpp\") pod \"0f466047-1e51-43ae-9fa2-a29dc4ee0c40\" (UID: \"0f466047-1e51-43ae-9fa2-a29dc4ee0c40\") " Feb 27 17:08:05 crc kubenswrapper[4700]: I0227 17:08:05.073258 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f466047-1e51-43ae-9fa2-a29dc4ee0c40-kube-api-access-g5tpp" (OuterVolumeSpecName: "kube-api-access-g5tpp") pod "0f466047-1e51-43ae-9fa2-a29dc4ee0c40" (UID: "0f466047-1e51-43ae-9fa2-a29dc4ee0c40"). InnerVolumeSpecName "kube-api-access-g5tpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:08:05 crc kubenswrapper[4700]: I0227 17:08:05.167615 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5tpp\" (UniqueName: \"kubernetes.io/projected/0f466047-1e51-43ae-9fa2-a29dc4ee0c40-kube-api-access-g5tpp\") on node \"crc\" DevicePath \"\"" Feb 27 17:08:05 crc kubenswrapper[4700]: I0227 17:08:05.602270 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536868-vf4cp" event={"ID":"0f466047-1e51-43ae-9fa2-a29dc4ee0c40","Type":"ContainerDied","Data":"9914b0e575f121abfde54c917bc4bfadbb5ea2f393797bf443ff5426474785ac"} Feb 27 17:08:05 crc kubenswrapper[4700]: I0227 17:08:05.602381 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9914b0e575f121abfde54c917bc4bfadbb5ea2f393797bf443ff5426474785ac" Feb 27 17:08:05 crc kubenswrapper[4700]: I0227 17:08:05.602340 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536868-vf4cp" Feb 27 17:08:06 crc kubenswrapper[4700]: I0227 17:08:06.411131 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:08:06 crc kubenswrapper[4700]: I0227 17:08:06.411449 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.407852 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5n6qh"] Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.411309 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5n6qh" podUID="a72bdcb3-42e2-4a45-8c98-7b6e97afabab" containerName="registry-server" containerID="cri-o://3ce0b02ad05b57bf02c6d6f8289417e613941767abac9195e2d52912363ab7e3" gracePeriod=30 Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.416094 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9t75w"] Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.416445 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9t75w" podUID="89845e2b-1804-445b-8462-36b2350ae663" containerName="registry-server" containerID="cri-o://d3dd09bfd09d07207b3bb46e8ad9c0e295fa697c60f2c5f65b86bcd9a9e755c8" gracePeriod=30 Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.426008 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hc86c"] Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.426789 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-hc86c" podUID="0cc33e9e-5984-454d-880c-5187145a2c36" containerName="marketplace-operator" containerID="cri-o://56bc59b0b2d3f5679646a78732d5bb245b651d31ba05be659b7825b6bfb8078b" gracePeriod=30 Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.451697 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gpb6h"] Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.452029 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gpb6h" podUID="8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2" containerName="registry-server" containerID="cri-o://ecda766de52f66adfbef2adbf1f8b9e59f9348d240a1d0cdf36242fb57e9d755" gracePeriod=30 Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.467927 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wlsk4"] Feb 27 17:08:13 crc kubenswrapper[4700]: E0227 17:08:13.468185 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f466047-1e51-43ae-9fa2-a29dc4ee0c40" containerName="oc" Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.468213 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f466047-1e51-43ae-9fa2-a29dc4ee0c40" containerName="oc" Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.468331 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f466047-1e51-43ae-9fa2-a29dc4ee0c40" containerName="oc" Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.468773 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wlsk4" Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.473647 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ltq9s"] Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.473912 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ltq9s" podUID="fe1b00ba-9936-45d5-9a6e-f4f93ce09b44" containerName="registry-server" containerID="cri-o://1906e32028c1280a1bcf8c6e2eb3953f22b5cbeac74a3f7e0dbd3515469080b3" gracePeriod=30 Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.478735 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wlsk4"] Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.542658 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmqk6\" (UniqueName: \"kubernetes.io/projected/aca7409e-ab2c-4d9b-a775-e82f2cd958d5-kube-api-access-vmqk6\") pod \"marketplace-operator-79b997595-wlsk4\" (UID: \"aca7409e-ab2c-4d9b-a775-e82f2cd958d5\") " pod="openshift-marketplace/marketplace-operator-79b997595-wlsk4" Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.542729 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/aca7409e-ab2c-4d9b-a775-e82f2cd958d5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wlsk4\" (UID: \"aca7409e-ab2c-4d9b-a775-e82f2cd958d5\") " pod="openshift-marketplace/marketplace-operator-79b997595-wlsk4" Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.542768 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aca7409e-ab2c-4d9b-a775-e82f2cd958d5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wlsk4\" (UID: \"aca7409e-ab2c-4d9b-a775-e82f2cd958d5\") " pod="openshift-marketplace/marketplace-operator-79b997595-wlsk4" Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.644856 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmqk6\" (UniqueName: \"kubernetes.io/projected/aca7409e-ab2c-4d9b-a775-e82f2cd958d5-kube-api-access-vmqk6\") pod \"marketplace-operator-79b997595-wlsk4\" (UID: \"aca7409e-ab2c-4d9b-a775-e82f2cd958d5\") " pod="openshift-marketplace/marketplace-operator-79b997595-wlsk4" Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.644916 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/aca7409e-ab2c-4d9b-a775-e82f2cd958d5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wlsk4\" (UID: \"aca7409e-ab2c-4d9b-a775-e82f2cd958d5\") " pod="openshift-marketplace/marketplace-operator-79b997595-wlsk4" Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.644954 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aca7409e-ab2c-4d9b-a775-e82f2cd958d5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wlsk4\" (UID: \"aca7409e-ab2c-4d9b-a775-e82f2cd958d5\") " pod="openshift-marketplace/marketplace-operator-79b997595-wlsk4" Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.646658 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aca7409e-ab2c-4d9b-a775-e82f2cd958d5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wlsk4\" (UID: \"aca7409e-ab2c-4d9b-a775-e82f2cd958d5\") " pod="openshift-marketplace/marketplace-operator-79b997595-wlsk4" Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.652161 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/aca7409e-ab2c-4d9b-a775-e82f2cd958d5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wlsk4\" (UID: \"aca7409e-ab2c-4d9b-a775-e82f2cd958d5\") " pod="openshift-marketplace/marketplace-operator-79b997595-wlsk4" Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.663366 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmqk6\" (UniqueName: \"kubernetes.io/projected/aca7409e-ab2c-4d9b-a775-e82f2cd958d5-kube-api-access-vmqk6\") pod \"marketplace-operator-79b997595-wlsk4\" (UID: \"aca7409e-ab2c-4d9b-a775-e82f2cd958d5\") " pod="openshift-marketplace/marketplace-operator-79b997595-wlsk4" Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.675220 4700 generic.go:334] "Generic (PLEG): container finished" podID="a72bdcb3-42e2-4a45-8c98-7b6e97afabab" containerID="3ce0b02ad05b57bf02c6d6f8289417e613941767abac9195e2d52912363ab7e3" exitCode=0 Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.675306 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5n6qh" event={"ID":"a72bdcb3-42e2-4a45-8c98-7b6e97afabab","Type":"ContainerDied","Data":"3ce0b02ad05b57bf02c6d6f8289417e613941767abac9195e2d52912363ab7e3"} Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.685798 4700 generic.go:334] "Generic (PLEG): container finished" podID="8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2" containerID="ecda766de52f66adfbef2adbf1f8b9e59f9348d240a1d0cdf36242fb57e9d755" exitCode=0 Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.685868 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gpb6h" event={"ID":"8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2","Type":"ContainerDied","Data":"ecda766de52f66adfbef2adbf1f8b9e59f9348d240a1d0cdf36242fb57e9d755"} Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.688434 4700 generic.go:334] "Generic (PLEG): container finished" podID="fe1b00ba-9936-45d5-9a6e-f4f93ce09b44" containerID="1906e32028c1280a1bcf8c6e2eb3953f22b5cbeac74a3f7e0dbd3515469080b3" exitCode=0 Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.688505 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ltq9s" event={"ID":"fe1b00ba-9936-45d5-9a6e-f4f93ce09b44","Type":"ContainerDied","Data":"1906e32028c1280a1bcf8c6e2eb3953f22b5cbeac74a3f7e0dbd3515469080b3"} Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.692601 4700 generic.go:334] "Generic (PLEG): container finished" podID="89845e2b-1804-445b-8462-36b2350ae663" containerID="d3dd09bfd09d07207b3bb46e8ad9c0e295fa697c60f2c5f65b86bcd9a9e755c8" exitCode=0 Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.692639 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9t75w" event={"ID":"89845e2b-1804-445b-8462-36b2350ae663","Type":"ContainerDied","Data":"d3dd09bfd09d07207b3bb46e8ad9c0e295fa697c60f2c5f65b86bcd9a9e755c8"} Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.696144 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-hc86c_0cc33e9e-5984-454d-880c-5187145a2c36/marketplace-operator/1.log" Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.696194 4700 generic.go:334] "Generic (PLEG): container finished" podID="0cc33e9e-5984-454d-880c-5187145a2c36" containerID="56bc59b0b2d3f5679646a78732d5bb245b651d31ba05be659b7825b6bfb8078b" exitCode=0 Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.696227 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hc86c" event={"ID":"0cc33e9e-5984-454d-880c-5187145a2c36","Type":"ContainerDied","Data":"56bc59b0b2d3f5679646a78732d5bb245b651d31ba05be659b7825b6bfb8078b"} Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.696265 4700 scope.go:117] "RemoveContainer" containerID="3582c66af7d192ae89576057e17a16a2426a9994835101bbc236e3316945298e" Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.852477 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wlsk4" Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.856587 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9t75w" Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.860383 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5n6qh" Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.878895 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-hc86c" Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.946995 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gpb6h" Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.948618 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89845e2b-1804-445b-8462-36b2350ae663-utilities\") pod \"89845e2b-1804-445b-8462-36b2350ae663\" (UID: \"89845e2b-1804-445b-8462-36b2350ae663\") " Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.948685 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4kwkl\" (UniqueName: \"kubernetes.io/projected/a72bdcb3-42e2-4a45-8c98-7b6e97afabab-kube-api-access-4kwkl\") pod \"a72bdcb3-42e2-4a45-8c98-7b6e97afabab\" (UID: \"a72bdcb3-42e2-4a45-8c98-7b6e97afabab\") " Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.948753 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qphmr\" (UniqueName: \"kubernetes.io/projected/0cc33e9e-5984-454d-880c-5187145a2c36-kube-api-access-qphmr\") pod \"0cc33e9e-5984-454d-880c-5187145a2c36\" (UID: \"0cc33e9e-5984-454d-880c-5187145a2c36\") " Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.948789 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0cc33e9e-5984-454d-880c-5187145a2c36-marketplace-trusted-ca\") pod \"0cc33e9e-5984-454d-880c-5187145a2c36\" (UID: \"0cc33e9e-5984-454d-880c-5187145a2c36\") " Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.948812 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89845e2b-1804-445b-8462-36b2350ae663-catalog-content\") pod \"89845e2b-1804-445b-8462-36b2350ae663\" (UID: \"89845e2b-1804-445b-8462-36b2350ae663\") " Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.948836 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0cc33e9e-5984-454d-880c-5187145a2c36-marketplace-operator-metrics\") pod \"0cc33e9e-5984-454d-880c-5187145a2c36\" (UID: \"0cc33e9e-5984-454d-880c-5187145a2c36\") " Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.948855 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qsvvr\" (UniqueName: \"kubernetes.io/projected/89845e2b-1804-445b-8462-36b2350ae663-kube-api-access-qsvvr\") pod \"89845e2b-1804-445b-8462-36b2350ae663\" (UID: \"89845e2b-1804-445b-8462-36b2350ae663\") " Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.948870 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a72bdcb3-42e2-4a45-8c98-7b6e97afabab-utilities\") pod \"a72bdcb3-42e2-4a45-8c98-7b6e97afabab\" (UID: \"a72bdcb3-42e2-4a45-8c98-7b6e97afabab\") " Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.948894 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a72bdcb3-42e2-4a45-8c98-7b6e97afabab-catalog-content\") pod \"a72bdcb3-42e2-4a45-8c98-7b6e97afabab\" (UID: \"a72bdcb3-42e2-4a45-8c98-7b6e97afabab\") " Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.950185 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cc33e9e-5984-454d-880c-5187145a2c36-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "0cc33e9e-5984-454d-880c-5187145a2c36" (UID: "0cc33e9e-5984-454d-880c-5187145a2c36"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.950268 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a72bdcb3-42e2-4a45-8c98-7b6e97afabab-utilities" (OuterVolumeSpecName: "utilities") pod "a72bdcb3-42e2-4a45-8c98-7b6e97afabab" (UID: "a72bdcb3-42e2-4a45-8c98-7b6e97afabab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.950586 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89845e2b-1804-445b-8462-36b2350ae663-utilities" (OuterVolumeSpecName: "utilities") pod "89845e2b-1804-445b-8462-36b2350ae663" (UID: "89845e2b-1804-445b-8462-36b2350ae663"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.953191 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a72bdcb3-42e2-4a45-8c98-7b6e97afabab-kube-api-access-4kwkl" (OuterVolumeSpecName: "kube-api-access-4kwkl") pod "a72bdcb3-42e2-4a45-8c98-7b6e97afabab" (UID: "a72bdcb3-42e2-4a45-8c98-7b6e97afabab"). InnerVolumeSpecName "kube-api-access-4kwkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.953542 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cc33e9e-5984-454d-880c-5187145a2c36-kube-api-access-qphmr" (OuterVolumeSpecName: "kube-api-access-qphmr") pod "0cc33e9e-5984-454d-880c-5187145a2c36" (UID: "0cc33e9e-5984-454d-880c-5187145a2c36"). InnerVolumeSpecName "kube-api-access-qphmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.958905 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cc33e9e-5984-454d-880c-5187145a2c36-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "0cc33e9e-5984-454d-880c-5187145a2c36" (UID: "0cc33e9e-5984-454d-880c-5187145a2c36"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.960565 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89845e2b-1804-445b-8462-36b2350ae663-kube-api-access-qsvvr" (OuterVolumeSpecName: "kube-api-access-qsvvr") pod "89845e2b-1804-445b-8462-36b2350ae663" (UID: "89845e2b-1804-445b-8462-36b2350ae663"). InnerVolumeSpecName "kube-api-access-qsvvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:08:13 crc kubenswrapper[4700]: I0227 17:08:13.960813 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ltq9s" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.014649 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a72bdcb3-42e2-4a45-8c98-7b6e97afabab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a72bdcb3-42e2-4a45-8c98-7b6e97afabab" (UID: "a72bdcb3-42e2-4a45-8c98-7b6e97afabab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.043588 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89845e2b-1804-445b-8462-36b2350ae663-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "89845e2b-1804-445b-8462-36b2350ae663" (UID: "89845e2b-1804-445b-8462-36b2350ae663"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.050301 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hq655\" (UniqueName: \"kubernetes.io/projected/8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2-kube-api-access-hq655\") pod \"8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2\" (UID: \"8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2\") " Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.050417 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe1b00ba-9936-45d5-9a6e-f4f93ce09b44-utilities\") pod \"fe1b00ba-9936-45d5-9a6e-f4f93ce09b44\" (UID: \"fe1b00ba-9936-45d5-9a6e-f4f93ce09b44\") " Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.050450 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2-utilities\") pod \"8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2\" (UID: \"8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2\") " Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.050511 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2-catalog-content\") pod \"8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2\" (UID: \"8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2\") " Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.050586 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe1b00ba-9936-45d5-9a6e-f4f93ce09b44-catalog-content\") pod \"fe1b00ba-9936-45d5-9a6e-f4f93ce09b44\" (UID: \"fe1b00ba-9936-45d5-9a6e-f4f93ce09b44\") " Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.050606 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fppfr\" (UniqueName: \"kubernetes.io/projected/fe1b00ba-9936-45d5-9a6e-f4f93ce09b44-kube-api-access-fppfr\") pod \"fe1b00ba-9936-45d5-9a6e-f4f93ce09b44\" (UID: \"fe1b00ba-9936-45d5-9a6e-f4f93ce09b44\") " Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.050850 4700 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0cc33e9e-5984-454d-880c-5187145a2c36-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.050863 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89845e2b-1804-445b-8462-36b2350ae663-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.050872 4700 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0cc33e9e-5984-454d-880c-5187145a2c36-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.050882 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a72bdcb3-42e2-4a45-8c98-7b6e97afabab-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.050891 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qsvvr\" (UniqueName: \"kubernetes.io/projected/89845e2b-1804-445b-8462-36b2350ae663-kube-api-access-qsvvr\") on node \"crc\" DevicePath \"\"" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.050900 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a72bdcb3-42e2-4a45-8c98-7b6e97afabab-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.050911 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89845e2b-1804-445b-8462-36b2350ae663-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.050920 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4kwkl\" (UniqueName: \"kubernetes.io/projected/a72bdcb3-42e2-4a45-8c98-7b6e97afabab-kube-api-access-4kwkl\") on node \"crc\" DevicePath \"\"" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.050931 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qphmr\" (UniqueName: \"kubernetes.io/projected/0cc33e9e-5984-454d-880c-5187145a2c36-kube-api-access-qphmr\") on node \"crc\" DevicePath \"\"" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.051477 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe1b00ba-9936-45d5-9a6e-f4f93ce09b44-utilities" (OuterVolumeSpecName: "utilities") pod "fe1b00ba-9936-45d5-9a6e-f4f93ce09b44" (UID: "fe1b00ba-9936-45d5-9a6e-f4f93ce09b44"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.053087 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2-utilities" (OuterVolumeSpecName: "utilities") pod "8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2" (UID: "8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.053528 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe1b00ba-9936-45d5-9a6e-f4f93ce09b44-kube-api-access-fppfr" (OuterVolumeSpecName: "kube-api-access-fppfr") pod "fe1b00ba-9936-45d5-9a6e-f4f93ce09b44" (UID: "fe1b00ba-9936-45d5-9a6e-f4f93ce09b44"). InnerVolumeSpecName "kube-api-access-fppfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.054032 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2-kube-api-access-hq655" (OuterVolumeSpecName: "kube-api-access-hq655") pod "8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2" (UID: "8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2"). InnerVolumeSpecName "kube-api-access-hq655". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.062661 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wlsk4"] Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.092253 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2" (UID: "8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.151676 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe1b00ba-9936-45d5-9a6e-f4f93ce09b44-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.151721 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.151731 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.151743 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fppfr\" (UniqueName: \"kubernetes.io/projected/fe1b00ba-9936-45d5-9a6e-f4f93ce09b44-kube-api-access-fppfr\") on node \"crc\" DevicePath \"\"" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.151752 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hq655\" (UniqueName: \"kubernetes.io/projected/8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2-kube-api-access-hq655\") on node \"crc\" DevicePath \"\"" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.167749 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe1b00ba-9936-45d5-9a6e-f4f93ce09b44-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fe1b00ba-9936-45d5-9a6e-f4f93ce09b44" (UID: "fe1b00ba-9936-45d5-9a6e-f4f93ce09b44"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.253081 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe1b00ba-9936-45d5-9a6e-f4f93ce09b44-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.702331 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wlsk4" event={"ID":"aca7409e-ab2c-4d9b-a775-e82f2cd958d5","Type":"ContainerStarted","Data":"6444c996d486b3d2e541e173b4aa9f264d5a5c8bec9e15880d3617c06ab7dcef"} Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.702395 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wlsk4" event={"ID":"aca7409e-ab2c-4d9b-a775-e82f2cd958d5","Type":"ContainerStarted","Data":"40d4efd9b526c2dc111fe61df04582718da95a4a9d9a6f355a42d59df3653491"} Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.702742 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-wlsk4" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.704392 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9t75w" event={"ID":"89845e2b-1804-445b-8462-36b2350ae663","Type":"ContainerDied","Data":"d8d53a3427897d1d1b8664f4d398b544f13f89bf8d97f79d4d2f9a125ecc3d8a"} Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.704441 4700 scope.go:117] "RemoveContainer" containerID="d3dd09bfd09d07207b3bb46e8ad9c0e295fa697c60f2c5f65b86bcd9a9e755c8" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.704613 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9t75w" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.707217 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hc86c" event={"ID":"0cc33e9e-5984-454d-880c-5187145a2c36","Type":"ContainerDied","Data":"743e746f4f36b8dd56462dcd3823d64d50abdf0a2e6e2c23b57905962aa3d276"} Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.707351 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-hc86c" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.709326 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-wlsk4" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.710351 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5n6qh" event={"ID":"a72bdcb3-42e2-4a45-8c98-7b6e97afabab","Type":"ContainerDied","Data":"dc416ff2c3de8978b2e6f3e281afa5c3b969ec0e08913bacb226109e67ad1237"} Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.710556 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5n6qh" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.716339 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gpb6h" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.716322 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gpb6h" event={"ID":"8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2","Type":"ContainerDied","Data":"864cd6a11e2a2f8e31a8c1fe61b497b560567f12393e162ef383a2df5b73c9c5"} Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.725512 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ltq9s" event={"ID":"fe1b00ba-9936-45d5-9a6e-f4f93ce09b44","Type":"ContainerDied","Data":"17dcd3d7a115422a23c0178e740471b6249e855ce09dd68bc02a4b6763747218"} Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.727587 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ltq9s" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.728360 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-wlsk4" podStartSLOduration=1.728349228 podStartE2EDuration="1.728349228s" podCreationTimestamp="2026-02-27 17:08:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:08:14.728275406 +0000 UTC m=+454.713588163" watchObservedRunningTime="2026-02-27 17:08:14.728349228 +0000 UTC m=+454.713661975" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.745860 4700 scope.go:117] "RemoveContainer" containerID="ec5bde0329f6c2c6c159ece770d807e1c316d222e5db57f90e3841fbf5deb8a1" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.770438 4700 scope.go:117] "RemoveContainer" containerID="b75bbb481dfa12e9f0984d0c52f051571dc47d60b84622ed1ebe8e65b6de2783" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.797128 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9t75w"] Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.800387 4700 scope.go:117] "RemoveContainer" containerID="56bc59b0b2d3f5679646a78732d5bb245b651d31ba05be659b7825b6bfb8078b" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.803988 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9t75w"] Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.818630 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gpb6h"] Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.822665 4700 scope.go:117] "RemoveContainer" containerID="3ce0b02ad05b57bf02c6d6f8289417e613941767abac9195e2d52912363ab7e3" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.825791 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gpb6h"] Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.834586 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hc86c"] Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.841126 4700 scope.go:117] "RemoveContainer" containerID="b3893d88bbfde9f99743b4604b171f418e7a77f1939d38d528cb0a8c6f03f1de" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.843926 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hc86c"] Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.849550 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5n6qh"] Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.853620 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5n6qh"] Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.859493 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ltq9s"] Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.859677 4700 scope.go:117] "RemoveContainer" containerID="59d9e1f169d44712ad38898168ed8a7ece3561c0bc8f0930a718b175778cad8c" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.863203 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ltq9s"] Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.881404 4700 scope.go:117] "RemoveContainer" containerID="ecda766de52f66adfbef2adbf1f8b9e59f9348d240a1d0cdf36242fb57e9d755" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.896368 4700 scope.go:117] "RemoveContainer" containerID="fa9e2fbb1c27f16e3b6698982e627ab8b7cbee499d71e03e21ab93e57fcb9d66" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.910301 4700 scope.go:117] "RemoveContainer" containerID="2bfba51a7a76772d0f6968119fb575eba783f431af139fbda9ff262e0010306d" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.926478 4700 scope.go:117] "RemoveContainer" containerID="1906e32028c1280a1bcf8c6e2eb3953f22b5cbeac74a3f7e0dbd3515469080b3" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.941948 4700 scope.go:117] "RemoveContainer" containerID="80839cb344fc98980a3f8b13fa2f1cd29349f2491d81ca794b0062991465053a" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.956785 4700 scope.go:117] "RemoveContainer" containerID="f1e5ff4267abdf9ff0078425b5b6b2499ae863eb907438b538f3e53c30cd262a" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.986241 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cc33e9e-5984-454d-880c-5187145a2c36" path="/var/lib/kubelet/pods/0cc33e9e-5984-454d-880c-5187145a2c36/volumes" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.986725 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89845e2b-1804-445b-8462-36b2350ae663" path="/var/lib/kubelet/pods/89845e2b-1804-445b-8462-36b2350ae663/volumes" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.987252 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2" path="/var/lib/kubelet/pods/8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2/volumes" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.988189 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a72bdcb3-42e2-4a45-8c98-7b6e97afabab" path="/var/lib/kubelet/pods/a72bdcb3-42e2-4a45-8c98-7b6e97afabab/volumes" Feb 27 17:08:14 crc kubenswrapper[4700]: I0227 17:08:14.988727 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe1b00ba-9936-45d5-9a6e-f4f93ce09b44" path="/var/lib/kubelet/pods/fe1b00ba-9936-45d5-9a6e-f4f93ce09b44/volumes" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.512070 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-sl4bz"] Feb 27 17:08:15 crc kubenswrapper[4700]: E0227 17:08:15.512669 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cc33e9e-5984-454d-880c-5187145a2c36" containerName="marketplace-operator" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.512696 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cc33e9e-5984-454d-880c-5187145a2c36" containerName="marketplace-operator" Feb 27 17:08:15 crc kubenswrapper[4700]: E0227 17:08:15.512713 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe1b00ba-9936-45d5-9a6e-f4f93ce09b44" containerName="registry-server" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.512726 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe1b00ba-9936-45d5-9a6e-f4f93ce09b44" containerName="registry-server" Feb 27 17:08:15 crc kubenswrapper[4700]: E0227 17:08:15.512746 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a72bdcb3-42e2-4a45-8c98-7b6e97afabab" containerName="registry-server" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.512760 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="a72bdcb3-42e2-4a45-8c98-7b6e97afabab" containerName="registry-server" Feb 27 17:08:15 crc kubenswrapper[4700]: E0227 17:08:15.512776 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89845e2b-1804-445b-8462-36b2350ae663" containerName="extract-content" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.512790 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="89845e2b-1804-445b-8462-36b2350ae663" containerName="extract-content" Feb 27 17:08:15 crc kubenswrapper[4700]: E0227 17:08:15.512807 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89845e2b-1804-445b-8462-36b2350ae663" containerName="extract-utilities" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.512820 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="89845e2b-1804-445b-8462-36b2350ae663" containerName="extract-utilities" Feb 27 17:08:15 crc kubenswrapper[4700]: E0227 17:08:15.512834 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2" containerName="extract-utilities" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.512846 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2" containerName="extract-utilities" Feb 27 17:08:15 crc kubenswrapper[4700]: E0227 17:08:15.512869 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89845e2b-1804-445b-8462-36b2350ae663" containerName="registry-server" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.512881 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="89845e2b-1804-445b-8462-36b2350ae663" containerName="registry-server" Feb 27 17:08:15 crc kubenswrapper[4700]: E0227 17:08:15.512896 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2" containerName="registry-server" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.512908 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2" containerName="registry-server" Feb 27 17:08:15 crc kubenswrapper[4700]: E0227 17:08:15.512928 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cc33e9e-5984-454d-880c-5187145a2c36" containerName="marketplace-operator" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.512940 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cc33e9e-5984-454d-880c-5187145a2c36" containerName="marketplace-operator" Feb 27 17:08:15 crc kubenswrapper[4700]: E0227 17:08:15.512958 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2" containerName="extract-content" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.512970 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2" containerName="extract-content" Feb 27 17:08:15 crc kubenswrapper[4700]: E0227 17:08:15.512982 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cc33e9e-5984-454d-880c-5187145a2c36" containerName="marketplace-operator" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.512994 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cc33e9e-5984-454d-880c-5187145a2c36" containerName="marketplace-operator" Feb 27 17:08:15 crc kubenswrapper[4700]: E0227 17:08:15.513015 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe1b00ba-9936-45d5-9a6e-f4f93ce09b44" containerName="extract-content" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.513026 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe1b00ba-9936-45d5-9a6e-f4f93ce09b44" containerName="extract-content" Feb 27 17:08:15 crc kubenswrapper[4700]: E0227 17:08:15.513041 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a72bdcb3-42e2-4a45-8c98-7b6e97afabab" containerName="extract-content" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.513054 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="a72bdcb3-42e2-4a45-8c98-7b6e97afabab" containerName="extract-content" Feb 27 17:08:15 crc kubenswrapper[4700]: E0227 17:08:15.513072 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe1b00ba-9936-45d5-9a6e-f4f93ce09b44" containerName="extract-utilities" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.513084 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe1b00ba-9936-45d5-9a6e-f4f93ce09b44" containerName="extract-utilities" Feb 27 17:08:15 crc kubenswrapper[4700]: E0227 17:08:15.513104 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a72bdcb3-42e2-4a45-8c98-7b6e97afabab" containerName="extract-utilities" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.513115 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="a72bdcb3-42e2-4a45-8c98-7b6e97afabab" containerName="extract-utilities" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.513275 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="a72bdcb3-42e2-4a45-8c98-7b6e97afabab" containerName="registry-server" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.513299 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cbbb8e8-05e4-4cb2-8500-91f8eb7212a2" containerName="registry-server" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.513318 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="89845e2b-1804-445b-8462-36b2350ae663" containerName="registry-server" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.513334 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cc33e9e-5984-454d-880c-5187145a2c36" containerName="marketplace-operator" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.513351 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cc33e9e-5984-454d-880c-5187145a2c36" containerName="marketplace-operator" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.513367 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cc33e9e-5984-454d-880c-5187145a2c36" containerName="marketplace-operator" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.513383 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe1b00ba-9936-45d5-9a6e-f4f93ce09b44" containerName="registry-server" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.514528 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sl4bz" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.516571 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sl4bz"] Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.517615 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.572055 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53351da8-d63b-4af8-912e-9a7aa5da939a-catalog-content\") pod \"redhat-marketplace-sl4bz\" (UID: \"53351da8-d63b-4af8-912e-9a7aa5da939a\") " pod="openshift-marketplace/redhat-marketplace-sl4bz" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.572127 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53351da8-d63b-4af8-912e-9a7aa5da939a-utilities\") pod \"redhat-marketplace-sl4bz\" (UID: \"53351da8-d63b-4af8-912e-9a7aa5da939a\") " pod="openshift-marketplace/redhat-marketplace-sl4bz" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.572163 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m62fb\" (UniqueName: \"kubernetes.io/projected/53351da8-d63b-4af8-912e-9a7aa5da939a-kube-api-access-m62fb\") pod \"redhat-marketplace-sl4bz\" (UID: \"53351da8-d63b-4af8-912e-9a7aa5da939a\") " pod="openshift-marketplace/redhat-marketplace-sl4bz" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.673545 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m62fb\" (UniqueName: \"kubernetes.io/projected/53351da8-d63b-4af8-912e-9a7aa5da939a-kube-api-access-m62fb\") pod \"redhat-marketplace-sl4bz\" (UID: \"53351da8-d63b-4af8-912e-9a7aa5da939a\") " pod="openshift-marketplace/redhat-marketplace-sl4bz" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.673630 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53351da8-d63b-4af8-912e-9a7aa5da939a-catalog-content\") pod \"redhat-marketplace-sl4bz\" (UID: \"53351da8-d63b-4af8-912e-9a7aa5da939a\") " pod="openshift-marketplace/redhat-marketplace-sl4bz" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.673669 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53351da8-d63b-4af8-912e-9a7aa5da939a-utilities\") pod \"redhat-marketplace-sl4bz\" (UID: \"53351da8-d63b-4af8-912e-9a7aa5da939a\") " pod="openshift-marketplace/redhat-marketplace-sl4bz" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.674511 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53351da8-d63b-4af8-912e-9a7aa5da939a-utilities\") pod \"redhat-marketplace-sl4bz\" (UID: \"53351da8-d63b-4af8-912e-9a7aa5da939a\") " pod="openshift-marketplace/redhat-marketplace-sl4bz" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.674558 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53351da8-d63b-4af8-912e-9a7aa5da939a-catalog-content\") pod \"redhat-marketplace-sl4bz\" (UID: \"53351da8-d63b-4af8-912e-9a7aa5da939a\") " pod="openshift-marketplace/redhat-marketplace-sl4bz" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.698102 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m62fb\" (UniqueName: \"kubernetes.io/projected/53351da8-d63b-4af8-912e-9a7aa5da939a-kube-api-access-m62fb\") pod \"redhat-marketplace-sl4bz\" (UID: \"53351da8-d63b-4af8-912e-9a7aa5da939a\") " pod="openshift-marketplace/redhat-marketplace-sl4bz" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.707020 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-q8xmq"] Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.708189 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q8xmq" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.711141 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.717767 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q8xmq"] Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.844037 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sl4bz" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.881051 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9dae327-14da-40ce-8e98-b29453b0352f-utilities\") pod \"redhat-operators-q8xmq\" (UID: \"f9dae327-14da-40ce-8e98-b29453b0352f\") " pod="openshift-marketplace/redhat-operators-q8xmq" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.881288 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rxnc\" (UniqueName: \"kubernetes.io/projected/f9dae327-14da-40ce-8e98-b29453b0352f-kube-api-access-6rxnc\") pod \"redhat-operators-q8xmq\" (UID: \"f9dae327-14da-40ce-8e98-b29453b0352f\") " pod="openshift-marketplace/redhat-operators-q8xmq" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.881529 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9dae327-14da-40ce-8e98-b29453b0352f-catalog-content\") pod \"redhat-operators-q8xmq\" (UID: \"f9dae327-14da-40ce-8e98-b29453b0352f\") " pod="openshift-marketplace/redhat-operators-q8xmq" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.982242 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9dae327-14da-40ce-8e98-b29453b0352f-utilities\") pod \"redhat-operators-q8xmq\" (UID: \"f9dae327-14da-40ce-8e98-b29453b0352f\") " pod="openshift-marketplace/redhat-operators-q8xmq" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.982627 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rxnc\" (UniqueName: \"kubernetes.io/projected/f9dae327-14da-40ce-8e98-b29453b0352f-kube-api-access-6rxnc\") pod \"redhat-operators-q8xmq\" (UID: \"f9dae327-14da-40ce-8e98-b29453b0352f\") " pod="openshift-marketplace/redhat-operators-q8xmq" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.982692 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9dae327-14da-40ce-8e98-b29453b0352f-catalog-content\") pod \"redhat-operators-q8xmq\" (UID: \"f9dae327-14da-40ce-8e98-b29453b0352f\") " pod="openshift-marketplace/redhat-operators-q8xmq" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.983106 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9dae327-14da-40ce-8e98-b29453b0352f-utilities\") pod \"redhat-operators-q8xmq\" (UID: \"f9dae327-14da-40ce-8e98-b29453b0352f\") " pod="openshift-marketplace/redhat-operators-q8xmq" Feb 27 17:08:15 crc kubenswrapper[4700]: I0227 17:08:15.983125 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9dae327-14da-40ce-8e98-b29453b0352f-catalog-content\") pod \"redhat-operators-q8xmq\" (UID: \"f9dae327-14da-40ce-8e98-b29453b0352f\") " pod="openshift-marketplace/redhat-operators-q8xmq" Feb 27 17:08:16 crc kubenswrapper[4700]: I0227 17:08:16.013001 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rxnc\" (UniqueName: \"kubernetes.io/projected/f9dae327-14da-40ce-8e98-b29453b0352f-kube-api-access-6rxnc\") pod \"redhat-operators-q8xmq\" (UID: \"f9dae327-14da-40ce-8e98-b29453b0352f\") " pod="openshift-marketplace/redhat-operators-q8xmq" Feb 27 17:08:16 crc kubenswrapper[4700]: I0227 17:08:16.030120 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sl4bz"] Feb 27 17:08:16 crc kubenswrapper[4700]: W0227 17:08:16.039393 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod53351da8_d63b_4af8_912e_9a7aa5da939a.slice/crio-5c7ac57f85eeb4c04ec375eac19b32744866d2bab95921a2cd767a8ab4bf2bcc WatchSource:0}: Error finding container 5c7ac57f85eeb4c04ec375eac19b32744866d2bab95921a2cd767a8ab4bf2bcc: Status 404 returned error can't find the container with id 5c7ac57f85eeb4c04ec375eac19b32744866d2bab95921a2cd767a8ab4bf2bcc Feb 27 17:08:16 crc kubenswrapper[4700]: I0227 17:08:16.111020 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q8xmq" Feb 27 17:08:16 crc kubenswrapper[4700]: I0227 17:08:16.297873 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q8xmq"] Feb 27 17:08:16 crc kubenswrapper[4700]: W0227 17:08:16.304253 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf9dae327_14da_40ce_8e98_b29453b0352f.slice/crio-a3ef2aa4aaff6c47c84e22acfdb979a903948d7c53564f891c79dd927117742f WatchSource:0}: Error finding container a3ef2aa4aaff6c47c84e22acfdb979a903948d7c53564f891c79dd927117742f: Status 404 returned error can't find the container with id a3ef2aa4aaff6c47c84e22acfdb979a903948d7c53564f891c79dd927117742f Feb 27 17:08:16 crc kubenswrapper[4700]: I0227 17:08:16.804910 4700 generic.go:334] "Generic (PLEG): container finished" podID="53351da8-d63b-4af8-912e-9a7aa5da939a" containerID="1ef1a017ab0d0f7357bd880ebe9ab4da0cfa644a0ba5175237f1eecbf9269905" exitCode=0 Feb 27 17:08:16 crc kubenswrapper[4700]: I0227 17:08:16.804980 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sl4bz" event={"ID":"53351da8-d63b-4af8-912e-9a7aa5da939a","Type":"ContainerDied","Data":"1ef1a017ab0d0f7357bd880ebe9ab4da0cfa644a0ba5175237f1eecbf9269905"} Feb 27 17:08:16 crc kubenswrapper[4700]: I0227 17:08:16.805010 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sl4bz" event={"ID":"53351da8-d63b-4af8-912e-9a7aa5da939a","Type":"ContainerStarted","Data":"5c7ac57f85eeb4c04ec375eac19b32744866d2bab95921a2cd767a8ab4bf2bcc"} Feb 27 17:08:16 crc kubenswrapper[4700]: I0227 17:08:16.807934 4700 generic.go:334] "Generic (PLEG): container finished" podID="f9dae327-14da-40ce-8e98-b29453b0352f" containerID="56e2da25c094bef9c00c9646193edd1773f6edeb0fca041d6e7e9b4cbadf8119" exitCode=0 Feb 27 17:08:16 crc kubenswrapper[4700]: I0227 17:08:16.808012 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8xmq" event={"ID":"f9dae327-14da-40ce-8e98-b29453b0352f","Type":"ContainerDied","Data":"56e2da25c094bef9c00c9646193edd1773f6edeb0fca041d6e7e9b4cbadf8119"} Feb 27 17:08:16 crc kubenswrapper[4700]: I0227 17:08:16.808061 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8xmq" event={"ID":"f9dae327-14da-40ce-8e98-b29453b0352f","Type":"ContainerStarted","Data":"a3ef2aa4aaff6c47c84e22acfdb979a903948d7c53564f891c79dd927117742f"} Feb 27 17:08:17 crc kubenswrapper[4700]: I0227 17:08:17.906675 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qjw68"] Feb 27 17:08:17 crc kubenswrapper[4700]: I0227 17:08:17.909347 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qjw68" Feb 27 17:08:17 crc kubenswrapper[4700]: I0227 17:08:17.911997 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 27 17:08:17 crc kubenswrapper[4700]: I0227 17:08:17.935616 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qjw68"] Feb 27 17:08:18 crc kubenswrapper[4700]: I0227 17:08:18.006761 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be7f5a17-2d75-446f-90ed-b5802f57b07a-catalog-content\") pod \"certified-operators-qjw68\" (UID: \"be7f5a17-2d75-446f-90ed-b5802f57b07a\") " pod="openshift-marketplace/certified-operators-qjw68" Feb 27 17:08:18 crc kubenswrapper[4700]: I0227 17:08:18.006845 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srwkf\" (UniqueName: \"kubernetes.io/projected/be7f5a17-2d75-446f-90ed-b5802f57b07a-kube-api-access-srwkf\") pod \"certified-operators-qjw68\" (UID: \"be7f5a17-2d75-446f-90ed-b5802f57b07a\") " pod="openshift-marketplace/certified-operators-qjw68" Feb 27 17:08:18 crc kubenswrapper[4700]: I0227 17:08:18.006990 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be7f5a17-2d75-446f-90ed-b5802f57b07a-utilities\") pod \"certified-operators-qjw68\" (UID: \"be7f5a17-2d75-446f-90ed-b5802f57b07a\") " pod="openshift-marketplace/certified-operators-qjw68" Feb 27 17:08:18 crc kubenswrapper[4700]: I0227 17:08:18.111236 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be7f5a17-2d75-446f-90ed-b5802f57b07a-utilities\") pod \"certified-operators-qjw68\" (UID: \"be7f5a17-2d75-446f-90ed-b5802f57b07a\") " pod="openshift-marketplace/certified-operators-qjw68" Feb 27 17:08:18 crc kubenswrapper[4700]: I0227 17:08:18.111377 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be7f5a17-2d75-446f-90ed-b5802f57b07a-catalog-content\") pod \"certified-operators-qjw68\" (UID: \"be7f5a17-2d75-446f-90ed-b5802f57b07a\") " pod="openshift-marketplace/certified-operators-qjw68" Feb 27 17:08:18 crc kubenswrapper[4700]: I0227 17:08:18.111446 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srwkf\" (UniqueName: \"kubernetes.io/projected/be7f5a17-2d75-446f-90ed-b5802f57b07a-kube-api-access-srwkf\") pod \"certified-operators-qjw68\" (UID: \"be7f5a17-2d75-446f-90ed-b5802f57b07a\") " pod="openshift-marketplace/certified-operators-qjw68" Feb 27 17:08:18 crc kubenswrapper[4700]: I0227 17:08:18.113572 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be7f5a17-2d75-446f-90ed-b5802f57b07a-utilities\") pod \"certified-operators-qjw68\" (UID: \"be7f5a17-2d75-446f-90ed-b5802f57b07a\") " pod="openshift-marketplace/certified-operators-qjw68" Feb 27 17:08:18 crc kubenswrapper[4700]: I0227 17:08:18.114074 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be7f5a17-2d75-446f-90ed-b5802f57b07a-catalog-content\") pod \"certified-operators-qjw68\" (UID: \"be7f5a17-2d75-446f-90ed-b5802f57b07a\") " pod="openshift-marketplace/certified-operators-qjw68" Feb 27 17:08:18 crc kubenswrapper[4700]: I0227 17:08:18.119278 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-98xlg"] Feb 27 17:08:18 crc kubenswrapper[4700]: I0227 17:08:18.123089 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-98xlg" Feb 27 17:08:18 crc kubenswrapper[4700]: I0227 17:08:18.128051 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 27 17:08:18 crc kubenswrapper[4700]: I0227 17:08:18.128738 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-98xlg"] Feb 27 17:08:18 crc kubenswrapper[4700]: I0227 17:08:18.154812 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srwkf\" (UniqueName: \"kubernetes.io/projected/be7f5a17-2d75-446f-90ed-b5802f57b07a-kube-api-access-srwkf\") pod \"certified-operators-qjw68\" (UID: \"be7f5a17-2d75-446f-90ed-b5802f57b07a\") " pod="openshift-marketplace/certified-operators-qjw68" Feb 27 17:08:18 crc kubenswrapper[4700]: I0227 17:08:18.211930 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/622df037-57ba-4a94-8e21-d78cfb08a79c-utilities\") pod \"community-operators-98xlg\" (UID: \"622df037-57ba-4a94-8e21-d78cfb08a79c\") " pod="openshift-marketplace/community-operators-98xlg" Feb 27 17:08:18 crc kubenswrapper[4700]: I0227 17:08:18.211985 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/622df037-57ba-4a94-8e21-d78cfb08a79c-catalog-content\") pod \"community-operators-98xlg\" (UID: \"622df037-57ba-4a94-8e21-d78cfb08a79c\") " pod="openshift-marketplace/community-operators-98xlg" Feb 27 17:08:18 crc kubenswrapper[4700]: I0227 17:08:18.212011 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5wwp\" (UniqueName: \"kubernetes.io/projected/622df037-57ba-4a94-8e21-d78cfb08a79c-kube-api-access-b5wwp\") pod \"community-operators-98xlg\" (UID: \"622df037-57ba-4a94-8e21-d78cfb08a79c\") " pod="openshift-marketplace/community-operators-98xlg" Feb 27 17:08:18 crc kubenswrapper[4700]: I0227 17:08:18.228709 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qjw68" Feb 27 17:08:18 crc kubenswrapper[4700]: I0227 17:08:18.312932 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/622df037-57ba-4a94-8e21-d78cfb08a79c-utilities\") pod \"community-operators-98xlg\" (UID: \"622df037-57ba-4a94-8e21-d78cfb08a79c\") " pod="openshift-marketplace/community-operators-98xlg" Feb 27 17:08:18 crc kubenswrapper[4700]: I0227 17:08:18.313031 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/622df037-57ba-4a94-8e21-d78cfb08a79c-catalog-content\") pod \"community-operators-98xlg\" (UID: \"622df037-57ba-4a94-8e21-d78cfb08a79c\") " pod="openshift-marketplace/community-operators-98xlg" Feb 27 17:08:18 crc kubenswrapper[4700]: I0227 17:08:18.313085 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5wwp\" (UniqueName: \"kubernetes.io/projected/622df037-57ba-4a94-8e21-d78cfb08a79c-kube-api-access-b5wwp\") pod \"community-operators-98xlg\" (UID: \"622df037-57ba-4a94-8e21-d78cfb08a79c\") " pod="openshift-marketplace/community-operators-98xlg" Feb 27 17:08:18 crc kubenswrapper[4700]: I0227 17:08:18.313872 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/622df037-57ba-4a94-8e21-d78cfb08a79c-utilities\") pod \"community-operators-98xlg\" (UID: \"622df037-57ba-4a94-8e21-d78cfb08a79c\") " pod="openshift-marketplace/community-operators-98xlg" Feb 27 17:08:18 crc kubenswrapper[4700]: I0227 17:08:18.314261 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/622df037-57ba-4a94-8e21-d78cfb08a79c-catalog-content\") pod \"community-operators-98xlg\" (UID: \"622df037-57ba-4a94-8e21-d78cfb08a79c\") " pod="openshift-marketplace/community-operators-98xlg" Feb 27 17:08:18 crc kubenswrapper[4700]: I0227 17:08:18.349126 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5wwp\" (UniqueName: \"kubernetes.io/projected/622df037-57ba-4a94-8e21-d78cfb08a79c-kube-api-access-b5wwp\") pod \"community-operators-98xlg\" (UID: \"622df037-57ba-4a94-8e21-d78cfb08a79c\") " pod="openshift-marketplace/community-operators-98xlg" Feb 27 17:08:18 crc kubenswrapper[4700]: I0227 17:08:18.482282 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qjw68"] Feb 27 17:08:18 crc kubenswrapper[4700]: W0227 17:08:18.490148 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbe7f5a17_2d75_446f_90ed_b5802f57b07a.slice/crio-3609c6636ec2088b66d518e9299d36dce19fd54fdd030360c99d633c7bcbe5a8 WatchSource:0}: Error finding container 3609c6636ec2088b66d518e9299d36dce19fd54fdd030360c99d633c7bcbe5a8: Status 404 returned error can't find the container with id 3609c6636ec2088b66d518e9299d36dce19fd54fdd030360c99d633c7bcbe5a8 Feb 27 17:08:18 crc kubenswrapper[4700]: I0227 17:08:18.496171 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-98xlg" Feb 27 17:08:18 crc kubenswrapper[4700]: I0227 17:08:18.697941 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-98xlg"] Feb 27 17:08:18 crc kubenswrapper[4700]: W0227 17:08:18.703583 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod622df037_57ba_4a94_8e21_d78cfb08a79c.slice/crio-d72d56c06e86b4bdb3d42094b29891bb988c7b08c9b6fff43cf4646f8184572f WatchSource:0}: Error finding container d72d56c06e86b4bdb3d42094b29891bb988c7b08c9b6fff43cf4646f8184572f: Status 404 returned error can't find the container with id d72d56c06e86b4bdb3d42094b29891bb988c7b08c9b6fff43cf4646f8184572f Feb 27 17:08:18 crc kubenswrapper[4700]: I0227 17:08:18.822504 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-98xlg" event={"ID":"622df037-57ba-4a94-8e21-d78cfb08a79c","Type":"ContainerStarted","Data":"d72d56c06e86b4bdb3d42094b29891bb988c7b08c9b6fff43cf4646f8184572f"} Feb 27 17:08:18 crc kubenswrapper[4700]: I0227 17:08:18.846048 4700 generic.go:334] "Generic (PLEG): container finished" podID="53351da8-d63b-4af8-912e-9a7aa5da939a" containerID="21b937530a807497ae25a62c9ed9db14b12b86e1d2014f4cf3bdbea1865d7076" exitCode=0 Feb 27 17:08:18 crc kubenswrapper[4700]: I0227 17:08:18.846115 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sl4bz" event={"ID":"53351da8-d63b-4af8-912e-9a7aa5da939a","Type":"ContainerDied","Data":"21b937530a807497ae25a62c9ed9db14b12b86e1d2014f4cf3bdbea1865d7076"} Feb 27 17:08:18 crc kubenswrapper[4700]: I0227 17:08:18.854036 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qjw68" event={"ID":"be7f5a17-2d75-446f-90ed-b5802f57b07a","Type":"ContainerStarted","Data":"3fa4c7ab2a569c2126247a326cf410d76863b937c971a582179761dc4d86b276"} Feb 27 17:08:18 crc kubenswrapper[4700]: I0227 17:08:18.854109 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qjw68" event={"ID":"be7f5a17-2d75-446f-90ed-b5802f57b07a","Type":"ContainerStarted","Data":"3609c6636ec2088b66d518e9299d36dce19fd54fdd030360c99d633c7bcbe5a8"} Feb 27 17:08:19 crc kubenswrapper[4700]: I0227 17:08:19.860985 4700 generic.go:334] "Generic (PLEG): container finished" podID="622df037-57ba-4a94-8e21-d78cfb08a79c" containerID="5b3a2ce5a9db54662678f7ddea8d4f91a9a37b70b7a051088df57367fbe501d5" exitCode=0 Feb 27 17:08:19 crc kubenswrapper[4700]: I0227 17:08:19.861070 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-98xlg" event={"ID":"622df037-57ba-4a94-8e21-d78cfb08a79c","Type":"ContainerDied","Data":"5b3a2ce5a9db54662678f7ddea8d4f91a9a37b70b7a051088df57367fbe501d5"} Feb 27 17:08:19 crc kubenswrapper[4700]: I0227 17:08:19.864352 4700 generic.go:334] "Generic (PLEG): container finished" podID="be7f5a17-2d75-446f-90ed-b5802f57b07a" containerID="3fa4c7ab2a569c2126247a326cf410d76863b937c971a582179761dc4d86b276" exitCode=0 Feb 27 17:08:19 crc kubenswrapper[4700]: I0227 17:08:19.864489 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qjw68" event={"ID":"be7f5a17-2d75-446f-90ed-b5802f57b07a","Type":"ContainerDied","Data":"3fa4c7ab2a569c2126247a326cf410d76863b937c971a582179761dc4d86b276"} Feb 27 17:08:19 crc kubenswrapper[4700]: I0227 17:08:19.866705 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8xmq" event={"ID":"f9dae327-14da-40ce-8e98-b29453b0352f","Type":"ContainerStarted","Data":"a40cc8895c05f4acb855e561117052a299983b86354534480ee93c90e7ae06dc"} Feb 27 17:08:20 crc kubenswrapper[4700]: I0227 17:08:20.878644 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sl4bz" event={"ID":"53351da8-d63b-4af8-912e-9a7aa5da939a","Type":"ContainerStarted","Data":"93064ee9cf7bc8fe1208e2c2f939f1c013de98c2d93878c1aeb3f28b0855d52b"} Feb 27 17:08:20 crc kubenswrapper[4700]: I0227 17:08:20.881622 4700 generic.go:334] "Generic (PLEG): container finished" podID="f9dae327-14da-40ce-8e98-b29453b0352f" containerID="a40cc8895c05f4acb855e561117052a299983b86354534480ee93c90e7ae06dc" exitCode=0 Feb 27 17:08:20 crc kubenswrapper[4700]: I0227 17:08:20.881676 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8xmq" event={"ID":"f9dae327-14da-40ce-8e98-b29453b0352f","Type":"ContainerDied","Data":"a40cc8895c05f4acb855e561117052a299983b86354534480ee93c90e7ae06dc"} Feb 27 17:08:21 crc kubenswrapper[4700]: I0227 17:08:21.912810 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-sl4bz" podStartSLOduration=3.505559772 podStartE2EDuration="6.912791792s" podCreationTimestamp="2026-02-27 17:08:15 +0000 UTC" firstStartedPulling="2026-02-27 17:08:16.806342401 +0000 UTC m=+456.791655168" lastFinishedPulling="2026-02-27 17:08:20.213574441 +0000 UTC m=+460.198887188" observedRunningTime="2026-02-27 17:08:21.912512625 +0000 UTC m=+461.897825412" watchObservedRunningTime="2026-02-27 17:08:21.912791792 +0000 UTC m=+461.898104549" Feb 27 17:08:24 crc kubenswrapper[4700]: I0227 17:08:24.904870 4700 generic.go:334] "Generic (PLEG): container finished" podID="622df037-57ba-4a94-8e21-d78cfb08a79c" containerID="a98875af84e5473f2430114dfc8c9bd1941218a456e5865f73595a2a54e42295" exitCode=0 Feb 27 17:08:24 crc kubenswrapper[4700]: I0227 17:08:24.904973 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-98xlg" event={"ID":"622df037-57ba-4a94-8e21-d78cfb08a79c","Type":"ContainerDied","Data":"a98875af84e5473f2430114dfc8c9bd1941218a456e5865f73595a2a54e42295"} Feb 27 17:08:24 crc kubenswrapper[4700]: I0227 17:08:24.907447 4700 generic.go:334] "Generic (PLEG): container finished" podID="be7f5a17-2d75-446f-90ed-b5802f57b07a" containerID="79f033dd71797894f9ab36ca15d1f70844f19243ce596b305dc0ac7ad61fbcc9" exitCode=0 Feb 27 17:08:24 crc kubenswrapper[4700]: I0227 17:08:24.907523 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qjw68" event={"ID":"be7f5a17-2d75-446f-90ed-b5802f57b07a","Type":"ContainerDied","Data":"79f033dd71797894f9ab36ca15d1f70844f19243ce596b305dc0ac7ad61fbcc9"} Feb 27 17:08:24 crc kubenswrapper[4700]: I0227 17:08:24.910567 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8xmq" event={"ID":"f9dae327-14da-40ce-8e98-b29453b0352f","Type":"ContainerStarted","Data":"f359d56b3c0bb99cf3310d1e4831d3307ebfee69ff08c27e14f1feb7f5010ad1"} Feb 27 17:08:24 crc kubenswrapper[4700]: I0227 17:08:24.964355 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-q8xmq" podStartSLOduration=3.175889499 podStartE2EDuration="9.964337141s" podCreationTimestamp="2026-02-27 17:08:15 +0000 UTC" firstStartedPulling="2026-02-27 17:08:16.809735834 +0000 UTC m=+456.795048581" lastFinishedPulling="2026-02-27 17:08:23.598183436 +0000 UTC m=+463.583496223" observedRunningTime="2026-02-27 17:08:24.962935033 +0000 UTC m=+464.948247780" watchObservedRunningTime="2026-02-27 17:08:24.964337141 +0000 UTC m=+464.949649888" Feb 27 17:08:25 crc kubenswrapper[4700]: I0227 17:08:25.791885 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-hj6xf"] Feb 27 17:08:25 crc kubenswrapper[4700]: I0227 17:08:25.792644 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-hj6xf" Feb 27 17:08:25 crc kubenswrapper[4700]: I0227 17:08:25.805233 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-hj6xf"] Feb 27 17:08:25 crc kubenswrapper[4700]: I0227 17:08:25.844120 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-sl4bz" Feb 27 17:08:25 crc kubenswrapper[4700]: I0227 17:08:25.844377 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-sl4bz" Feb 27 17:08:25 crc kubenswrapper[4700]: I0227 17:08:25.896490 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-sl4bz" Feb 27 17:08:25 crc kubenswrapper[4700]: I0227 17:08:25.915349 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-hj6xf\" (UID: \"4a895fb8-c5a1-4523-9089-c37a9ae92d5f\") " pod="openshift-image-registry/image-registry-66df7c8f76-hj6xf" Feb 27 17:08:25 crc kubenswrapper[4700]: I0227 17:08:25.915408 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4a895fb8-c5a1-4523-9089-c37a9ae92d5f-trusted-ca\") pod \"image-registry-66df7c8f76-hj6xf\" (UID: \"4a895fb8-c5a1-4523-9089-c37a9ae92d5f\") " pod="openshift-image-registry/image-registry-66df7c8f76-hj6xf" Feb 27 17:08:25 crc kubenswrapper[4700]: I0227 17:08:25.915441 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jb8qt\" (UniqueName: \"kubernetes.io/projected/4a895fb8-c5a1-4523-9089-c37a9ae92d5f-kube-api-access-jb8qt\") pod \"image-registry-66df7c8f76-hj6xf\" (UID: \"4a895fb8-c5a1-4523-9089-c37a9ae92d5f\") " pod="openshift-image-registry/image-registry-66df7c8f76-hj6xf" Feb 27 17:08:25 crc kubenswrapper[4700]: I0227 17:08:25.915477 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4a895fb8-c5a1-4523-9089-c37a9ae92d5f-registry-tls\") pod \"image-registry-66df7c8f76-hj6xf\" (UID: \"4a895fb8-c5a1-4523-9089-c37a9ae92d5f\") " pod="openshift-image-registry/image-registry-66df7c8f76-hj6xf" Feb 27 17:08:25 crc kubenswrapper[4700]: I0227 17:08:25.915499 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4a895fb8-c5a1-4523-9089-c37a9ae92d5f-installation-pull-secrets\") pod \"image-registry-66df7c8f76-hj6xf\" (UID: \"4a895fb8-c5a1-4523-9089-c37a9ae92d5f\") " pod="openshift-image-registry/image-registry-66df7c8f76-hj6xf" Feb 27 17:08:25 crc kubenswrapper[4700]: I0227 17:08:25.915529 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4a895fb8-c5a1-4523-9089-c37a9ae92d5f-registry-certificates\") pod \"image-registry-66df7c8f76-hj6xf\" (UID: \"4a895fb8-c5a1-4523-9089-c37a9ae92d5f\") " pod="openshift-image-registry/image-registry-66df7c8f76-hj6xf" Feb 27 17:08:25 crc kubenswrapper[4700]: I0227 17:08:25.915577 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4a895fb8-c5a1-4523-9089-c37a9ae92d5f-ca-trust-extracted\") pod \"image-registry-66df7c8f76-hj6xf\" (UID: \"4a895fb8-c5a1-4523-9089-c37a9ae92d5f\") " pod="openshift-image-registry/image-registry-66df7c8f76-hj6xf" Feb 27 17:08:25 crc kubenswrapper[4700]: I0227 17:08:25.915594 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4a895fb8-c5a1-4523-9089-c37a9ae92d5f-bound-sa-token\") pod \"image-registry-66df7c8f76-hj6xf\" (UID: \"4a895fb8-c5a1-4523-9089-c37a9ae92d5f\") " pod="openshift-image-registry/image-registry-66df7c8f76-hj6xf" Feb 27 17:08:25 crc kubenswrapper[4700]: I0227 17:08:25.920189 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qjw68" event={"ID":"be7f5a17-2d75-446f-90ed-b5802f57b07a","Type":"ContainerStarted","Data":"ec7b94744c9a0936ac2d59e837f01b7cca71f21f1aa6341c6d941c7bed387f31"} Feb 27 17:08:25 crc kubenswrapper[4700]: I0227 17:08:25.925411 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-98xlg" event={"ID":"622df037-57ba-4a94-8e21-d78cfb08a79c","Type":"ContainerStarted","Data":"3aaf72070a5e1c14a4ff4e200e9098003f007b0b9d239a61cafe45b6b147f327"} Feb 27 17:08:25 crc kubenswrapper[4700]: I0227 17:08:25.954391 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qjw68" podStartSLOduration=3.744475406 podStartE2EDuration="8.954296888s" podCreationTimestamp="2026-02-27 17:08:17 +0000 UTC" firstStartedPulling="2026-02-27 17:08:20.211260128 +0000 UTC m=+460.196572875" lastFinishedPulling="2026-02-27 17:08:25.42108157 +0000 UTC m=+465.406394357" observedRunningTime="2026-02-27 17:08:25.950051282 +0000 UTC m=+465.935364029" watchObservedRunningTime="2026-02-27 17:08:25.954296888 +0000 UTC m=+465.939609635" Feb 27 17:08:25 crc kubenswrapper[4700]: I0227 17:08:25.957059 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-hj6xf\" (UID: \"4a895fb8-c5a1-4523-9089-c37a9ae92d5f\") " pod="openshift-image-registry/image-registry-66df7c8f76-hj6xf" Feb 27 17:08:25 crc kubenswrapper[4700]: I0227 17:08:25.979854 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-sl4bz" Feb 27 17:08:25 crc kubenswrapper[4700]: I0227 17:08:25.985701 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-98xlg" podStartSLOduration=2.707560043 podStartE2EDuration="7.985672801s" podCreationTimestamp="2026-02-27 17:08:18 +0000 UTC" firstStartedPulling="2026-02-27 17:08:20.211497205 +0000 UTC m=+460.196809952" lastFinishedPulling="2026-02-27 17:08:25.489609963 +0000 UTC m=+465.474922710" observedRunningTime="2026-02-27 17:08:25.980102199 +0000 UTC m=+465.965414946" watchObservedRunningTime="2026-02-27 17:08:25.985672801 +0000 UTC m=+465.970985548" Feb 27 17:08:26 crc kubenswrapper[4700]: I0227 17:08:26.017214 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4a895fb8-c5a1-4523-9089-c37a9ae92d5f-registry-certificates\") pod \"image-registry-66df7c8f76-hj6xf\" (UID: \"4a895fb8-c5a1-4523-9089-c37a9ae92d5f\") " pod="openshift-image-registry/image-registry-66df7c8f76-hj6xf" Feb 27 17:08:26 crc kubenswrapper[4700]: I0227 17:08:26.017272 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4a895fb8-c5a1-4523-9089-c37a9ae92d5f-ca-trust-extracted\") pod \"image-registry-66df7c8f76-hj6xf\" (UID: \"4a895fb8-c5a1-4523-9089-c37a9ae92d5f\") " pod="openshift-image-registry/image-registry-66df7c8f76-hj6xf" Feb 27 17:08:26 crc kubenswrapper[4700]: I0227 17:08:26.017290 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4a895fb8-c5a1-4523-9089-c37a9ae92d5f-bound-sa-token\") pod \"image-registry-66df7c8f76-hj6xf\" (UID: \"4a895fb8-c5a1-4523-9089-c37a9ae92d5f\") " pod="openshift-image-registry/image-registry-66df7c8f76-hj6xf" Feb 27 17:08:26 crc kubenswrapper[4700]: I0227 17:08:26.017344 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4a895fb8-c5a1-4523-9089-c37a9ae92d5f-trusted-ca\") pod \"image-registry-66df7c8f76-hj6xf\" (UID: \"4a895fb8-c5a1-4523-9089-c37a9ae92d5f\") " pod="openshift-image-registry/image-registry-66df7c8f76-hj6xf" Feb 27 17:08:26 crc kubenswrapper[4700]: I0227 17:08:26.017366 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jb8qt\" (UniqueName: \"kubernetes.io/projected/4a895fb8-c5a1-4523-9089-c37a9ae92d5f-kube-api-access-jb8qt\") pod \"image-registry-66df7c8f76-hj6xf\" (UID: \"4a895fb8-c5a1-4523-9089-c37a9ae92d5f\") " pod="openshift-image-registry/image-registry-66df7c8f76-hj6xf" Feb 27 17:08:26 crc kubenswrapper[4700]: I0227 17:08:26.017383 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4a895fb8-c5a1-4523-9089-c37a9ae92d5f-registry-tls\") pod \"image-registry-66df7c8f76-hj6xf\" (UID: \"4a895fb8-c5a1-4523-9089-c37a9ae92d5f\") " pod="openshift-image-registry/image-registry-66df7c8f76-hj6xf" Feb 27 17:08:26 crc kubenswrapper[4700]: I0227 17:08:26.017402 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4a895fb8-c5a1-4523-9089-c37a9ae92d5f-installation-pull-secrets\") pod \"image-registry-66df7c8f76-hj6xf\" (UID: \"4a895fb8-c5a1-4523-9089-c37a9ae92d5f\") " pod="openshift-image-registry/image-registry-66df7c8f76-hj6xf" Feb 27 17:08:26 crc kubenswrapper[4700]: I0227 17:08:26.019101 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4a895fb8-c5a1-4523-9089-c37a9ae92d5f-ca-trust-extracted\") pod \"image-registry-66df7c8f76-hj6xf\" (UID: \"4a895fb8-c5a1-4523-9089-c37a9ae92d5f\") " pod="openshift-image-registry/image-registry-66df7c8f76-hj6xf" Feb 27 17:08:26 crc kubenswrapper[4700]: I0227 17:08:26.019765 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4a895fb8-c5a1-4523-9089-c37a9ae92d5f-trusted-ca\") pod \"image-registry-66df7c8f76-hj6xf\" (UID: \"4a895fb8-c5a1-4523-9089-c37a9ae92d5f\") " pod="openshift-image-registry/image-registry-66df7c8f76-hj6xf" Feb 27 17:08:26 crc kubenswrapper[4700]: I0227 17:08:26.020341 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4a895fb8-c5a1-4523-9089-c37a9ae92d5f-registry-certificates\") pod \"image-registry-66df7c8f76-hj6xf\" (UID: \"4a895fb8-c5a1-4523-9089-c37a9ae92d5f\") " pod="openshift-image-registry/image-registry-66df7c8f76-hj6xf" Feb 27 17:08:26 crc kubenswrapper[4700]: I0227 17:08:26.025055 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4a895fb8-c5a1-4523-9089-c37a9ae92d5f-installation-pull-secrets\") pod \"image-registry-66df7c8f76-hj6xf\" (UID: \"4a895fb8-c5a1-4523-9089-c37a9ae92d5f\") " pod="openshift-image-registry/image-registry-66df7c8f76-hj6xf" Feb 27 17:08:26 crc kubenswrapper[4700]: I0227 17:08:26.025062 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4a895fb8-c5a1-4523-9089-c37a9ae92d5f-registry-tls\") pod \"image-registry-66df7c8f76-hj6xf\" (UID: \"4a895fb8-c5a1-4523-9089-c37a9ae92d5f\") " pod="openshift-image-registry/image-registry-66df7c8f76-hj6xf" Feb 27 17:08:26 crc kubenswrapper[4700]: I0227 17:08:26.034838 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4a895fb8-c5a1-4523-9089-c37a9ae92d5f-bound-sa-token\") pod \"image-registry-66df7c8f76-hj6xf\" (UID: \"4a895fb8-c5a1-4523-9089-c37a9ae92d5f\") " pod="openshift-image-registry/image-registry-66df7c8f76-hj6xf" Feb 27 17:08:26 crc kubenswrapper[4700]: I0227 17:08:26.035003 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jb8qt\" (UniqueName: \"kubernetes.io/projected/4a895fb8-c5a1-4523-9089-c37a9ae92d5f-kube-api-access-jb8qt\") pod \"image-registry-66df7c8f76-hj6xf\" (UID: \"4a895fb8-c5a1-4523-9089-c37a9ae92d5f\") " pod="openshift-image-registry/image-registry-66df7c8f76-hj6xf" Feb 27 17:08:26 crc kubenswrapper[4700]: I0227 17:08:26.112354 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-q8xmq" Feb 27 17:08:26 crc kubenswrapper[4700]: I0227 17:08:26.112411 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-q8xmq" Feb 27 17:08:26 crc kubenswrapper[4700]: I0227 17:08:26.163753 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-hj6xf" Feb 27 17:08:26 crc kubenswrapper[4700]: I0227 17:08:26.363505 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-hj6xf"] Feb 27 17:08:26 crc kubenswrapper[4700]: I0227 17:08:26.930248 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-hj6xf" event={"ID":"4a895fb8-c5a1-4523-9089-c37a9ae92d5f","Type":"ContainerStarted","Data":"6f1d11b09e1049189ab87e75966ebdf582d1a9e8958c5ceff29ccd993ffbf192"} Feb 27 17:08:26 crc kubenswrapper[4700]: I0227 17:08:26.930295 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-hj6xf" event={"ID":"4a895fb8-c5a1-4523-9089-c37a9ae92d5f","Type":"ContainerStarted","Data":"571284584d85753da9a9b6af05889d994ed302c5c00ae773303900112bac9ec7"} Feb 27 17:08:27 crc kubenswrapper[4700]: I0227 17:08:27.152754 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-q8xmq" podUID="f9dae327-14da-40ce-8e98-b29453b0352f" containerName="registry-server" probeResult="failure" output=< Feb 27 17:08:27 crc kubenswrapper[4700]: timeout: failed to connect service ":50051" within 1s Feb 27 17:08:27 crc kubenswrapper[4700]: > Feb 27 17:08:27 crc kubenswrapper[4700]: I0227 17:08:27.943703 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-hj6xf" Feb 27 17:08:28 crc kubenswrapper[4700]: I0227 17:08:28.229846 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qjw68" Feb 27 17:08:28 crc kubenswrapper[4700]: I0227 17:08:28.229933 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qjw68" Feb 27 17:08:28 crc kubenswrapper[4700]: I0227 17:08:28.275249 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qjw68" Feb 27 17:08:28 crc kubenswrapper[4700]: I0227 17:08:28.289855 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-hj6xf" podStartSLOduration=3.2898356189999998 podStartE2EDuration="3.289835619s" podCreationTimestamp="2026-02-27 17:08:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:08:26.952072056 +0000 UTC m=+466.937384843" watchObservedRunningTime="2026-02-27 17:08:28.289835619 +0000 UTC m=+468.275148376" Feb 27 17:08:28 crc kubenswrapper[4700]: I0227 17:08:28.497156 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-98xlg" Feb 27 17:08:28 crc kubenswrapper[4700]: I0227 17:08:28.497229 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-98xlg" Feb 27 17:08:28 crc kubenswrapper[4700]: I0227 17:08:28.567333 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-98xlg" Feb 27 17:08:36 crc kubenswrapper[4700]: I0227 17:08:36.176297 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-q8xmq" Feb 27 17:08:36 crc kubenswrapper[4700]: I0227 17:08:36.256223 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-q8xmq" Feb 27 17:08:36 crc kubenswrapper[4700]: I0227 17:08:36.410798 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:08:36 crc kubenswrapper[4700]: I0227 17:08:36.410862 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:08:36 crc kubenswrapper[4700]: I0227 17:08:36.410904 4700 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" Feb 27 17:08:36 crc kubenswrapper[4700]: I0227 17:08:36.411447 4700 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8776130d3c26488575d3ae9c1c22338916b362ebc2dc54951e165061dd752d23"} pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 17:08:36 crc kubenswrapper[4700]: I0227 17:08:36.411517 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" containerID="cri-o://8776130d3c26488575d3ae9c1c22338916b362ebc2dc54951e165061dd752d23" gracePeriod=600 Feb 27 17:08:38 crc kubenswrapper[4700]: I0227 17:08:38.010199 4700 generic.go:334] "Generic (PLEG): container finished" podID="84b45511-b94c-479f-98db-fd2c4eceec46" containerID="8776130d3c26488575d3ae9c1c22338916b362ebc2dc54951e165061dd752d23" exitCode=0 Feb 27 17:08:38 crc kubenswrapper[4700]: I0227 17:08:38.010275 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerDied","Data":"8776130d3c26488575d3ae9c1c22338916b362ebc2dc54951e165061dd752d23"} Feb 27 17:08:38 crc kubenswrapper[4700]: I0227 17:08:38.010907 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerStarted","Data":"e1a1a6356e2245cab05c4d1addb6b5e53e3e02d22a816739c9a355a221dd2d7b"} Feb 27 17:08:38 crc kubenswrapper[4700]: I0227 17:08:38.010934 4700 scope.go:117] "RemoveContainer" containerID="2ef7c37b61de7a039756248f222bd1769b774703c406de5ad11abd9fb2f46dd2" Feb 27 17:08:38 crc kubenswrapper[4700]: I0227 17:08:38.292873 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qjw68" Feb 27 17:08:38 crc kubenswrapper[4700]: I0227 17:08:38.564753 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-98xlg" Feb 27 17:08:46 crc kubenswrapper[4700]: I0227 17:08:46.171497 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-hj6xf" Feb 27 17:08:46 crc kubenswrapper[4700]: I0227 17:08:46.224353 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-k97gv"] Feb 27 17:09:11 crc kubenswrapper[4700]: I0227 17:09:11.265261 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" podUID="ee57d6e4-3550-46d0-947e-fe6db5cf3291" containerName="registry" containerID="cri-o://53a859b46b5d3051299ef73410d20f19b56dfdfc84bc08c14fd21bbda1c8a1e5" gracePeriod=30 Feb 27 17:09:11 crc kubenswrapper[4700]: I0227 17:09:11.700291 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:09:11 crc kubenswrapper[4700]: I0227 17:09:11.782628 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " Feb 27 17:09:11 crc kubenswrapper[4700]: I0227 17:09:11.782707 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ee57d6e4-3550-46d0-947e-fe6db5cf3291-trusted-ca\") pod \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " Feb 27 17:09:11 crc kubenswrapper[4700]: I0227 17:09:11.782740 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ee57d6e4-3550-46d0-947e-fe6db5cf3291-installation-pull-secrets\") pod \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " Feb 27 17:09:11 crc kubenswrapper[4700]: I0227 17:09:11.782762 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ee57d6e4-3550-46d0-947e-fe6db5cf3291-bound-sa-token\") pod \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " Feb 27 17:09:11 crc kubenswrapper[4700]: I0227 17:09:11.782790 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ee57d6e4-3550-46d0-947e-fe6db5cf3291-registry-certificates\") pod \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " Feb 27 17:09:11 crc kubenswrapper[4700]: I0227 17:09:11.782850 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ee57d6e4-3550-46d0-947e-fe6db5cf3291-registry-tls\") pod \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " Feb 27 17:09:11 crc kubenswrapper[4700]: I0227 17:09:11.782873 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6pzj\" (UniqueName: \"kubernetes.io/projected/ee57d6e4-3550-46d0-947e-fe6db5cf3291-kube-api-access-z6pzj\") pod \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " Feb 27 17:09:11 crc kubenswrapper[4700]: I0227 17:09:11.782923 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ee57d6e4-3550-46d0-947e-fe6db5cf3291-ca-trust-extracted\") pod \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\" (UID: \"ee57d6e4-3550-46d0-947e-fe6db5cf3291\") " Feb 27 17:09:11 crc kubenswrapper[4700]: I0227 17:09:11.785831 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee57d6e4-3550-46d0-947e-fe6db5cf3291-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "ee57d6e4-3550-46d0-947e-fe6db5cf3291" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:09:11 crc kubenswrapper[4700]: I0227 17:09:11.789815 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee57d6e4-3550-46d0-947e-fe6db5cf3291-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "ee57d6e4-3550-46d0-947e-fe6db5cf3291" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:09:11 crc kubenswrapper[4700]: I0227 17:09:11.790202 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee57d6e4-3550-46d0-947e-fe6db5cf3291-kube-api-access-z6pzj" (OuterVolumeSpecName: "kube-api-access-z6pzj") pod "ee57d6e4-3550-46d0-947e-fe6db5cf3291" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291"). InnerVolumeSpecName "kube-api-access-z6pzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:09:11 crc kubenswrapper[4700]: I0227 17:09:11.790613 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee57d6e4-3550-46d0-947e-fe6db5cf3291-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "ee57d6e4-3550-46d0-947e-fe6db5cf3291" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:09:11 crc kubenswrapper[4700]: I0227 17:09:11.793080 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee57d6e4-3550-46d0-947e-fe6db5cf3291-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "ee57d6e4-3550-46d0-947e-fe6db5cf3291" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:09:11 crc kubenswrapper[4700]: I0227 17:09:11.793345 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee57d6e4-3550-46d0-947e-fe6db5cf3291-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "ee57d6e4-3550-46d0-947e-fe6db5cf3291" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:09:11 crc kubenswrapper[4700]: I0227 17:09:11.795819 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "ee57d6e4-3550-46d0-947e-fe6db5cf3291" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 27 17:09:11 crc kubenswrapper[4700]: I0227 17:09:11.807291 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee57d6e4-3550-46d0-947e-fe6db5cf3291-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "ee57d6e4-3550-46d0-947e-fe6db5cf3291" (UID: "ee57d6e4-3550-46d0-947e-fe6db5cf3291"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:09:11 crc kubenswrapper[4700]: I0227 17:09:11.884214 4700 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ee57d6e4-3550-46d0-947e-fe6db5cf3291-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 27 17:09:11 crc kubenswrapper[4700]: I0227 17:09:11.884279 4700 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ee57d6e4-3550-46d0-947e-fe6db5cf3291-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 27 17:09:11 crc kubenswrapper[4700]: I0227 17:09:11.884301 4700 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ee57d6e4-3550-46d0-947e-fe6db5cf3291-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 27 17:09:11 crc kubenswrapper[4700]: I0227 17:09:11.884321 4700 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ee57d6e4-3550-46d0-947e-fe6db5cf3291-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 27 17:09:11 crc kubenswrapper[4700]: I0227 17:09:11.884340 4700 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ee57d6e4-3550-46d0-947e-fe6db5cf3291-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 27 17:09:11 crc kubenswrapper[4700]: I0227 17:09:11.884359 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6pzj\" (UniqueName: \"kubernetes.io/projected/ee57d6e4-3550-46d0-947e-fe6db5cf3291-kube-api-access-z6pzj\") on node \"crc\" DevicePath \"\"" Feb 27 17:09:11 crc kubenswrapper[4700]: I0227 17:09:11.884377 4700 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ee57d6e4-3550-46d0-947e-fe6db5cf3291-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 27 17:09:12 crc kubenswrapper[4700]: I0227 17:09:12.235155 4700 generic.go:334] "Generic (PLEG): container finished" podID="ee57d6e4-3550-46d0-947e-fe6db5cf3291" containerID="53a859b46b5d3051299ef73410d20f19b56dfdfc84bc08c14fd21bbda1c8a1e5" exitCode=0 Feb 27 17:09:12 crc kubenswrapper[4700]: I0227 17:09:12.235279 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" event={"ID":"ee57d6e4-3550-46d0-947e-fe6db5cf3291","Type":"ContainerDied","Data":"53a859b46b5d3051299ef73410d20f19b56dfdfc84bc08c14fd21bbda1c8a1e5"} Feb 27 17:09:12 crc kubenswrapper[4700]: I0227 17:09:12.235309 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" Feb 27 17:09:12 crc kubenswrapper[4700]: I0227 17:09:12.235811 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-k97gv" event={"ID":"ee57d6e4-3550-46d0-947e-fe6db5cf3291","Type":"ContainerDied","Data":"90e87d19965ec592df14b6d13c12968724f1a43bbfa6d69fbb0fcbbdf57aa07f"} Feb 27 17:09:12 crc kubenswrapper[4700]: I0227 17:09:12.235953 4700 scope.go:117] "RemoveContainer" containerID="53a859b46b5d3051299ef73410d20f19b56dfdfc84bc08c14fd21bbda1c8a1e5" Feb 27 17:09:12 crc kubenswrapper[4700]: I0227 17:09:12.266202 4700 scope.go:117] "RemoveContainer" containerID="53a859b46b5d3051299ef73410d20f19b56dfdfc84bc08c14fd21bbda1c8a1e5" Feb 27 17:09:12 crc kubenswrapper[4700]: E0227 17:09:12.268487 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53a859b46b5d3051299ef73410d20f19b56dfdfc84bc08c14fd21bbda1c8a1e5\": container with ID starting with 53a859b46b5d3051299ef73410d20f19b56dfdfc84bc08c14fd21bbda1c8a1e5 not found: ID does not exist" containerID="53a859b46b5d3051299ef73410d20f19b56dfdfc84bc08c14fd21bbda1c8a1e5" Feb 27 17:09:12 crc kubenswrapper[4700]: I0227 17:09:12.268553 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53a859b46b5d3051299ef73410d20f19b56dfdfc84bc08c14fd21bbda1c8a1e5"} err="failed to get container status \"53a859b46b5d3051299ef73410d20f19b56dfdfc84bc08c14fd21bbda1c8a1e5\": rpc error: code = NotFound desc = could not find container \"53a859b46b5d3051299ef73410d20f19b56dfdfc84bc08c14fd21bbda1c8a1e5\": container with ID starting with 53a859b46b5d3051299ef73410d20f19b56dfdfc84bc08c14fd21bbda1c8a1e5 not found: ID does not exist" Feb 27 17:09:12 crc kubenswrapper[4700]: I0227 17:09:12.298014 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-k97gv"] Feb 27 17:09:12 crc kubenswrapper[4700]: I0227 17:09:12.302628 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-k97gv"] Feb 27 17:09:12 crc kubenswrapper[4700]: I0227 17:09:12.992077 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee57d6e4-3550-46d0-947e-fe6db5cf3291" path="/var/lib/kubelet/pods/ee57d6e4-3550-46d0-947e-fe6db5cf3291/volumes" Feb 27 17:10:00 crc kubenswrapper[4700]: I0227 17:10:00.123343 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536870-ss2vg"] Feb 27 17:10:00 crc kubenswrapper[4700]: E0227 17:10:00.124067 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee57d6e4-3550-46d0-947e-fe6db5cf3291" containerName="registry" Feb 27 17:10:00 crc kubenswrapper[4700]: I0227 17:10:00.124081 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee57d6e4-3550-46d0-947e-fe6db5cf3291" containerName="registry" Feb 27 17:10:00 crc kubenswrapper[4700]: I0227 17:10:00.124190 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee57d6e4-3550-46d0-947e-fe6db5cf3291" containerName="registry" Feb 27 17:10:00 crc kubenswrapper[4700]: I0227 17:10:00.124555 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536870-ss2vg" Feb 27 17:10:00 crc kubenswrapper[4700]: I0227 17:10:00.126402 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:10:00 crc kubenswrapper[4700]: I0227 17:10:00.127130 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d62kc" Feb 27 17:10:00 crc kubenswrapper[4700]: I0227 17:10:00.129657 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:10:00 crc kubenswrapper[4700]: I0227 17:10:00.135227 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536870-ss2vg"] Feb 27 17:10:00 crc kubenswrapper[4700]: I0227 17:10:00.280381 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvh24\" (UniqueName: \"kubernetes.io/projected/1c8bc912-83a1-4ed4-8df7-ec5060b47700-kube-api-access-hvh24\") pod \"auto-csr-approver-29536870-ss2vg\" (UID: \"1c8bc912-83a1-4ed4-8df7-ec5060b47700\") " pod="openshift-infra/auto-csr-approver-29536870-ss2vg" Feb 27 17:10:00 crc kubenswrapper[4700]: I0227 17:10:00.382045 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvh24\" (UniqueName: \"kubernetes.io/projected/1c8bc912-83a1-4ed4-8df7-ec5060b47700-kube-api-access-hvh24\") pod \"auto-csr-approver-29536870-ss2vg\" (UID: \"1c8bc912-83a1-4ed4-8df7-ec5060b47700\") " pod="openshift-infra/auto-csr-approver-29536870-ss2vg" Feb 27 17:10:00 crc kubenswrapper[4700]: I0227 17:10:00.407220 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvh24\" (UniqueName: \"kubernetes.io/projected/1c8bc912-83a1-4ed4-8df7-ec5060b47700-kube-api-access-hvh24\") pod \"auto-csr-approver-29536870-ss2vg\" (UID: \"1c8bc912-83a1-4ed4-8df7-ec5060b47700\") " pod="openshift-infra/auto-csr-approver-29536870-ss2vg" Feb 27 17:10:00 crc kubenswrapper[4700]: I0227 17:10:00.445947 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536870-ss2vg" Feb 27 17:10:00 crc kubenswrapper[4700]: I0227 17:10:00.696623 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536870-ss2vg"] Feb 27 17:10:00 crc kubenswrapper[4700]: I0227 17:10:00.711915 4700 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 17:10:01 crc kubenswrapper[4700]: I0227 17:10:01.586248 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536870-ss2vg" event={"ID":"1c8bc912-83a1-4ed4-8df7-ec5060b47700","Type":"ContainerStarted","Data":"c9056172902742409f39d608253f1b6fb3636b21d2ba232d31d0c29b1bc24efe"} Feb 27 17:10:02 crc kubenswrapper[4700]: I0227 17:10:02.593894 4700 generic.go:334] "Generic (PLEG): container finished" podID="1c8bc912-83a1-4ed4-8df7-ec5060b47700" containerID="a3b8232566af1f16791722b65b5aacf80836bcde9541da4ef5ed04b439a21043" exitCode=0 Feb 27 17:10:02 crc kubenswrapper[4700]: I0227 17:10:02.593931 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536870-ss2vg" event={"ID":"1c8bc912-83a1-4ed4-8df7-ec5060b47700","Type":"ContainerDied","Data":"a3b8232566af1f16791722b65b5aacf80836bcde9541da4ef5ed04b439a21043"} Feb 27 17:10:03 crc kubenswrapper[4700]: I0227 17:10:03.911147 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536870-ss2vg" Feb 27 17:10:04 crc kubenswrapper[4700]: I0227 17:10:04.031035 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvh24\" (UniqueName: \"kubernetes.io/projected/1c8bc912-83a1-4ed4-8df7-ec5060b47700-kube-api-access-hvh24\") pod \"1c8bc912-83a1-4ed4-8df7-ec5060b47700\" (UID: \"1c8bc912-83a1-4ed4-8df7-ec5060b47700\") " Feb 27 17:10:04 crc kubenswrapper[4700]: I0227 17:10:04.039237 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c8bc912-83a1-4ed4-8df7-ec5060b47700-kube-api-access-hvh24" (OuterVolumeSpecName: "kube-api-access-hvh24") pod "1c8bc912-83a1-4ed4-8df7-ec5060b47700" (UID: "1c8bc912-83a1-4ed4-8df7-ec5060b47700"). InnerVolumeSpecName "kube-api-access-hvh24". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:10:04 crc kubenswrapper[4700]: I0227 17:10:04.132340 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvh24\" (UniqueName: \"kubernetes.io/projected/1c8bc912-83a1-4ed4-8df7-ec5060b47700-kube-api-access-hvh24\") on node \"crc\" DevicePath \"\"" Feb 27 17:10:04 crc kubenswrapper[4700]: I0227 17:10:04.623532 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536870-ss2vg" event={"ID":"1c8bc912-83a1-4ed4-8df7-ec5060b47700","Type":"ContainerDied","Data":"c9056172902742409f39d608253f1b6fb3636b21d2ba232d31d0c29b1bc24efe"} Feb 27 17:10:04 crc kubenswrapper[4700]: I0227 17:10:04.623600 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9056172902742409f39d608253f1b6fb3636b21d2ba232d31d0c29b1bc24efe" Feb 27 17:10:04 crc kubenswrapper[4700]: I0227 17:10:04.623662 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536870-ss2vg" Feb 27 17:10:04 crc kubenswrapper[4700]: I0227 17:10:04.997177 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536864-kmfkv"] Feb 27 17:10:05 crc kubenswrapper[4700]: I0227 17:10:05.005836 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536864-kmfkv"] Feb 27 17:10:06 crc kubenswrapper[4700]: I0227 17:10:06.992925 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3166d7d3-7842-4655-b10b-24e5731d77e0" path="/var/lib/kubelet/pods/3166d7d3-7842-4655-b10b-24e5731d77e0/volumes" Feb 27 17:11:06 crc kubenswrapper[4700]: I0227 17:11:06.410834 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:11:06 crc kubenswrapper[4700]: I0227 17:11:06.412096 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:11:36 crc kubenswrapper[4700]: I0227 17:11:36.410775 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:11:36 crc kubenswrapper[4700]: I0227 17:11:36.411409 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:11:42 crc kubenswrapper[4700]: I0227 17:11:42.562448 4700 scope.go:117] "RemoveContainer" containerID="150cee9a20fea156a81b992257b8aa0474a7374f0bf2b8aefeb83d1ea77030d4" Feb 27 17:12:00 crc kubenswrapper[4700]: I0227 17:12:00.161199 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536872-tkkz5"] Feb 27 17:12:00 crc kubenswrapper[4700]: E0227 17:12:00.162135 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c8bc912-83a1-4ed4-8df7-ec5060b47700" containerName="oc" Feb 27 17:12:00 crc kubenswrapper[4700]: I0227 17:12:00.162155 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c8bc912-83a1-4ed4-8df7-ec5060b47700" containerName="oc" Feb 27 17:12:00 crc kubenswrapper[4700]: I0227 17:12:00.162341 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c8bc912-83a1-4ed4-8df7-ec5060b47700" containerName="oc" Feb 27 17:12:00 crc kubenswrapper[4700]: I0227 17:12:00.162978 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536872-tkkz5" Feb 27 17:12:00 crc kubenswrapper[4700]: I0227 17:12:00.168379 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:12:00 crc kubenswrapper[4700]: I0227 17:12:00.168427 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:12:00 crc kubenswrapper[4700]: I0227 17:12:00.168515 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d62kc" Feb 27 17:12:00 crc kubenswrapper[4700]: I0227 17:12:00.174760 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536872-tkkz5"] Feb 27 17:12:00 crc kubenswrapper[4700]: I0227 17:12:00.206605 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29dnn\" (UniqueName: \"kubernetes.io/projected/6e5bcb83-503a-4684-9e4c-3d3c9badbae5-kube-api-access-29dnn\") pod \"auto-csr-approver-29536872-tkkz5\" (UID: \"6e5bcb83-503a-4684-9e4c-3d3c9badbae5\") " pod="openshift-infra/auto-csr-approver-29536872-tkkz5" Feb 27 17:12:00 crc kubenswrapper[4700]: I0227 17:12:00.307823 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29dnn\" (UniqueName: \"kubernetes.io/projected/6e5bcb83-503a-4684-9e4c-3d3c9badbae5-kube-api-access-29dnn\") pod \"auto-csr-approver-29536872-tkkz5\" (UID: \"6e5bcb83-503a-4684-9e4c-3d3c9badbae5\") " pod="openshift-infra/auto-csr-approver-29536872-tkkz5" Feb 27 17:12:00 crc kubenswrapper[4700]: I0227 17:12:00.346295 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29dnn\" (UniqueName: \"kubernetes.io/projected/6e5bcb83-503a-4684-9e4c-3d3c9badbae5-kube-api-access-29dnn\") pod \"auto-csr-approver-29536872-tkkz5\" (UID: \"6e5bcb83-503a-4684-9e4c-3d3c9badbae5\") " pod="openshift-infra/auto-csr-approver-29536872-tkkz5" Feb 27 17:12:00 crc kubenswrapper[4700]: I0227 17:12:00.490959 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536872-tkkz5" Feb 27 17:12:00 crc kubenswrapper[4700]: I0227 17:12:00.776332 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536872-tkkz5"] Feb 27 17:12:01 crc kubenswrapper[4700]: I0227 17:12:01.520453 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536872-tkkz5" event={"ID":"6e5bcb83-503a-4684-9e4c-3d3c9badbae5","Type":"ContainerStarted","Data":"b57ca5590e925bbe6d81b498cc7cb7c7fd4e54252e3aad006ff15af4a8ef0d20"} Feb 27 17:12:02 crc kubenswrapper[4700]: I0227 17:12:02.529919 4700 generic.go:334] "Generic (PLEG): container finished" podID="6e5bcb83-503a-4684-9e4c-3d3c9badbae5" containerID="f21ac7cba5e13686e60de613f90c6209fb6b1b815f8b55deec107290160ab448" exitCode=0 Feb 27 17:12:02 crc kubenswrapper[4700]: I0227 17:12:02.529984 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536872-tkkz5" event={"ID":"6e5bcb83-503a-4684-9e4c-3d3c9badbae5","Type":"ContainerDied","Data":"f21ac7cba5e13686e60de613f90c6209fb6b1b815f8b55deec107290160ab448"} Feb 27 17:12:03 crc kubenswrapper[4700]: I0227 17:12:03.916029 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536872-tkkz5" Feb 27 17:12:04 crc kubenswrapper[4700]: I0227 17:12:04.056689 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29dnn\" (UniqueName: \"kubernetes.io/projected/6e5bcb83-503a-4684-9e4c-3d3c9badbae5-kube-api-access-29dnn\") pod \"6e5bcb83-503a-4684-9e4c-3d3c9badbae5\" (UID: \"6e5bcb83-503a-4684-9e4c-3d3c9badbae5\") " Feb 27 17:12:04 crc kubenswrapper[4700]: I0227 17:12:04.067081 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e5bcb83-503a-4684-9e4c-3d3c9badbae5-kube-api-access-29dnn" (OuterVolumeSpecName: "kube-api-access-29dnn") pod "6e5bcb83-503a-4684-9e4c-3d3c9badbae5" (UID: "6e5bcb83-503a-4684-9e4c-3d3c9badbae5"). InnerVolumeSpecName "kube-api-access-29dnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:12:04 crc kubenswrapper[4700]: I0227 17:12:04.158919 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29dnn\" (UniqueName: \"kubernetes.io/projected/6e5bcb83-503a-4684-9e4c-3d3c9badbae5-kube-api-access-29dnn\") on node \"crc\" DevicePath \"\"" Feb 27 17:12:04 crc kubenswrapper[4700]: I0227 17:12:04.546382 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536872-tkkz5" event={"ID":"6e5bcb83-503a-4684-9e4c-3d3c9badbae5","Type":"ContainerDied","Data":"b57ca5590e925bbe6d81b498cc7cb7c7fd4e54252e3aad006ff15af4a8ef0d20"} Feb 27 17:12:04 crc kubenswrapper[4700]: I0227 17:12:04.546886 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b57ca5590e925bbe6d81b498cc7cb7c7fd4e54252e3aad006ff15af4a8ef0d20" Feb 27 17:12:04 crc kubenswrapper[4700]: I0227 17:12:04.546768 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536872-tkkz5" Feb 27 17:12:04 crc kubenswrapper[4700]: I0227 17:12:04.998565 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536866-th9ft"] Feb 27 17:12:05 crc kubenswrapper[4700]: I0227 17:12:05.005294 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536866-th9ft"] Feb 27 17:12:06 crc kubenswrapper[4700]: I0227 17:12:06.410682 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:12:06 crc kubenswrapper[4700]: I0227 17:12:06.410751 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:12:06 crc kubenswrapper[4700]: I0227 17:12:06.410802 4700 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" Feb 27 17:12:06 crc kubenswrapper[4700]: I0227 17:12:06.411448 4700 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e1a1a6356e2245cab05c4d1addb6b5e53e3e02d22a816739c9a355a221dd2d7b"} pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 17:12:06 crc kubenswrapper[4700]: I0227 17:12:06.411528 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" containerID="cri-o://e1a1a6356e2245cab05c4d1addb6b5e53e3e02d22a816739c9a355a221dd2d7b" gracePeriod=600 Feb 27 17:12:06 crc kubenswrapper[4700]: I0227 17:12:06.566279 4700 generic.go:334] "Generic (PLEG): container finished" podID="84b45511-b94c-479f-98db-fd2c4eceec46" containerID="e1a1a6356e2245cab05c4d1addb6b5e53e3e02d22a816739c9a355a221dd2d7b" exitCode=0 Feb 27 17:12:06 crc kubenswrapper[4700]: I0227 17:12:06.566355 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerDied","Data":"e1a1a6356e2245cab05c4d1addb6b5e53e3e02d22a816739c9a355a221dd2d7b"} Feb 27 17:12:06 crc kubenswrapper[4700]: I0227 17:12:06.566434 4700 scope.go:117] "RemoveContainer" containerID="8776130d3c26488575d3ae9c1c22338916b362ebc2dc54951e165061dd752d23" Feb 27 17:12:06 crc kubenswrapper[4700]: I0227 17:12:06.995058 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f0dd4fc-ce75-4088-be4d-664e4b5a31ea" path="/var/lib/kubelet/pods/4f0dd4fc-ce75-4088-be4d-664e4b5a31ea/volumes" Feb 27 17:12:07 crc kubenswrapper[4700]: I0227 17:12:07.576946 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerStarted","Data":"510d6e176ab30bcaf35acfa6f85490773219083ba09de332e6a13153619b2fac"} Feb 27 17:13:21 crc kubenswrapper[4700]: I0227 17:13:21.213230 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-fdwz4"] Feb 27 17:13:21 crc kubenswrapper[4700]: E0227 17:13:21.214011 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e5bcb83-503a-4684-9e4c-3d3c9badbae5" containerName="oc" Feb 27 17:13:21 crc kubenswrapper[4700]: I0227 17:13:21.214027 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e5bcb83-503a-4684-9e4c-3d3c9badbae5" containerName="oc" Feb 27 17:13:21 crc kubenswrapper[4700]: I0227 17:13:21.214155 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e5bcb83-503a-4684-9e4c-3d3c9badbae5" containerName="oc" Feb 27 17:13:21 crc kubenswrapper[4700]: I0227 17:13:21.214576 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-fdwz4" Feb 27 17:13:21 crc kubenswrapper[4700]: I0227 17:13:21.216782 4700 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-jpcrk" Feb 27 17:13:21 crc kubenswrapper[4700]: I0227 17:13:21.218055 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Feb 27 17:13:21 crc kubenswrapper[4700]: I0227 17:13:21.220819 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-fdwz4"] Feb 27 17:13:21 crc kubenswrapper[4700]: I0227 17:13:21.222305 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Feb 27 17:13:21 crc kubenswrapper[4700]: I0227 17:13:21.241086 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-szl9g"] Feb 27 17:13:21 crc kubenswrapper[4700]: I0227 17:13:21.242128 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-szl9g" Feb 27 17:13:21 crc kubenswrapper[4700]: I0227 17:13:21.243561 4700 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-tphf9" Feb 27 17:13:21 crc kubenswrapper[4700]: I0227 17:13:21.249129 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-6xcnk"] Feb 27 17:13:21 crc kubenswrapper[4700]: I0227 17:13:21.249947 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-6xcnk" Feb 27 17:13:21 crc kubenswrapper[4700]: I0227 17:13:21.251645 4700 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-nqkhp" Feb 27 17:13:21 crc kubenswrapper[4700]: I0227 17:13:21.255336 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-szl9g"] Feb 27 17:13:21 crc kubenswrapper[4700]: I0227 17:13:21.290537 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-6xcnk"] Feb 27 17:13:21 crc kubenswrapper[4700]: I0227 17:13:21.311835 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bq442\" (UniqueName: \"kubernetes.io/projected/e1d9b74a-c2c9-4654-b440-b4eb4f2deef1-kube-api-access-bq442\") pod \"cert-manager-webhook-687f57d79b-szl9g\" (UID: \"e1d9b74a-c2c9-4654-b440-b4eb4f2deef1\") " pod="cert-manager/cert-manager-webhook-687f57d79b-szl9g" Feb 27 17:13:21 crc kubenswrapper[4700]: I0227 17:13:21.311911 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q96pd\" (UniqueName: \"kubernetes.io/projected/c00612dd-19ac-4a4b-870f-d1e70fa1604f-kube-api-access-q96pd\") pod \"cert-manager-cainjector-cf98fcc89-fdwz4\" (UID: \"c00612dd-19ac-4a4b-870f-d1e70fa1604f\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-fdwz4" Feb 27 17:13:21 crc kubenswrapper[4700]: I0227 17:13:21.311936 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7lc8\" (UniqueName: \"kubernetes.io/projected/fc05be47-71de-4f97-b1f2-b58bacfc752e-kube-api-access-n7lc8\") pod \"cert-manager-858654f9db-6xcnk\" (UID: \"fc05be47-71de-4f97-b1f2-b58bacfc752e\") " pod="cert-manager/cert-manager-858654f9db-6xcnk" Feb 27 17:13:21 crc kubenswrapper[4700]: I0227 17:13:21.415041 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bq442\" (UniqueName: \"kubernetes.io/projected/e1d9b74a-c2c9-4654-b440-b4eb4f2deef1-kube-api-access-bq442\") pod \"cert-manager-webhook-687f57d79b-szl9g\" (UID: \"e1d9b74a-c2c9-4654-b440-b4eb4f2deef1\") " pod="cert-manager/cert-manager-webhook-687f57d79b-szl9g" Feb 27 17:13:21 crc kubenswrapper[4700]: I0227 17:13:21.415119 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q96pd\" (UniqueName: \"kubernetes.io/projected/c00612dd-19ac-4a4b-870f-d1e70fa1604f-kube-api-access-q96pd\") pod \"cert-manager-cainjector-cf98fcc89-fdwz4\" (UID: \"c00612dd-19ac-4a4b-870f-d1e70fa1604f\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-fdwz4" Feb 27 17:13:21 crc kubenswrapper[4700]: I0227 17:13:21.415148 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7lc8\" (UniqueName: \"kubernetes.io/projected/fc05be47-71de-4f97-b1f2-b58bacfc752e-kube-api-access-n7lc8\") pod \"cert-manager-858654f9db-6xcnk\" (UID: \"fc05be47-71de-4f97-b1f2-b58bacfc752e\") " pod="cert-manager/cert-manager-858654f9db-6xcnk" Feb 27 17:13:21 crc kubenswrapper[4700]: I0227 17:13:21.437881 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7lc8\" (UniqueName: \"kubernetes.io/projected/fc05be47-71de-4f97-b1f2-b58bacfc752e-kube-api-access-n7lc8\") pod \"cert-manager-858654f9db-6xcnk\" (UID: \"fc05be47-71de-4f97-b1f2-b58bacfc752e\") " pod="cert-manager/cert-manager-858654f9db-6xcnk" Feb 27 17:13:21 crc kubenswrapper[4700]: I0227 17:13:21.438391 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bq442\" (UniqueName: \"kubernetes.io/projected/e1d9b74a-c2c9-4654-b440-b4eb4f2deef1-kube-api-access-bq442\") pod \"cert-manager-webhook-687f57d79b-szl9g\" (UID: \"e1d9b74a-c2c9-4654-b440-b4eb4f2deef1\") " pod="cert-manager/cert-manager-webhook-687f57d79b-szl9g" Feb 27 17:13:21 crc kubenswrapper[4700]: I0227 17:13:21.439419 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q96pd\" (UniqueName: \"kubernetes.io/projected/c00612dd-19ac-4a4b-870f-d1e70fa1604f-kube-api-access-q96pd\") pod \"cert-manager-cainjector-cf98fcc89-fdwz4\" (UID: \"c00612dd-19ac-4a4b-870f-d1e70fa1604f\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-fdwz4" Feb 27 17:13:21 crc kubenswrapper[4700]: I0227 17:13:21.530294 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-fdwz4" Feb 27 17:13:21 crc kubenswrapper[4700]: I0227 17:13:21.560823 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-szl9g" Feb 27 17:13:21 crc kubenswrapper[4700]: I0227 17:13:21.578233 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-6xcnk" Feb 27 17:13:22 crc kubenswrapper[4700]: I0227 17:13:22.031438 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-fdwz4"] Feb 27 17:13:22 crc kubenswrapper[4700]: W0227 17:13:22.036783 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc05be47_71de_4f97_b1f2_b58bacfc752e.slice/crio-b3af7e13a701637c9827944608950e6755b2f9f27e6a1b084957cc5292f14cc9 WatchSource:0}: Error finding container b3af7e13a701637c9827944608950e6755b2f9f27e6a1b084957cc5292f14cc9: Status 404 returned error can't find the container with id b3af7e13a701637c9827944608950e6755b2f9f27e6a1b084957cc5292f14cc9 Feb 27 17:13:22 crc kubenswrapper[4700]: I0227 17:13:22.040519 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-6xcnk"] Feb 27 17:13:22 crc kubenswrapper[4700]: I0227 17:13:22.045974 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-szl9g"] Feb 27 17:13:22 crc kubenswrapper[4700]: W0227 17:13:22.047253 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1d9b74a_c2c9_4654_b440_b4eb4f2deef1.slice/crio-a0370d1d34a9e7ace44c0da8fce08e0db93952b8aeabcae08e78c568f6b354f2 WatchSource:0}: Error finding container a0370d1d34a9e7ace44c0da8fce08e0db93952b8aeabcae08e78c568f6b354f2: Status 404 returned error can't find the container with id a0370d1d34a9e7ace44c0da8fce08e0db93952b8aeabcae08e78c568f6b354f2 Feb 27 17:13:22 crc kubenswrapper[4700]: I0227 17:13:22.084399 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-szl9g" event={"ID":"e1d9b74a-c2c9-4654-b440-b4eb4f2deef1","Type":"ContainerStarted","Data":"a0370d1d34a9e7ace44c0da8fce08e0db93952b8aeabcae08e78c568f6b354f2"} Feb 27 17:13:22 crc kubenswrapper[4700]: I0227 17:13:22.086364 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-fdwz4" event={"ID":"c00612dd-19ac-4a4b-870f-d1e70fa1604f","Type":"ContainerStarted","Data":"1e663ba781a929463f38f11616961ffb0ae300cfa4f8edd1d345df5ed10f0f0d"} Feb 27 17:13:22 crc kubenswrapper[4700]: I0227 17:13:22.089062 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-6xcnk" event={"ID":"fc05be47-71de-4f97-b1f2-b58bacfc752e","Type":"ContainerStarted","Data":"b3af7e13a701637c9827944608950e6755b2f9f27e6a1b084957cc5292f14cc9"} Feb 27 17:13:25 crc kubenswrapper[4700]: I0227 17:13:25.108655 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-6xcnk" event={"ID":"fc05be47-71de-4f97-b1f2-b58bacfc752e","Type":"ContainerStarted","Data":"e76cf046b2c08af2871ed1ee705a0b0e2d804d7911938d39e183beff7f4fb67a"} Feb 27 17:13:25 crc kubenswrapper[4700]: I0227 17:13:25.126711 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-6xcnk" podStartSLOduration=1.478869376 podStartE2EDuration="4.126683095s" podCreationTimestamp="2026-02-27 17:13:21 +0000 UTC" firstStartedPulling="2026-02-27 17:13:22.03996398 +0000 UTC m=+762.025276757" lastFinishedPulling="2026-02-27 17:13:24.687777719 +0000 UTC m=+764.673090476" observedRunningTime="2026-02-27 17:13:25.126382227 +0000 UTC m=+765.111694974" watchObservedRunningTime="2026-02-27 17:13:25.126683095 +0000 UTC m=+765.111995882" Feb 27 17:13:26 crc kubenswrapper[4700]: I0227 17:13:26.117651 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-fdwz4" event={"ID":"c00612dd-19ac-4a4b-870f-d1e70fa1604f","Type":"ContainerStarted","Data":"9936f89b57608d45bca27ea7d51fbea0d7b87ffedd39a6a573e361312c260637"} Feb 27 17:13:26 crc kubenswrapper[4700]: I0227 17:13:26.148281 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-fdwz4" podStartSLOduration=1.540657063 podStartE2EDuration="5.148062493s" podCreationTimestamp="2026-02-27 17:13:21 +0000 UTC" firstStartedPulling="2026-02-27 17:13:22.035106531 +0000 UTC m=+762.020419288" lastFinishedPulling="2026-02-27 17:13:25.642511941 +0000 UTC m=+765.627824718" observedRunningTime="2026-02-27 17:13:26.140371738 +0000 UTC m=+766.125684485" watchObservedRunningTime="2026-02-27 17:13:26.148062493 +0000 UTC m=+766.133375250" Feb 27 17:13:27 crc kubenswrapper[4700]: I0227 17:13:27.127115 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-szl9g" event={"ID":"e1d9b74a-c2c9-4654-b440-b4eb4f2deef1","Type":"ContainerStarted","Data":"6af9ac841f381cdcea87143b2c00881193bf4c5e3c24bcb2e1d8dae30233c5aa"} Feb 27 17:13:27 crc kubenswrapper[4700]: I0227 17:13:27.157900 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-szl9g" podStartSLOduration=1.3587411249999999 podStartE2EDuration="6.157869632s" podCreationTimestamp="2026-02-27 17:13:21 +0000 UTC" firstStartedPulling="2026-02-27 17:13:22.049070063 +0000 UTC m=+762.034382810" lastFinishedPulling="2026-02-27 17:13:26.84819857 +0000 UTC m=+766.833511317" observedRunningTime="2026-02-27 17:13:27.148864002 +0000 UTC m=+767.134176789" watchObservedRunningTime="2026-02-27 17:13:27.157869632 +0000 UTC m=+767.143182389" Feb 27 17:13:28 crc kubenswrapper[4700]: I0227 17:13:28.134025 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-szl9g" Feb 27 17:13:31 crc kubenswrapper[4700]: I0227 17:13:31.565014 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-szl9g" Feb 27 17:13:38 crc kubenswrapper[4700]: I0227 17:13:38.597055 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-jtbqn"] Feb 27 17:13:38 crc kubenswrapper[4700]: I0227 17:13:38.598093 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="ovn-controller" containerID="cri-o://b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0" gracePeriod=30 Feb 27 17:13:38 crc kubenswrapper[4700]: I0227 17:13:38.598203 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="kube-rbac-proxy-node" containerID="cri-o://a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a" gracePeriod=30 Feb 27 17:13:38 crc kubenswrapper[4700]: I0227 17:13:38.598202 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64" gracePeriod=30 Feb 27 17:13:38 crc kubenswrapper[4700]: I0227 17:13:38.598269 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="ovn-acl-logging" containerID="cri-o://ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8" gracePeriod=30 Feb 27 17:13:38 crc kubenswrapper[4700]: I0227 17:13:38.598327 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="northd" containerID="cri-o://c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf" gracePeriod=30 Feb 27 17:13:38 crc kubenswrapper[4700]: I0227 17:13:38.598516 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="sbdb" containerID="cri-o://3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191" gracePeriod=30 Feb 27 17:13:38 crc kubenswrapper[4700]: I0227 17:13:38.598553 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="nbdb" containerID="cri-o://611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265" gracePeriod=30 Feb 27 17:13:38 crc kubenswrapper[4700]: I0227 17:13:38.641086 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="ovnkube-controller" containerID="cri-o://e2e83f8a1f5d3909754673dab13333df53c0524f2c5c6b33831c04129dc65302" gracePeriod=30 Feb 27 17:13:38 crc kubenswrapper[4700]: I0227 17:13:38.948787 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jtbqn_3a34c0e3-2513-4e81-b6b1-80d1230475fd/ovnkube-controller/3.log" Feb 27 17:13:38 crc kubenswrapper[4700]: I0227 17:13:38.954981 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jtbqn_3a34c0e3-2513-4e81-b6b1-80d1230475fd/ovn-acl-logging/0.log" Feb 27 17:13:38 crc kubenswrapper[4700]: I0227 17:13:38.956059 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jtbqn_3a34c0e3-2513-4e81-b6b1-80d1230475fd/ovn-controller/0.log" Feb 27 17:13:38 crc kubenswrapper[4700]: I0227 17:13:38.957038 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.033735 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-nn687"] Feb 27 17:13:39 crc kubenswrapper[4700]: E0227 17:13:39.034038 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="kube-rbac-proxy-ovn-metrics" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.034059 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="kube-rbac-proxy-ovn-metrics" Feb 27 17:13:39 crc kubenswrapper[4700]: E0227 17:13:39.034080 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="nbdb" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.034095 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="nbdb" Feb 27 17:13:39 crc kubenswrapper[4700]: E0227 17:13:39.034108 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="ovnkube-controller" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.034122 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="ovnkube-controller" Feb 27 17:13:39 crc kubenswrapper[4700]: E0227 17:13:39.034136 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="ovn-acl-logging" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.034148 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="ovn-acl-logging" Feb 27 17:13:39 crc kubenswrapper[4700]: E0227 17:13:39.034166 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="sbdb" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.034178 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="sbdb" Feb 27 17:13:39 crc kubenswrapper[4700]: E0227 17:13:39.034192 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="ovn-controller" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.034205 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="ovn-controller" Feb 27 17:13:39 crc kubenswrapper[4700]: E0227 17:13:39.034221 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="northd" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.034234 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="northd" Feb 27 17:13:39 crc kubenswrapper[4700]: E0227 17:13:39.034256 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="kube-rbac-proxy-node" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.034268 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="kube-rbac-proxy-node" Feb 27 17:13:39 crc kubenswrapper[4700]: E0227 17:13:39.034288 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="kubecfg-setup" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.034302 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="kubecfg-setup" Feb 27 17:13:39 crc kubenswrapper[4700]: E0227 17:13:39.034316 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="ovnkube-controller" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.034328 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="ovnkube-controller" Feb 27 17:13:39 crc kubenswrapper[4700]: E0227 17:13:39.034345 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="ovnkube-controller" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.034357 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="ovnkube-controller" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.034559 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="sbdb" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.034577 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="northd" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.034594 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="ovnkube-controller" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.034606 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="ovn-acl-logging" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.034625 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="ovn-controller" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.034644 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="ovnkube-controller" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.034658 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="kube-rbac-proxy-ovn-metrics" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.034673 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="nbdb" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.034690 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="ovnkube-controller" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.034707 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="kube-rbac-proxy-node" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.034726 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="ovnkube-controller" Feb 27 17:13:39 crc kubenswrapper[4700]: E0227 17:13:39.034877 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="ovnkube-controller" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.034941 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="ovnkube-controller" Feb 27 17:13:39 crc kubenswrapper[4700]: E0227 17:13:39.034967 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="ovnkube-controller" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.034981 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="ovnkube-controller" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.035177 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerName="ovnkube-controller" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.038452 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.155741 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3a34c0e3-2513-4e81-b6b1-80d1230475fd-ovn-node-metrics-cert\") pod \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.155799 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-systemd-units\") pod \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.155838 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.155875 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-slash\") pod \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.155904 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-run-systemd\") pod \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.155938 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-cni-bin\") pod \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.155991 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3a34c0e3-2513-4e81-b6b1-80d1230475fd-env-overrides\") pod \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.155997 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "3a34c0e3-2513-4e81-b6b1-80d1230475fd" (UID: "3a34c0e3-2513-4e81-b6b1-80d1230475fd"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.156027 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-cni-netd\") pod \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.156054 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-kubelet\") pod \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.156088 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-log-socket\") pod \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.156006 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "3a34c0e3-2513-4e81-b6b1-80d1230475fd" (UID: "3a34c0e3-2513-4e81-b6b1-80d1230475fd"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.156163 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-slash" (OuterVolumeSpecName: "host-slash") pod "3a34c0e3-2513-4e81-b6b1-80d1230475fd" (UID: "3a34c0e3-2513-4e81-b6b1-80d1230475fd"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.156194 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "3a34c0e3-2513-4e81-b6b1-80d1230475fd" (UID: "3a34c0e3-2513-4e81-b6b1-80d1230475fd"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.156189 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "3a34c0e3-2513-4e81-b6b1-80d1230475fd" (UID: "3a34c0e3-2513-4e81-b6b1-80d1230475fd"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.156201 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-log-socket" (OuterVolumeSpecName: "log-socket") pod "3a34c0e3-2513-4e81-b6b1-80d1230475fd" (UID: "3a34c0e3-2513-4e81-b6b1-80d1230475fd"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.156248 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-run-ovn\") pod \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.156247 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "3a34c0e3-2513-4e81-b6b1-80d1230475fd" (UID: "3a34c0e3-2513-4e81-b6b1-80d1230475fd"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.156333 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-run-ovn-kubernetes\") pod \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.156327 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "3a34c0e3-2513-4e81-b6b1-80d1230475fd" (UID: "3a34c0e3-2513-4e81-b6b1-80d1230475fd"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.156395 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "3a34c0e3-2513-4e81-b6b1-80d1230475fd" (UID: "3a34c0e3-2513-4e81-b6b1-80d1230475fd"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.156425 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-run-openvswitch\") pod \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.156485 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-var-lib-openvswitch\") pod \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.156519 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-node-log\") pod \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.156554 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "3a34c0e3-2513-4e81-b6b1-80d1230475fd" (UID: "3a34c0e3-2513-4e81-b6b1-80d1230475fd"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.156553 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "3a34c0e3-2513-4e81-b6b1-80d1230475fd" (UID: "3a34c0e3-2513-4e81-b6b1-80d1230475fd"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.156559 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3a34c0e3-2513-4e81-b6b1-80d1230475fd-ovnkube-config\") pod \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.156640 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-node-log" (OuterVolumeSpecName: "node-log") pod "3a34c0e3-2513-4e81-b6b1-80d1230475fd" (UID: "3a34c0e3-2513-4e81-b6b1-80d1230475fd"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.156718 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-748j7\" (UniqueName: \"kubernetes.io/projected/3a34c0e3-2513-4e81-b6b1-80d1230475fd-kube-api-access-748j7\") pod \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.156808 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-etc-openvswitch\") pod \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.156861 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3a34c0e3-2513-4e81-b6b1-80d1230475fd-ovnkube-script-lib\") pod \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.156909 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-run-netns\") pod \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\" (UID: \"3a34c0e3-2513-4e81-b6b1-80d1230475fd\") " Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.156914 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "3a34c0e3-2513-4e81-b6b1-80d1230475fd" (UID: "3a34c0e3-2513-4e81-b6b1-80d1230475fd"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.156961 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "3a34c0e3-2513-4e81-b6b1-80d1230475fd" (UID: "3a34c0e3-2513-4e81-b6b1-80d1230475fd"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.157103 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a34c0e3-2513-4e81-b6b1-80d1230475fd-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "3a34c0e3-2513-4e81-b6b1-80d1230475fd" (UID: "3a34c0e3-2513-4e81-b6b1-80d1230475fd"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.157184 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a34c0e3-2513-4e81-b6b1-80d1230475fd-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "3a34c0e3-2513-4e81-b6b1-80d1230475fd" (UID: "3a34c0e3-2513-4e81-b6b1-80d1230475fd"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.157186 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-systemd-units\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.157250 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-run-ovn\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.157316 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-host-run-ovn-kubernetes\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.157436 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-etc-openvswitch\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.157651 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-host-kubelet\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.157742 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-host-run-netns\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.157815 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-log-socket\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.157871 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-host-cni-bin\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.158008 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/20d19220-2826-4990-b39e-bd24304ead43-ovn-node-metrics-cert\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.158054 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-host-slash\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.158144 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/20d19220-2826-4990-b39e-bd24304ead43-env-overrides\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.158185 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-run-systemd\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.158216 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/20d19220-2826-4990-b39e-bd24304ead43-ovnkube-config\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.158315 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-run-openvswitch\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.158360 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.158398 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-host-cni-netd\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.158514 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-node-log\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.158603 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-var-lib-openvswitch\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.158671 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcz9w\" (UniqueName: \"kubernetes.io/projected/20d19220-2826-4990-b39e-bd24304ead43-kube-api-access-lcz9w\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.158731 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a34c0e3-2513-4e81-b6b1-80d1230475fd-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "3a34c0e3-2513-4e81-b6b1-80d1230475fd" (UID: "3a34c0e3-2513-4e81-b6b1-80d1230475fd"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.158857 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/20d19220-2826-4990-b39e-bd24304ead43-ovnkube-script-lib\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.159019 4700 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-systemd-units\") on node \"crc\" DevicePath \"\"" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.159042 4700 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.159086 4700 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-slash\") on node \"crc\" DevicePath \"\"" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.159105 4700 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-cni-bin\") on node \"crc\" DevicePath \"\"" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.159125 4700 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3a34c0e3-2513-4e81-b6b1-80d1230475fd-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.159141 4700 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-cni-netd\") on node \"crc\" DevicePath \"\"" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.159157 4700 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-kubelet\") on node \"crc\" DevicePath \"\"" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.159175 4700 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-log-socket\") on node \"crc\" DevicePath \"\"" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.159193 4700 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.159229 4700 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.159260 4700 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-run-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.159278 4700 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.159295 4700 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-node-log\") on node \"crc\" DevicePath \"\"" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.159311 4700 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3a34c0e3-2513-4e81-b6b1-80d1230475fd-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.159329 4700 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.159345 4700 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-host-run-netns\") on node \"crc\" DevicePath \"\"" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.159362 4700 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3a34c0e3-2513-4e81-b6b1-80d1230475fd-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.162510 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a34c0e3-2513-4e81-b6b1-80d1230475fd-kube-api-access-748j7" (OuterVolumeSpecName: "kube-api-access-748j7") pod "3a34c0e3-2513-4e81-b6b1-80d1230475fd" (UID: "3a34c0e3-2513-4e81-b6b1-80d1230475fd"). InnerVolumeSpecName "kube-api-access-748j7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.163971 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a34c0e3-2513-4e81-b6b1-80d1230475fd-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "3a34c0e3-2513-4e81-b6b1-80d1230475fd" (UID: "3a34c0e3-2513-4e81-b6b1-80d1230475fd"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.180155 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "3a34c0e3-2513-4e81-b6b1-80d1230475fd" (UID: "3a34c0e3-2513-4e81-b6b1-80d1230475fd"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.214091 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jtbqn_3a34c0e3-2513-4e81-b6b1-80d1230475fd/ovnkube-controller/3.log" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.218442 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jtbqn_3a34c0e3-2513-4e81-b6b1-80d1230475fd/ovn-acl-logging/0.log" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.219569 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jtbqn_3a34c0e3-2513-4e81-b6b1-80d1230475fd/ovn-controller/0.log" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.220266 4700 generic.go:334] "Generic (PLEG): container finished" podID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerID="e2e83f8a1f5d3909754673dab13333df53c0524f2c5c6b33831c04129dc65302" exitCode=0 Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.220318 4700 generic.go:334] "Generic (PLEG): container finished" podID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerID="3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191" exitCode=0 Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.220343 4700 generic.go:334] "Generic (PLEG): container finished" podID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerID="611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265" exitCode=0 Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.220360 4700 generic.go:334] "Generic (PLEG): container finished" podID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerID="c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf" exitCode=0 Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.220378 4700 generic.go:334] "Generic (PLEG): container finished" podID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerID="6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64" exitCode=0 Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.220395 4700 generic.go:334] "Generic (PLEG): container finished" podID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerID="a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a" exitCode=0 Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.220401 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.220412 4700 generic.go:334] "Generic (PLEG): container finished" podID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerID="ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8" exitCode=143 Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.220511 4700 generic.go:334] "Generic (PLEG): container finished" podID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" containerID="b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0" exitCode=143 Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.220388 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" event={"ID":"3a34c0e3-2513-4e81-b6b1-80d1230475fd","Type":"ContainerDied","Data":"e2e83f8a1f5d3909754673dab13333df53c0524f2c5c6b33831c04129dc65302"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.220647 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" event={"ID":"3a34c0e3-2513-4e81-b6b1-80d1230475fd","Type":"ContainerDied","Data":"3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.220715 4700 scope.go:117] "RemoveContainer" containerID="e2e83f8a1f5d3909754673dab13333df53c0524f2c5c6b33831c04129dc65302" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.220732 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" event={"ID":"3a34c0e3-2513-4e81-b6b1-80d1230475fd","Type":"ContainerDied","Data":"611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.220774 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" event={"ID":"3a34c0e3-2513-4e81-b6b1-80d1230475fd","Type":"ContainerDied","Data":"c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.220862 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" event={"ID":"3a34c0e3-2513-4e81-b6b1-80d1230475fd","Type":"ContainerDied","Data":"6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.220945 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" event={"ID":"3a34c0e3-2513-4e81-b6b1-80d1230475fd","Type":"ContainerDied","Data":"a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.220974 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.221039 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.221056 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.221070 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.221137 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.221570 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.221632 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.221651 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.221667 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.221693 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" event={"ID":"3a34c0e3-2513-4e81-b6b1-80d1230475fd","Type":"ContainerDied","Data":"ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.221723 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e2e83f8a1f5d3909754673dab13333df53c0524f2c5c6b33831c04129dc65302"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.221743 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.221757 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.221771 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.221785 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.221798 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.221812 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.221824 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.221838 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.221852 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.221872 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" event={"ID":"3a34c0e3-2513-4e81-b6b1-80d1230475fd","Type":"ContainerDied","Data":"b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.221897 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e2e83f8a1f5d3909754673dab13333df53c0524f2c5c6b33831c04129dc65302"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.221915 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.221929 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.221944 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.221959 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.221976 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.221989 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.222002 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.222018 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.222032 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.222052 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jtbqn" event={"ID":"3a34c0e3-2513-4e81-b6b1-80d1230475fd","Type":"ContainerDied","Data":"59b3e89a8e1c655b3a026b9f6a993dcc53a00351a6765f6762f67e5f72758623"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.222075 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e2e83f8a1f5d3909754673dab13333df53c0524f2c5c6b33831c04129dc65302"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.222090 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.222104 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.222117 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.222131 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.222145 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.222158 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.222171 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.222185 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.222198 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.224455 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-rbptz_66eb22fb-e593-40ff-9ff3-5bb32912972d/kube-multus/2.log" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.228437 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-rbptz_66eb22fb-e593-40ff-9ff3-5bb32912972d/kube-multus/1.log" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.228569 4700 generic.go:334] "Generic (PLEG): container finished" podID="66eb22fb-e593-40ff-9ff3-5bb32912972d" containerID="ebdea7f95d1c1fbe317f3b2abbfaf311938c90e4554c498e7dd55e758928933f" exitCode=2 Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.228707 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-rbptz" event={"ID":"66eb22fb-e593-40ff-9ff3-5bb32912972d","Type":"ContainerDied","Data":"ebdea7f95d1c1fbe317f3b2abbfaf311938c90e4554c498e7dd55e758928933f"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.228753 4700 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9c675574fe85ad3f55e0431b8d36fd1b83523ee502e675165aafad8551ba4560"} Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.230814 4700 scope.go:117] "RemoveContainer" containerID="ebdea7f95d1c1fbe317f3b2abbfaf311938c90e4554c498e7dd55e758928933f" Feb 27 17:13:39 crc kubenswrapper[4700]: E0227 17:13:39.231682 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-rbptz_openshift-multus(66eb22fb-e593-40ff-9ff3-5bb32912972d)\"" pod="openshift-multus/multus-rbptz" podUID="66eb22fb-e593-40ff-9ff3-5bb32912972d" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.260732 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-var-lib-openvswitch\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.261160 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcz9w\" (UniqueName: \"kubernetes.io/projected/20d19220-2826-4990-b39e-bd24304ead43-kube-api-access-lcz9w\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.261216 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/20d19220-2826-4990-b39e-bd24304ead43-ovnkube-script-lib\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.262101 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-systemd-units\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.262153 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-run-ovn\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.262184 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-host-run-ovn-kubernetes\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.262217 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-etc-openvswitch\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.262258 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-host-kubelet\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.262288 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-host-run-netns\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.262314 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-log-socket\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.262350 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-host-cni-bin\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.262384 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/20d19220-2826-4990-b39e-bd24304ead43-ovn-node-metrics-cert\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.262413 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-host-slash\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.262536 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/20d19220-2826-4990-b39e-bd24304ead43-env-overrides\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.264420 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-run-systemd\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.264538 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/20d19220-2826-4990-b39e-bd24304ead43-ovnkube-config\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.264600 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-run-openvswitch\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.264632 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.264720 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-host-cni-netd\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.264756 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-node-log\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.263324 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-host-run-ovn-kubernetes\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.263331 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-etc-openvswitch\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.263369 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-host-slash\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.263146 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-systemd-units\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.263190 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-host-cni-bin\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.263185 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-run-ovn\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.263221 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-host-run-netns\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.263206 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-host-kubelet\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.263274 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-log-socket\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.266194 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/20d19220-2826-4990-b39e-bd24304ead43-env-overrides\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.268107 4700 scope.go:117] "RemoveContainer" containerID="da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.268753 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/20d19220-2826-4990-b39e-bd24304ead43-ovnkube-script-lib\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.270152 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/20d19220-2826-4990-b39e-bd24304ead43-ovn-node-metrics-cert\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.270595 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-748j7\" (UniqueName: \"kubernetes.io/projected/3a34c0e3-2513-4e81-b6b1-80d1230475fd-kube-api-access-748j7\") on node \"crc\" DevicePath \"\"" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.270651 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-var-lib-openvswitch\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.270748 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-run-systemd\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.270828 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-host-cni-netd\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.270861 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-node-log\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.270822 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.270953 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/20d19220-2826-4990-b39e-bd24304ead43-run-openvswitch\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.272088 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/20d19220-2826-4990-b39e-bd24304ead43-ovnkube-config\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.273203 4700 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3a34c0e3-2513-4e81-b6b1-80d1230475fd-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.273283 4700 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3a34c0e3-2513-4e81-b6b1-80d1230475fd-run-systemd\") on node \"crc\" DevicePath \"\"" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.304522 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-jtbqn"] Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.305262 4700 scope.go:117] "RemoveContainer" containerID="3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.307355 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcz9w\" (UniqueName: \"kubernetes.io/projected/20d19220-2826-4990-b39e-bd24304ead43-kube-api-access-lcz9w\") pod \"ovnkube-node-nn687\" (UID: \"20d19220-2826-4990-b39e-bd24304ead43\") " pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.311884 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-jtbqn"] Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.328451 4700 scope.go:117] "RemoveContainer" containerID="611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.350887 4700 scope.go:117] "RemoveContainer" containerID="c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.364723 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.369910 4700 scope.go:117] "RemoveContainer" containerID="6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.394279 4700 scope.go:117] "RemoveContainer" containerID="a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.413455 4700 scope.go:117] "RemoveContainer" containerID="ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8" Feb 27 17:13:39 crc kubenswrapper[4700]: W0227 17:13:39.421626 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20d19220_2826_4990_b39e_bd24304ead43.slice/crio-f2c8e7d82adbc489ccb61d0906afee14709844f94bf890dbc1b92fc8ec17b3a2 WatchSource:0}: Error finding container f2c8e7d82adbc489ccb61d0906afee14709844f94bf890dbc1b92fc8ec17b3a2: Status 404 returned error can't find the container with id f2c8e7d82adbc489ccb61d0906afee14709844f94bf890dbc1b92fc8ec17b3a2 Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.439817 4700 scope.go:117] "RemoveContainer" containerID="b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.464804 4700 scope.go:117] "RemoveContainer" containerID="d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.485998 4700 scope.go:117] "RemoveContainer" containerID="e2e83f8a1f5d3909754673dab13333df53c0524f2c5c6b33831c04129dc65302" Feb 27 17:13:39 crc kubenswrapper[4700]: E0227 17:13:39.486502 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2e83f8a1f5d3909754673dab13333df53c0524f2c5c6b33831c04129dc65302\": container with ID starting with e2e83f8a1f5d3909754673dab13333df53c0524f2c5c6b33831c04129dc65302 not found: ID does not exist" containerID="e2e83f8a1f5d3909754673dab13333df53c0524f2c5c6b33831c04129dc65302" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.486558 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2e83f8a1f5d3909754673dab13333df53c0524f2c5c6b33831c04129dc65302"} err="failed to get container status \"e2e83f8a1f5d3909754673dab13333df53c0524f2c5c6b33831c04129dc65302\": rpc error: code = NotFound desc = could not find container \"e2e83f8a1f5d3909754673dab13333df53c0524f2c5c6b33831c04129dc65302\": container with ID starting with e2e83f8a1f5d3909754673dab13333df53c0524f2c5c6b33831c04129dc65302 not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.486591 4700 scope.go:117] "RemoveContainer" containerID="da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49" Feb 27 17:13:39 crc kubenswrapper[4700]: E0227 17:13:39.486993 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49\": container with ID starting with da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49 not found: ID does not exist" containerID="da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.487033 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49"} err="failed to get container status \"da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49\": rpc error: code = NotFound desc = could not find container \"da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49\": container with ID starting with da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49 not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.487056 4700 scope.go:117] "RemoveContainer" containerID="3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191" Feb 27 17:13:39 crc kubenswrapper[4700]: E0227 17:13:39.487598 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191\": container with ID starting with 3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191 not found: ID does not exist" containerID="3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.487642 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191"} err="failed to get container status \"3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191\": rpc error: code = NotFound desc = could not find container \"3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191\": container with ID starting with 3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191 not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.487669 4700 scope.go:117] "RemoveContainer" containerID="611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265" Feb 27 17:13:39 crc kubenswrapper[4700]: E0227 17:13:39.488068 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265\": container with ID starting with 611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265 not found: ID does not exist" containerID="611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.488108 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265"} err="failed to get container status \"611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265\": rpc error: code = NotFound desc = could not find container \"611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265\": container with ID starting with 611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265 not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.488132 4700 scope.go:117] "RemoveContainer" containerID="c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf" Feb 27 17:13:39 crc kubenswrapper[4700]: E0227 17:13:39.488493 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf\": container with ID starting with c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf not found: ID does not exist" containerID="c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.488530 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf"} err="failed to get container status \"c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf\": rpc error: code = NotFound desc = could not find container \"c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf\": container with ID starting with c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.488552 4700 scope.go:117] "RemoveContainer" containerID="6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64" Feb 27 17:13:39 crc kubenswrapper[4700]: E0227 17:13:39.488950 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64\": container with ID starting with 6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64 not found: ID does not exist" containerID="6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.488993 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64"} err="failed to get container status \"6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64\": rpc error: code = NotFound desc = could not find container \"6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64\": container with ID starting with 6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64 not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.489021 4700 scope.go:117] "RemoveContainer" containerID="a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a" Feb 27 17:13:39 crc kubenswrapper[4700]: E0227 17:13:39.489345 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a\": container with ID starting with a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a not found: ID does not exist" containerID="a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.489382 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a"} err="failed to get container status \"a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a\": rpc error: code = NotFound desc = could not find container \"a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a\": container with ID starting with a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.489407 4700 scope.go:117] "RemoveContainer" containerID="ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8" Feb 27 17:13:39 crc kubenswrapper[4700]: E0227 17:13:39.489706 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8\": container with ID starting with ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8 not found: ID does not exist" containerID="ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.489742 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8"} err="failed to get container status \"ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8\": rpc error: code = NotFound desc = could not find container \"ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8\": container with ID starting with ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8 not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.489767 4700 scope.go:117] "RemoveContainer" containerID="b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0" Feb 27 17:13:39 crc kubenswrapper[4700]: E0227 17:13:39.490299 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0\": container with ID starting with b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0 not found: ID does not exist" containerID="b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.490340 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0"} err="failed to get container status \"b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0\": rpc error: code = NotFound desc = could not find container \"b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0\": container with ID starting with b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0 not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.490365 4700 scope.go:117] "RemoveContainer" containerID="d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1" Feb 27 17:13:39 crc kubenswrapper[4700]: E0227 17:13:39.490708 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\": container with ID starting with d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1 not found: ID does not exist" containerID="d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.490747 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1"} err="failed to get container status \"d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\": rpc error: code = NotFound desc = could not find container \"d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\": container with ID starting with d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1 not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.490774 4700 scope.go:117] "RemoveContainer" containerID="e2e83f8a1f5d3909754673dab13333df53c0524f2c5c6b33831c04129dc65302" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.491105 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2e83f8a1f5d3909754673dab13333df53c0524f2c5c6b33831c04129dc65302"} err="failed to get container status \"e2e83f8a1f5d3909754673dab13333df53c0524f2c5c6b33831c04129dc65302\": rpc error: code = NotFound desc = could not find container \"e2e83f8a1f5d3909754673dab13333df53c0524f2c5c6b33831c04129dc65302\": container with ID starting with e2e83f8a1f5d3909754673dab13333df53c0524f2c5c6b33831c04129dc65302 not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.491156 4700 scope.go:117] "RemoveContainer" containerID="da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.491591 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49"} err="failed to get container status \"da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49\": rpc error: code = NotFound desc = could not find container \"da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49\": container with ID starting with da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49 not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.491628 4700 scope.go:117] "RemoveContainer" containerID="3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.491945 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191"} err="failed to get container status \"3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191\": rpc error: code = NotFound desc = could not find container \"3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191\": container with ID starting with 3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191 not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.491981 4700 scope.go:117] "RemoveContainer" containerID="611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.492295 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265"} err="failed to get container status \"611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265\": rpc error: code = NotFound desc = could not find container \"611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265\": container with ID starting with 611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265 not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.492342 4700 scope.go:117] "RemoveContainer" containerID="c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.492831 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf"} err="failed to get container status \"c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf\": rpc error: code = NotFound desc = could not find container \"c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf\": container with ID starting with c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.492867 4700 scope.go:117] "RemoveContainer" containerID="6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.493193 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64"} err="failed to get container status \"6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64\": rpc error: code = NotFound desc = could not find container \"6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64\": container with ID starting with 6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64 not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.493225 4700 scope.go:117] "RemoveContainer" containerID="a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.493547 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a"} err="failed to get container status \"a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a\": rpc error: code = NotFound desc = could not find container \"a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a\": container with ID starting with a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.493603 4700 scope.go:117] "RemoveContainer" containerID="ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.494093 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8"} err="failed to get container status \"ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8\": rpc error: code = NotFound desc = could not find container \"ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8\": container with ID starting with ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8 not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.494131 4700 scope.go:117] "RemoveContainer" containerID="b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.494509 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0"} err="failed to get container status \"b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0\": rpc error: code = NotFound desc = could not find container \"b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0\": container with ID starting with b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0 not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.494543 4700 scope.go:117] "RemoveContainer" containerID="d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.494881 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1"} err="failed to get container status \"d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\": rpc error: code = NotFound desc = could not find container \"d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\": container with ID starting with d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1 not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.494931 4700 scope.go:117] "RemoveContainer" containerID="e2e83f8a1f5d3909754673dab13333df53c0524f2c5c6b33831c04129dc65302" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.495315 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2e83f8a1f5d3909754673dab13333df53c0524f2c5c6b33831c04129dc65302"} err="failed to get container status \"e2e83f8a1f5d3909754673dab13333df53c0524f2c5c6b33831c04129dc65302\": rpc error: code = NotFound desc = could not find container \"e2e83f8a1f5d3909754673dab13333df53c0524f2c5c6b33831c04129dc65302\": container with ID starting with e2e83f8a1f5d3909754673dab13333df53c0524f2c5c6b33831c04129dc65302 not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.495354 4700 scope.go:117] "RemoveContainer" containerID="da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.496444 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49"} err="failed to get container status \"da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49\": rpc error: code = NotFound desc = could not find container \"da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49\": container with ID starting with da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49 not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.496589 4700 scope.go:117] "RemoveContainer" containerID="3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.497348 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191"} err="failed to get container status \"3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191\": rpc error: code = NotFound desc = could not find container \"3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191\": container with ID starting with 3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191 not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.497392 4700 scope.go:117] "RemoveContainer" containerID="611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.497745 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265"} err="failed to get container status \"611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265\": rpc error: code = NotFound desc = could not find container \"611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265\": container with ID starting with 611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265 not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.497781 4700 scope.go:117] "RemoveContainer" containerID="c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.498214 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf"} err="failed to get container status \"c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf\": rpc error: code = NotFound desc = could not find container \"c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf\": container with ID starting with c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.498249 4700 scope.go:117] "RemoveContainer" containerID="6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.498606 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64"} err="failed to get container status \"6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64\": rpc error: code = NotFound desc = could not find container \"6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64\": container with ID starting with 6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64 not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.498642 4700 scope.go:117] "RemoveContainer" containerID="a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.498979 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a"} err="failed to get container status \"a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a\": rpc error: code = NotFound desc = could not find container \"a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a\": container with ID starting with a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.499029 4700 scope.go:117] "RemoveContainer" containerID="ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.499401 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8"} err="failed to get container status \"ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8\": rpc error: code = NotFound desc = could not find container \"ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8\": container with ID starting with ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8 not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.499441 4700 scope.go:117] "RemoveContainer" containerID="b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.499952 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0"} err="failed to get container status \"b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0\": rpc error: code = NotFound desc = could not find container \"b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0\": container with ID starting with b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0 not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.500013 4700 scope.go:117] "RemoveContainer" containerID="d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.500553 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1"} err="failed to get container status \"d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\": rpc error: code = NotFound desc = could not find container \"d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\": container with ID starting with d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1 not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.500595 4700 scope.go:117] "RemoveContainer" containerID="e2e83f8a1f5d3909754673dab13333df53c0524f2c5c6b33831c04129dc65302" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.501082 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2e83f8a1f5d3909754673dab13333df53c0524f2c5c6b33831c04129dc65302"} err="failed to get container status \"e2e83f8a1f5d3909754673dab13333df53c0524f2c5c6b33831c04129dc65302\": rpc error: code = NotFound desc = could not find container \"e2e83f8a1f5d3909754673dab13333df53c0524f2c5c6b33831c04129dc65302\": container with ID starting with e2e83f8a1f5d3909754673dab13333df53c0524f2c5c6b33831c04129dc65302 not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.501155 4700 scope.go:117] "RemoveContainer" containerID="da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.501601 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49"} err="failed to get container status \"da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49\": rpc error: code = NotFound desc = could not find container \"da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49\": container with ID starting with da7bf1d4af431161c7640eb3bd19623d6ceb0a9cc75b02aff1b07e0d8ec03d49 not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.501649 4700 scope.go:117] "RemoveContainer" containerID="3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.502112 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191"} err="failed to get container status \"3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191\": rpc error: code = NotFound desc = could not find container \"3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191\": container with ID starting with 3a47a267e849df081b1b4c0b166450c75794a7c65a45a43e18d0676f813c5191 not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.502147 4700 scope.go:117] "RemoveContainer" containerID="611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.502622 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265"} err="failed to get container status \"611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265\": rpc error: code = NotFound desc = could not find container \"611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265\": container with ID starting with 611c33492f4ae433ce05b3bb0874da3b13b15925cbc52c904ce1020cdb517265 not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.502672 4700 scope.go:117] "RemoveContainer" containerID="c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.502998 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf"} err="failed to get container status \"c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf\": rpc error: code = NotFound desc = could not find container \"c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf\": container with ID starting with c2d080e3c8bba1838e86309a08d83497c3cfd283e915756f8d15dbe41febd7cf not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.503042 4700 scope.go:117] "RemoveContainer" containerID="6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.503452 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64"} err="failed to get container status \"6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64\": rpc error: code = NotFound desc = could not find container \"6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64\": container with ID starting with 6993e854a623892476392727897c1ea45ed6f192f796852c6ca16912b166eb64 not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.503516 4700 scope.go:117] "RemoveContainer" containerID="a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.503852 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a"} err="failed to get container status \"a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a\": rpc error: code = NotFound desc = could not find container \"a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a\": container with ID starting with a6e43cb86a02bed9c0b65b67832478e87c09e62bd123792e647b20fffceabc8a not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.504014 4700 scope.go:117] "RemoveContainer" containerID="ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.504369 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8"} err="failed to get container status \"ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8\": rpc error: code = NotFound desc = could not find container \"ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8\": container with ID starting with ee8b49219056ce9ad42564722882aaf2ffba00e88b09c30f76f0ca565b2373b8 not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.504418 4700 scope.go:117] "RemoveContainer" containerID="b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.504843 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0"} err="failed to get container status \"b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0\": rpc error: code = NotFound desc = could not find container \"b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0\": container with ID starting with b653da7e6e9482b763e17bcd5e11865cdfd4fef4baaa3c390e28c21e019871a0 not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.504891 4700 scope.go:117] "RemoveContainer" containerID="d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.505247 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1"} err="failed to get container status \"d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\": rpc error: code = NotFound desc = could not find container \"d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1\": container with ID starting with d362ca10bee883685922492050238d918fe61ae643aa54b7a79cc92765fb0fb1 not found: ID does not exist" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.505287 4700 scope.go:117] "RemoveContainer" containerID="e2e83f8a1f5d3909754673dab13333df53c0524f2c5c6b33831c04129dc65302" Feb 27 17:13:39 crc kubenswrapper[4700]: I0227 17:13:39.505682 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2e83f8a1f5d3909754673dab13333df53c0524f2c5c6b33831c04129dc65302"} err="failed to get container status \"e2e83f8a1f5d3909754673dab13333df53c0524f2c5c6b33831c04129dc65302\": rpc error: code = NotFound desc = could not find container \"e2e83f8a1f5d3909754673dab13333df53c0524f2c5c6b33831c04129dc65302\": container with ID starting with e2e83f8a1f5d3909754673dab13333df53c0524f2c5c6b33831c04129dc65302 not found: ID does not exist" Feb 27 17:13:40 crc kubenswrapper[4700]: I0227 17:13:40.236396 4700 generic.go:334] "Generic (PLEG): container finished" podID="20d19220-2826-4990-b39e-bd24304ead43" containerID="49559d2b1814c2ec94fb9ad57eb24b0f44c5abdbe607f794cce1437b4e28b386" exitCode=0 Feb 27 17:13:40 crc kubenswrapper[4700]: I0227 17:13:40.236501 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nn687" event={"ID":"20d19220-2826-4990-b39e-bd24304ead43","Type":"ContainerDied","Data":"49559d2b1814c2ec94fb9ad57eb24b0f44c5abdbe607f794cce1437b4e28b386"} Feb 27 17:13:40 crc kubenswrapper[4700]: I0227 17:13:40.236689 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nn687" event={"ID":"20d19220-2826-4990-b39e-bd24304ead43","Type":"ContainerStarted","Data":"f2c8e7d82adbc489ccb61d0906afee14709844f94bf890dbc1b92fc8ec17b3a2"} Feb 27 17:13:40 crc kubenswrapper[4700]: I0227 17:13:40.994528 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a34c0e3-2513-4e81-b6b1-80d1230475fd" path="/var/lib/kubelet/pods/3a34c0e3-2513-4e81-b6b1-80d1230475fd/volumes" Feb 27 17:13:41 crc kubenswrapper[4700]: I0227 17:13:41.252590 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nn687" event={"ID":"20d19220-2826-4990-b39e-bd24304ead43","Type":"ContainerStarted","Data":"b449910e44ebd17a93539e51ef3eea691eee7fe13f4ffaa79a6c519bd91e3884"} Feb 27 17:13:41 crc kubenswrapper[4700]: I0227 17:13:41.252977 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nn687" event={"ID":"20d19220-2826-4990-b39e-bd24304ead43","Type":"ContainerStarted","Data":"b5daf1d611dda8fa63f5ba0498ab159a9df1857e3e67d685e85043a18b5e2e6c"} Feb 27 17:13:41 crc kubenswrapper[4700]: I0227 17:13:41.253000 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nn687" event={"ID":"20d19220-2826-4990-b39e-bd24304ead43","Type":"ContainerStarted","Data":"acb658e3d639d927ca40cb1918402e8a2eef75a5eea15f61e66ce266e90da905"} Feb 27 17:13:41 crc kubenswrapper[4700]: I0227 17:13:41.253019 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nn687" event={"ID":"20d19220-2826-4990-b39e-bd24304ead43","Type":"ContainerStarted","Data":"99bf98f74a1cef8128275033d8fcfcbe0d03260ef9ce12de7b98f137e1376a37"} Feb 27 17:13:41 crc kubenswrapper[4700]: I0227 17:13:41.253035 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nn687" event={"ID":"20d19220-2826-4990-b39e-bd24304ead43","Type":"ContainerStarted","Data":"1264baf10ca71c0afed1fe4cee4873b5205b1d0fc2574b615f3d8b70599669e5"} Feb 27 17:13:41 crc kubenswrapper[4700]: I0227 17:13:41.253051 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nn687" event={"ID":"20d19220-2826-4990-b39e-bd24304ead43","Type":"ContainerStarted","Data":"c6fba3a10a420c75571ac547c3427480132af52715ed65b2a9eab8ae38b89eea"} Feb 27 17:13:42 crc kubenswrapper[4700]: I0227 17:13:42.647044 4700 scope.go:117] "RemoveContainer" containerID="b7d75a20f6b55ff385d8e95852faafad27de83d67600bb63df0f57aceea1a925" Feb 27 17:13:42 crc kubenswrapper[4700]: I0227 17:13:42.691411 4700 scope.go:117] "RemoveContainer" containerID="9c675574fe85ad3f55e0431b8d36fd1b83523ee502e675165aafad8551ba4560" Feb 27 17:13:43 crc kubenswrapper[4700]: I0227 17:13:43.270863 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-rbptz_66eb22fb-e593-40ff-9ff3-5bb32912972d/kube-multus/2.log" Feb 27 17:13:44 crc kubenswrapper[4700]: I0227 17:13:44.285996 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nn687" event={"ID":"20d19220-2826-4990-b39e-bd24304ead43","Type":"ContainerStarted","Data":"01a9d0ce8037ebae45bcd03e2674363e591fdd7071fc9b0a4cb622fba9100166"} Feb 27 17:13:46 crc kubenswrapper[4700]: I0227 17:13:46.302435 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nn687" event={"ID":"20d19220-2826-4990-b39e-bd24304ead43","Type":"ContainerStarted","Data":"82d6580b577ae9baeaa5282f29333b388b5f86931e60b3f3fddc56a9f670c81c"} Feb 27 17:13:46 crc kubenswrapper[4700]: I0227 17:13:46.303815 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:46 crc kubenswrapper[4700]: I0227 17:13:46.303846 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:46 crc kubenswrapper[4700]: I0227 17:13:46.303859 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:46 crc kubenswrapper[4700]: I0227 17:13:46.341293 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-nn687" podStartSLOduration=7.341274346 podStartE2EDuration="7.341274346s" podCreationTimestamp="2026-02-27 17:13:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:13:46.335680067 +0000 UTC m=+786.320992834" watchObservedRunningTime="2026-02-27 17:13:46.341274346 +0000 UTC m=+786.326587103" Feb 27 17:13:46 crc kubenswrapper[4700]: I0227 17:13:46.344547 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:46 crc kubenswrapper[4700]: I0227 17:13:46.349824 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:13:52 crc kubenswrapper[4700]: I0227 17:13:52.989182 4700 scope.go:117] "RemoveContainer" containerID="ebdea7f95d1c1fbe317f3b2abbfaf311938c90e4554c498e7dd55e758928933f" Feb 27 17:13:53 crc kubenswrapper[4700]: I0227 17:13:53.354062 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-rbptz_66eb22fb-e593-40ff-9ff3-5bb32912972d/kube-multus/2.log" Feb 27 17:13:53 crc kubenswrapper[4700]: I0227 17:13:53.354450 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-rbptz" event={"ID":"66eb22fb-e593-40ff-9ff3-5bb32912972d","Type":"ContainerStarted","Data":"876e953f6e282ad669dabe4d418a4c332b8c99f580f9dd0f0c7d7efdb9d2bd6d"} Feb 27 17:13:59 crc kubenswrapper[4700]: I0227 17:13:59.963866 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ct5kj"] Feb 27 17:13:59 crc kubenswrapper[4700]: I0227 17:13:59.972136 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ct5kj" Feb 27 17:13:59 crc kubenswrapper[4700]: I0227 17:13:59.973510 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ct5kj"] Feb 27 17:13:59 crc kubenswrapper[4700]: I0227 17:13:59.976402 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 27 17:14:00 crc kubenswrapper[4700]: I0227 17:14:00.153783 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536874-g79g6"] Feb 27 17:14:00 crc kubenswrapper[4700]: I0227 17:14:00.154977 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536874-g79g6" Feb 27 17:14:00 crc kubenswrapper[4700]: I0227 17:14:00.158048 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:14:00 crc kubenswrapper[4700]: I0227 17:14:00.158332 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d62kc" Feb 27 17:14:00 crc kubenswrapper[4700]: I0227 17:14:00.159917 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536874-g79g6"] Feb 27 17:14:00 crc kubenswrapper[4700]: I0227 17:14:00.162174 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:14:00 crc kubenswrapper[4700]: I0227 17:14:00.170563 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0cc8109b-159b-4c74-90af-110d79f1b6f4-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ct5kj\" (UID: \"0cc8109b-159b-4c74-90af-110d79f1b6f4\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ct5kj" Feb 27 17:14:00 crc kubenswrapper[4700]: I0227 17:14:00.170754 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpq9g\" (UniqueName: \"kubernetes.io/projected/0cc8109b-159b-4c74-90af-110d79f1b6f4-kube-api-access-zpq9g\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ct5kj\" (UID: \"0cc8109b-159b-4c74-90af-110d79f1b6f4\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ct5kj" Feb 27 17:14:00 crc kubenswrapper[4700]: I0227 17:14:00.170906 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0cc8109b-159b-4c74-90af-110d79f1b6f4-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ct5kj\" (UID: \"0cc8109b-159b-4c74-90af-110d79f1b6f4\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ct5kj" Feb 27 17:14:00 crc kubenswrapper[4700]: I0227 17:14:00.271829 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpq9g\" (UniqueName: \"kubernetes.io/projected/0cc8109b-159b-4c74-90af-110d79f1b6f4-kube-api-access-zpq9g\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ct5kj\" (UID: \"0cc8109b-159b-4c74-90af-110d79f1b6f4\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ct5kj" Feb 27 17:14:00 crc kubenswrapper[4700]: I0227 17:14:00.272075 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tgq9\" (UniqueName: \"kubernetes.io/projected/5e2f6b19-ca84-4f18-b030-420a69fc6dce-kube-api-access-6tgq9\") pod \"auto-csr-approver-29536874-g79g6\" (UID: \"5e2f6b19-ca84-4f18-b030-420a69fc6dce\") " pod="openshift-infra/auto-csr-approver-29536874-g79g6" Feb 27 17:14:00 crc kubenswrapper[4700]: I0227 17:14:00.272126 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0cc8109b-159b-4c74-90af-110d79f1b6f4-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ct5kj\" (UID: \"0cc8109b-159b-4c74-90af-110d79f1b6f4\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ct5kj" Feb 27 17:14:00 crc kubenswrapper[4700]: I0227 17:14:00.272213 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0cc8109b-159b-4c74-90af-110d79f1b6f4-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ct5kj\" (UID: \"0cc8109b-159b-4c74-90af-110d79f1b6f4\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ct5kj" Feb 27 17:14:00 crc kubenswrapper[4700]: I0227 17:14:00.272631 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0cc8109b-159b-4c74-90af-110d79f1b6f4-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ct5kj\" (UID: \"0cc8109b-159b-4c74-90af-110d79f1b6f4\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ct5kj" Feb 27 17:14:00 crc kubenswrapper[4700]: I0227 17:14:00.273779 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0cc8109b-159b-4c74-90af-110d79f1b6f4-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ct5kj\" (UID: \"0cc8109b-159b-4c74-90af-110d79f1b6f4\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ct5kj" Feb 27 17:14:00 crc kubenswrapper[4700]: I0227 17:14:00.308634 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpq9g\" (UniqueName: \"kubernetes.io/projected/0cc8109b-159b-4c74-90af-110d79f1b6f4-kube-api-access-zpq9g\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ct5kj\" (UID: \"0cc8109b-159b-4c74-90af-110d79f1b6f4\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ct5kj" Feb 27 17:14:00 crc kubenswrapper[4700]: I0227 17:14:00.373991 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tgq9\" (UniqueName: \"kubernetes.io/projected/5e2f6b19-ca84-4f18-b030-420a69fc6dce-kube-api-access-6tgq9\") pod \"auto-csr-approver-29536874-g79g6\" (UID: \"5e2f6b19-ca84-4f18-b030-420a69fc6dce\") " pod="openshift-infra/auto-csr-approver-29536874-g79g6" Feb 27 17:14:00 crc kubenswrapper[4700]: I0227 17:14:00.403342 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tgq9\" (UniqueName: \"kubernetes.io/projected/5e2f6b19-ca84-4f18-b030-420a69fc6dce-kube-api-access-6tgq9\") pod \"auto-csr-approver-29536874-g79g6\" (UID: \"5e2f6b19-ca84-4f18-b030-420a69fc6dce\") " pod="openshift-infra/auto-csr-approver-29536874-g79g6" Feb 27 17:14:00 crc kubenswrapper[4700]: I0227 17:14:00.485635 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536874-g79g6" Feb 27 17:14:00 crc kubenswrapper[4700]: I0227 17:14:00.599123 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ct5kj" Feb 27 17:14:00 crc kubenswrapper[4700]: I0227 17:14:00.761306 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536874-g79g6"] Feb 27 17:14:00 crc kubenswrapper[4700]: W0227 17:14:00.796958 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e2f6b19_ca84_4f18_b030_420a69fc6dce.slice/crio-efe8963e7b038e85f10fe309dcf161b52f3a21a5a52e60e967e02130b78c5ea8 WatchSource:0}: Error finding container efe8963e7b038e85f10fe309dcf161b52f3a21a5a52e60e967e02130b78c5ea8: Status 404 returned error can't find the container with id efe8963e7b038e85f10fe309dcf161b52f3a21a5a52e60e967e02130b78c5ea8 Feb 27 17:14:00 crc kubenswrapper[4700]: I0227 17:14:00.878128 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ct5kj"] Feb 27 17:14:00 crc kubenswrapper[4700]: W0227 17:14:00.879397 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0cc8109b_159b_4c74_90af_110d79f1b6f4.slice/crio-135bfae3d55e6331494a7788ab2a08e2bcfe256e7585bca3db2bd5e15b50fdcb WatchSource:0}: Error finding container 135bfae3d55e6331494a7788ab2a08e2bcfe256e7585bca3db2bd5e15b50fdcb: Status 404 returned error can't find the container with id 135bfae3d55e6331494a7788ab2a08e2bcfe256e7585bca3db2bd5e15b50fdcb Feb 27 17:14:01 crc kubenswrapper[4700]: I0227 17:14:01.405677 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536874-g79g6" event={"ID":"5e2f6b19-ca84-4f18-b030-420a69fc6dce","Type":"ContainerStarted","Data":"efe8963e7b038e85f10fe309dcf161b52f3a21a5a52e60e967e02130b78c5ea8"} Feb 27 17:14:01 crc kubenswrapper[4700]: I0227 17:14:01.407357 4700 generic.go:334] "Generic (PLEG): container finished" podID="0cc8109b-159b-4c74-90af-110d79f1b6f4" containerID="043131bcd6a1b37f6a803d09aac495f3f7b81d3ea02cb867fedab7eb83d68ab9" exitCode=0 Feb 27 17:14:01 crc kubenswrapper[4700]: I0227 17:14:01.407393 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ct5kj" event={"ID":"0cc8109b-159b-4c74-90af-110d79f1b6f4","Type":"ContainerDied","Data":"043131bcd6a1b37f6a803d09aac495f3f7b81d3ea02cb867fedab7eb83d68ab9"} Feb 27 17:14:01 crc kubenswrapper[4700]: I0227 17:14:01.407421 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ct5kj" event={"ID":"0cc8109b-159b-4c74-90af-110d79f1b6f4","Type":"ContainerStarted","Data":"135bfae3d55e6331494a7788ab2a08e2bcfe256e7585bca3db2bd5e15b50fdcb"} Feb 27 17:14:02 crc kubenswrapper[4700]: I0227 17:14:02.414520 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536874-g79g6" event={"ID":"5e2f6b19-ca84-4f18-b030-420a69fc6dce","Type":"ContainerStarted","Data":"c90b908abe9c5264dc998e9123b4919a8f267032d44c1b05b1861f6277e3fef6"} Feb 27 17:14:02 crc kubenswrapper[4700]: I0227 17:14:02.429147 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536874-g79g6" podStartSLOduration=1.21314669 podStartE2EDuration="2.429128166s" podCreationTimestamp="2026-02-27 17:14:00 +0000 UTC" firstStartedPulling="2026-02-27 17:14:00.799292754 +0000 UTC m=+800.784605501" lastFinishedPulling="2026-02-27 17:14:02.01527423 +0000 UTC m=+802.000586977" observedRunningTime="2026-02-27 17:14:02.427261816 +0000 UTC m=+802.412574563" watchObservedRunningTime="2026-02-27 17:14:02.429128166 +0000 UTC m=+802.414440913" Feb 27 17:14:03 crc kubenswrapper[4700]: I0227 17:14:03.440371 4700 generic.go:334] "Generic (PLEG): container finished" podID="5e2f6b19-ca84-4f18-b030-420a69fc6dce" containerID="c90b908abe9c5264dc998e9123b4919a8f267032d44c1b05b1861f6277e3fef6" exitCode=0 Feb 27 17:14:03 crc kubenswrapper[4700]: I0227 17:14:03.440435 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536874-g79g6" event={"ID":"5e2f6b19-ca84-4f18-b030-420a69fc6dce","Type":"ContainerDied","Data":"c90b908abe9c5264dc998e9123b4919a8f267032d44c1b05b1861f6277e3fef6"} Feb 27 17:14:04 crc kubenswrapper[4700]: I0227 17:14:04.752906 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536874-g79g6" Feb 27 17:14:04 crc kubenswrapper[4700]: I0227 17:14:04.867964 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6tgq9\" (UniqueName: \"kubernetes.io/projected/5e2f6b19-ca84-4f18-b030-420a69fc6dce-kube-api-access-6tgq9\") pod \"5e2f6b19-ca84-4f18-b030-420a69fc6dce\" (UID: \"5e2f6b19-ca84-4f18-b030-420a69fc6dce\") " Feb 27 17:14:04 crc kubenswrapper[4700]: I0227 17:14:04.878132 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e2f6b19-ca84-4f18-b030-420a69fc6dce-kube-api-access-6tgq9" (OuterVolumeSpecName: "kube-api-access-6tgq9") pod "5e2f6b19-ca84-4f18-b030-420a69fc6dce" (UID: "5e2f6b19-ca84-4f18-b030-420a69fc6dce"). InnerVolumeSpecName "kube-api-access-6tgq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:14:04 crc kubenswrapper[4700]: I0227 17:14:04.970228 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6tgq9\" (UniqueName: \"kubernetes.io/projected/5e2f6b19-ca84-4f18-b030-420a69fc6dce-kube-api-access-6tgq9\") on node \"crc\" DevicePath \"\"" Feb 27 17:14:05 crc kubenswrapper[4700]: I0227 17:14:05.458587 4700 generic.go:334] "Generic (PLEG): container finished" podID="0cc8109b-159b-4c74-90af-110d79f1b6f4" containerID="8fd9b2aa95ba004cdd89a66bc3936789f0ef8ba1e4a131aee04665360344e4b7" exitCode=0 Feb 27 17:14:05 crc kubenswrapper[4700]: I0227 17:14:05.458660 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ct5kj" event={"ID":"0cc8109b-159b-4c74-90af-110d79f1b6f4","Type":"ContainerDied","Data":"8fd9b2aa95ba004cdd89a66bc3936789f0ef8ba1e4a131aee04665360344e4b7"} Feb 27 17:14:05 crc kubenswrapper[4700]: I0227 17:14:05.465121 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536874-g79g6" event={"ID":"5e2f6b19-ca84-4f18-b030-420a69fc6dce","Type":"ContainerDied","Data":"efe8963e7b038e85f10fe309dcf161b52f3a21a5a52e60e967e02130b78c5ea8"} Feb 27 17:14:05 crc kubenswrapper[4700]: I0227 17:14:05.465170 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="efe8963e7b038e85f10fe309dcf161b52f3a21a5a52e60e967e02130b78c5ea8" Feb 27 17:14:05 crc kubenswrapper[4700]: I0227 17:14:05.465256 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536874-g79g6" Feb 27 17:14:05 crc kubenswrapper[4700]: I0227 17:14:05.520368 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536868-vf4cp"] Feb 27 17:14:05 crc kubenswrapper[4700]: I0227 17:14:05.527298 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536868-vf4cp"] Feb 27 17:14:06 crc kubenswrapper[4700]: I0227 17:14:06.410422 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:14:06 crc kubenswrapper[4700]: I0227 17:14:06.411038 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:14:06 crc kubenswrapper[4700]: I0227 17:14:06.476083 4700 generic.go:334] "Generic (PLEG): container finished" podID="0cc8109b-159b-4c74-90af-110d79f1b6f4" containerID="43eeb0a0e1c679a460e037471a66d462ccf768d8e10e3c283086709ef26ddf8a" exitCode=0 Feb 27 17:14:06 crc kubenswrapper[4700]: I0227 17:14:06.476183 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ct5kj" event={"ID":"0cc8109b-159b-4c74-90af-110d79f1b6f4","Type":"ContainerDied","Data":"43eeb0a0e1c679a460e037471a66d462ccf768d8e10e3c283086709ef26ddf8a"} Feb 27 17:14:06 crc kubenswrapper[4700]: I0227 17:14:06.991692 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f466047-1e51-43ae-9fa2-a29dc4ee0c40" path="/var/lib/kubelet/pods/0f466047-1e51-43ae-9fa2-a29dc4ee0c40/volumes" Feb 27 17:14:07 crc kubenswrapper[4700]: I0227 17:14:07.752794 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ct5kj" Feb 27 17:14:07 crc kubenswrapper[4700]: I0227 17:14:07.911506 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0cc8109b-159b-4c74-90af-110d79f1b6f4-util\") pod \"0cc8109b-159b-4c74-90af-110d79f1b6f4\" (UID: \"0cc8109b-159b-4c74-90af-110d79f1b6f4\") " Feb 27 17:14:07 crc kubenswrapper[4700]: I0227 17:14:07.911612 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpq9g\" (UniqueName: \"kubernetes.io/projected/0cc8109b-159b-4c74-90af-110d79f1b6f4-kube-api-access-zpq9g\") pod \"0cc8109b-159b-4c74-90af-110d79f1b6f4\" (UID: \"0cc8109b-159b-4c74-90af-110d79f1b6f4\") " Feb 27 17:14:07 crc kubenswrapper[4700]: I0227 17:14:07.911674 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0cc8109b-159b-4c74-90af-110d79f1b6f4-bundle\") pod \"0cc8109b-159b-4c74-90af-110d79f1b6f4\" (UID: \"0cc8109b-159b-4c74-90af-110d79f1b6f4\") " Feb 27 17:14:07 crc kubenswrapper[4700]: I0227 17:14:07.918589 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0cc8109b-159b-4c74-90af-110d79f1b6f4-bundle" (OuterVolumeSpecName: "bundle") pod "0cc8109b-159b-4c74-90af-110d79f1b6f4" (UID: "0cc8109b-159b-4c74-90af-110d79f1b6f4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:14:07 crc kubenswrapper[4700]: I0227 17:14:07.920277 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cc8109b-159b-4c74-90af-110d79f1b6f4-kube-api-access-zpq9g" (OuterVolumeSpecName: "kube-api-access-zpq9g") pod "0cc8109b-159b-4c74-90af-110d79f1b6f4" (UID: "0cc8109b-159b-4c74-90af-110d79f1b6f4"). InnerVolumeSpecName "kube-api-access-zpq9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:14:07 crc kubenswrapper[4700]: I0227 17:14:07.935807 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0cc8109b-159b-4c74-90af-110d79f1b6f4-util" (OuterVolumeSpecName: "util") pod "0cc8109b-159b-4c74-90af-110d79f1b6f4" (UID: "0cc8109b-159b-4c74-90af-110d79f1b6f4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:14:08 crc kubenswrapper[4700]: I0227 17:14:08.012866 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpq9g\" (UniqueName: \"kubernetes.io/projected/0cc8109b-159b-4c74-90af-110d79f1b6f4-kube-api-access-zpq9g\") on node \"crc\" DevicePath \"\"" Feb 27 17:14:08 crc kubenswrapper[4700]: I0227 17:14:08.012903 4700 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0cc8109b-159b-4c74-90af-110d79f1b6f4-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:14:08 crc kubenswrapper[4700]: I0227 17:14:08.012916 4700 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0cc8109b-159b-4c74-90af-110d79f1b6f4-util\") on node \"crc\" DevicePath \"\"" Feb 27 17:14:08 crc kubenswrapper[4700]: I0227 17:14:08.492794 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ct5kj" event={"ID":"0cc8109b-159b-4c74-90af-110d79f1b6f4","Type":"ContainerDied","Data":"135bfae3d55e6331494a7788ab2a08e2bcfe256e7585bca3db2bd5e15b50fdcb"} Feb 27 17:14:08 crc kubenswrapper[4700]: I0227 17:14:08.493143 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="135bfae3d55e6331494a7788ab2a08e2bcfe256e7585bca3db2bd5e15b50fdcb" Feb 27 17:14:08 crc kubenswrapper[4700]: I0227 17:14:08.492861 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ct5kj" Feb 27 17:14:09 crc kubenswrapper[4700]: I0227 17:14:09.398063 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-nn687" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.507347 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-mw89d"] Feb 27 17:14:17 crc kubenswrapper[4700]: E0227 17:14:17.508250 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cc8109b-159b-4c74-90af-110d79f1b6f4" containerName="extract" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.508266 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cc8109b-159b-4c74-90af-110d79f1b6f4" containerName="extract" Feb 27 17:14:17 crc kubenswrapper[4700]: E0227 17:14:17.508285 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cc8109b-159b-4c74-90af-110d79f1b6f4" containerName="util" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.508292 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cc8109b-159b-4c74-90af-110d79f1b6f4" containerName="util" Feb 27 17:14:17 crc kubenswrapper[4700]: E0227 17:14:17.508327 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cc8109b-159b-4c74-90af-110d79f1b6f4" containerName="pull" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.508334 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cc8109b-159b-4c74-90af-110d79f1b6f4" containerName="pull" Feb 27 17:14:17 crc kubenswrapper[4700]: E0227 17:14:17.508348 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e2f6b19-ca84-4f18-b030-420a69fc6dce" containerName="oc" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.508356 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e2f6b19-ca84-4f18-b030-420a69fc6dce" containerName="oc" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.508532 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cc8109b-159b-4c74-90af-110d79f1b6f4" containerName="extract" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.508574 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e2f6b19-ca84-4f18-b030-420a69fc6dce" containerName="oc" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.509133 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-mw89d" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.511400 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.511913 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.511938 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-9m654" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.544231 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6894b8c5db-82ptt"] Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.545038 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6894b8c5db-82ptt" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.551806 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.551907 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-4vvt4" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.559163 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-mw89d"] Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.570216 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6894b8c5db-82ptt"] Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.574516 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6894b8c5db-7gblc"] Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.575281 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6894b8c5db-7gblc" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.585940 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6894b8c5db-7gblc"] Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.628744 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79dpf\" (UniqueName: \"kubernetes.io/projected/530391a5-04d2-480a-8a3d-99d343188028-kube-api-access-79dpf\") pod \"obo-prometheus-operator-68bc856cb9-mw89d\" (UID: \"530391a5-04d2-480a-8a3d-99d343188028\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-mw89d" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.730054 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/97b66bd1-71dd-493c-a8cf-86b371409c8c-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6894b8c5db-82ptt\" (UID: \"97b66bd1-71dd-493c-a8cf-86b371409c8c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6894b8c5db-82ptt" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.730110 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79dpf\" (UniqueName: \"kubernetes.io/projected/530391a5-04d2-480a-8a3d-99d343188028-kube-api-access-79dpf\") pod \"obo-prometheus-operator-68bc856cb9-mw89d\" (UID: \"530391a5-04d2-480a-8a3d-99d343188028\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-mw89d" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.730217 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fd6f868c-c0c2-4d1f-ac0a-8f81da9db16f-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6894b8c5db-7gblc\" (UID: \"fd6f868c-c0c2-4d1f-ac0a-8f81da9db16f\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6894b8c5db-7gblc" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.730229 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-hc2rv"] Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.730263 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/97b66bd1-71dd-493c-a8cf-86b371409c8c-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6894b8c5db-82ptt\" (UID: \"97b66bd1-71dd-493c-a8cf-86b371409c8c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6894b8c5db-82ptt" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.730347 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fd6f868c-c0c2-4d1f-ac0a-8f81da9db16f-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6894b8c5db-7gblc\" (UID: \"fd6f868c-c0c2-4d1f-ac0a-8f81da9db16f\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6894b8c5db-7gblc" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.731006 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-hc2rv" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.738074 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-d88pf" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.738105 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.756285 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-hc2rv"] Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.766350 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79dpf\" (UniqueName: \"kubernetes.io/projected/530391a5-04d2-480a-8a3d-99d343188028-kube-api-access-79dpf\") pod \"obo-prometheus-operator-68bc856cb9-mw89d\" (UID: \"530391a5-04d2-480a-8a3d-99d343188028\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-mw89d" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.824168 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-mw89d" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.831877 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fd6f868c-c0c2-4d1f-ac0a-8f81da9db16f-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6894b8c5db-7gblc\" (UID: \"fd6f868c-c0c2-4d1f-ac0a-8f81da9db16f\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6894b8c5db-7gblc" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.831920 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/eaaadf56-7fe6-4df6-a59b-7cdcab1dc47b-observability-operator-tls\") pod \"observability-operator-59bdc8b94-hc2rv\" (UID: \"eaaadf56-7fe6-4df6-a59b-7cdcab1dc47b\") " pod="openshift-operators/observability-operator-59bdc8b94-hc2rv" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.831959 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r56zw\" (UniqueName: \"kubernetes.io/projected/eaaadf56-7fe6-4df6-a59b-7cdcab1dc47b-kube-api-access-r56zw\") pod \"observability-operator-59bdc8b94-hc2rv\" (UID: \"eaaadf56-7fe6-4df6-a59b-7cdcab1dc47b\") " pod="openshift-operators/observability-operator-59bdc8b94-hc2rv" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.831988 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/97b66bd1-71dd-493c-a8cf-86b371409c8c-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6894b8c5db-82ptt\" (UID: \"97b66bd1-71dd-493c-a8cf-86b371409c8c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6894b8c5db-82ptt" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.832031 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fd6f868c-c0c2-4d1f-ac0a-8f81da9db16f-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6894b8c5db-7gblc\" (UID: \"fd6f868c-c0c2-4d1f-ac0a-8f81da9db16f\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6894b8c5db-7gblc" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.832091 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/97b66bd1-71dd-493c-a8cf-86b371409c8c-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6894b8c5db-82ptt\" (UID: \"97b66bd1-71dd-493c-a8cf-86b371409c8c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6894b8c5db-82ptt" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.836861 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fd6f868c-c0c2-4d1f-ac0a-8f81da9db16f-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6894b8c5db-7gblc\" (UID: \"fd6f868c-c0c2-4d1f-ac0a-8f81da9db16f\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6894b8c5db-7gblc" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.836868 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/97b66bd1-71dd-493c-a8cf-86b371409c8c-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6894b8c5db-82ptt\" (UID: \"97b66bd1-71dd-493c-a8cf-86b371409c8c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6894b8c5db-82ptt" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.838847 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/97b66bd1-71dd-493c-a8cf-86b371409c8c-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6894b8c5db-82ptt\" (UID: \"97b66bd1-71dd-493c-a8cf-86b371409c8c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6894b8c5db-82ptt" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.839221 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fd6f868c-c0c2-4d1f-ac0a-8f81da9db16f-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6894b8c5db-7gblc\" (UID: \"fd6f868c-c0c2-4d1f-ac0a-8f81da9db16f\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6894b8c5db-7gblc" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.859241 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6894b8c5db-82ptt" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.898705 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6894b8c5db-7gblc" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.935248 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/eaaadf56-7fe6-4df6-a59b-7cdcab1dc47b-observability-operator-tls\") pod \"observability-operator-59bdc8b94-hc2rv\" (UID: \"eaaadf56-7fe6-4df6-a59b-7cdcab1dc47b\") " pod="openshift-operators/observability-operator-59bdc8b94-hc2rv" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.935285 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r56zw\" (UniqueName: \"kubernetes.io/projected/eaaadf56-7fe6-4df6-a59b-7cdcab1dc47b-kube-api-access-r56zw\") pod \"observability-operator-59bdc8b94-hc2rv\" (UID: \"eaaadf56-7fe6-4df6-a59b-7cdcab1dc47b\") " pod="openshift-operators/observability-operator-59bdc8b94-hc2rv" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.942398 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/eaaadf56-7fe6-4df6-a59b-7cdcab1dc47b-observability-operator-tls\") pod \"observability-operator-59bdc8b94-hc2rv\" (UID: \"eaaadf56-7fe6-4df6-a59b-7cdcab1dc47b\") " pod="openshift-operators/observability-operator-59bdc8b94-hc2rv" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.953089 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r56zw\" (UniqueName: \"kubernetes.io/projected/eaaadf56-7fe6-4df6-a59b-7cdcab1dc47b-kube-api-access-r56zw\") pod \"observability-operator-59bdc8b94-hc2rv\" (UID: \"eaaadf56-7fe6-4df6-a59b-7cdcab1dc47b\") " pod="openshift-operators/observability-operator-59bdc8b94-hc2rv" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.953517 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-z6mk4"] Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.956003 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-z6mk4" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.965079 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-jb64r" Feb 27 17:14:17 crc kubenswrapper[4700]: I0227 17:14:17.966149 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-z6mk4"] Feb 27 17:14:18 crc kubenswrapper[4700]: I0227 17:14:18.045617 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-mw89d"] Feb 27 17:14:18 crc kubenswrapper[4700]: I0227 17:14:18.046984 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-hc2rv" Feb 27 17:14:18 crc kubenswrapper[4700]: I0227 17:14:18.137589 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/c4b67ce7-9585-4c61-a8d7-bf37151fb6ea-openshift-service-ca\") pod \"perses-operator-5bf474d74f-z6mk4\" (UID: \"c4b67ce7-9585-4c61-a8d7-bf37151fb6ea\") " pod="openshift-operators/perses-operator-5bf474d74f-z6mk4" Feb 27 17:14:18 crc kubenswrapper[4700]: I0227 17:14:18.138031 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brc2f\" (UniqueName: \"kubernetes.io/projected/c4b67ce7-9585-4c61-a8d7-bf37151fb6ea-kube-api-access-brc2f\") pod \"perses-operator-5bf474d74f-z6mk4\" (UID: \"c4b67ce7-9585-4c61-a8d7-bf37151fb6ea\") " pod="openshift-operators/perses-operator-5bf474d74f-z6mk4" Feb 27 17:14:18 crc kubenswrapper[4700]: I0227 17:14:18.239724 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brc2f\" (UniqueName: \"kubernetes.io/projected/c4b67ce7-9585-4c61-a8d7-bf37151fb6ea-kube-api-access-brc2f\") pod \"perses-operator-5bf474d74f-z6mk4\" (UID: \"c4b67ce7-9585-4c61-a8d7-bf37151fb6ea\") " pod="openshift-operators/perses-operator-5bf474d74f-z6mk4" Feb 27 17:14:18 crc kubenswrapper[4700]: I0227 17:14:18.239780 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/c4b67ce7-9585-4c61-a8d7-bf37151fb6ea-openshift-service-ca\") pod \"perses-operator-5bf474d74f-z6mk4\" (UID: \"c4b67ce7-9585-4c61-a8d7-bf37151fb6ea\") " pod="openshift-operators/perses-operator-5bf474d74f-z6mk4" Feb 27 17:14:18 crc kubenswrapper[4700]: I0227 17:14:18.240648 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/c4b67ce7-9585-4c61-a8d7-bf37151fb6ea-openshift-service-ca\") pod \"perses-operator-5bf474d74f-z6mk4\" (UID: \"c4b67ce7-9585-4c61-a8d7-bf37151fb6ea\") " pod="openshift-operators/perses-operator-5bf474d74f-z6mk4" Feb 27 17:14:18 crc kubenswrapper[4700]: I0227 17:14:18.268001 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brc2f\" (UniqueName: \"kubernetes.io/projected/c4b67ce7-9585-4c61-a8d7-bf37151fb6ea-kube-api-access-brc2f\") pod \"perses-operator-5bf474d74f-z6mk4\" (UID: \"c4b67ce7-9585-4c61-a8d7-bf37151fb6ea\") " pod="openshift-operators/perses-operator-5bf474d74f-z6mk4" Feb 27 17:14:18 crc kubenswrapper[4700]: I0227 17:14:18.277265 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6894b8c5db-82ptt"] Feb 27 17:14:18 crc kubenswrapper[4700]: I0227 17:14:18.281656 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-z6mk4" Feb 27 17:14:18 crc kubenswrapper[4700]: I0227 17:14:18.310020 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6894b8c5db-7gblc"] Feb 27 17:14:18 crc kubenswrapper[4700]: I0227 17:14:18.486296 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-hc2rv"] Feb 27 17:14:18 crc kubenswrapper[4700]: W0227 17:14:18.489408 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeaaadf56_7fe6_4df6_a59b_7cdcab1dc47b.slice/crio-44544fc235653e5c588c1bcc4da0d1a9db84e70eb16574fdc31c7eec93b9202d WatchSource:0}: Error finding container 44544fc235653e5c588c1bcc4da0d1a9db84e70eb16574fdc31c7eec93b9202d: Status 404 returned error can't find the container with id 44544fc235653e5c588c1bcc4da0d1a9db84e70eb16574fdc31c7eec93b9202d Feb 27 17:14:18 crc kubenswrapper[4700]: I0227 17:14:18.573741 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-hc2rv" event={"ID":"eaaadf56-7fe6-4df6-a59b-7cdcab1dc47b","Type":"ContainerStarted","Data":"44544fc235653e5c588c1bcc4da0d1a9db84e70eb16574fdc31c7eec93b9202d"} Feb 27 17:14:18 crc kubenswrapper[4700]: I0227 17:14:18.575180 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6894b8c5db-82ptt" event={"ID":"97b66bd1-71dd-493c-a8cf-86b371409c8c","Type":"ContainerStarted","Data":"13312a57bc0b93b7fed8f142f42244f09fed8337454076d3c3b47ced7172491c"} Feb 27 17:14:18 crc kubenswrapper[4700]: I0227 17:14:18.576635 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-mw89d" event={"ID":"530391a5-04d2-480a-8a3d-99d343188028","Type":"ContainerStarted","Data":"4cf3fb66970a45a1b3e8430c4e785eb43b616d9b6aea0828e5a00536677b274b"} Feb 27 17:14:18 crc kubenswrapper[4700]: I0227 17:14:18.585384 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6894b8c5db-7gblc" event={"ID":"fd6f868c-c0c2-4d1f-ac0a-8f81da9db16f","Type":"ContainerStarted","Data":"7dfd572deec42ecaca14a986edc5aa28ff72189804fa4b368b26eab2b56e1118"} Feb 27 17:14:18 crc kubenswrapper[4700]: I0227 17:14:18.762236 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-z6mk4"] Feb 27 17:14:19 crc kubenswrapper[4700]: I0227 17:14:19.592459 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-z6mk4" event={"ID":"c4b67ce7-9585-4c61-a8d7-bf37151fb6ea","Type":"ContainerStarted","Data":"f9617e8cee990efda567cad84aeba1f320ce4ca41036a33914565dffc8f7a6e4"} Feb 27 17:14:20 crc kubenswrapper[4700]: I0227 17:14:20.769866 4700 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 27 17:14:26 crc kubenswrapper[4700]: I0227 17:14:26.181827 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-clzsx"] Feb 27 17:14:26 crc kubenswrapper[4700]: I0227 17:14:26.183083 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-clzsx" Feb 27 17:14:26 crc kubenswrapper[4700]: I0227 17:14:26.193928 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-clzsx"] Feb 27 17:14:26 crc kubenswrapper[4700]: I0227 17:14:26.261437 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c4faed5-1a11-4e87-ab27-544d8ec596e0-utilities\") pod \"community-operators-clzsx\" (UID: \"8c4faed5-1a11-4e87-ab27-544d8ec596e0\") " pod="openshift-marketplace/community-operators-clzsx" Feb 27 17:14:26 crc kubenswrapper[4700]: I0227 17:14:26.261542 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c4faed5-1a11-4e87-ab27-544d8ec596e0-catalog-content\") pod \"community-operators-clzsx\" (UID: \"8c4faed5-1a11-4e87-ab27-544d8ec596e0\") " pod="openshift-marketplace/community-operators-clzsx" Feb 27 17:14:26 crc kubenswrapper[4700]: I0227 17:14:26.261587 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbl9q\" (UniqueName: \"kubernetes.io/projected/8c4faed5-1a11-4e87-ab27-544d8ec596e0-kube-api-access-jbl9q\") pod \"community-operators-clzsx\" (UID: \"8c4faed5-1a11-4e87-ab27-544d8ec596e0\") " pod="openshift-marketplace/community-operators-clzsx" Feb 27 17:14:26 crc kubenswrapper[4700]: I0227 17:14:26.363442 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbl9q\" (UniqueName: \"kubernetes.io/projected/8c4faed5-1a11-4e87-ab27-544d8ec596e0-kube-api-access-jbl9q\") pod \"community-operators-clzsx\" (UID: \"8c4faed5-1a11-4e87-ab27-544d8ec596e0\") " pod="openshift-marketplace/community-operators-clzsx" Feb 27 17:14:26 crc kubenswrapper[4700]: I0227 17:14:26.364682 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c4faed5-1a11-4e87-ab27-544d8ec596e0-utilities\") pod \"community-operators-clzsx\" (UID: \"8c4faed5-1a11-4e87-ab27-544d8ec596e0\") " pod="openshift-marketplace/community-operators-clzsx" Feb 27 17:14:26 crc kubenswrapper[4700]: I0227 17:14:26.363607 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c4faed5-1a11-4e87-ab27-544d8ec596e0-utilities\") pod \"community-operators-clzsx\" (UID: \"8c4faed5-1a11-4e87-ab27-544d8ec596e0\") " pod="openshift-marketplace/community-operators-clzsx" Feb 27 17:14:26 crc kubenswrapper[4700]: I0227 17:14:26.365907 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c4faed5-1a11-4e87-ab27-544d8ec596e0-catalog-content\") pod \"community-operators-clzsx\" (UID: \"8c4faed5-1a11-4e87-ab27-544d8ec596e0\") " pod="openshift-marketplace/community-operators-clzsx" Feb 27 17:14:26 crc kubenswrapper[4700]: I0227 17:14:26.366261 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c4faed5-1a11-4e87-ab27-544d8ec596e0-catalog-content\") pod \"community-operators-clzsx\" (UID: \"8c4faed5-1a11-4e87-ab27-544d8ec596e0\") " pod="openshift-marketplace/community-operators-clzsx" Feb 27 17:14:26 crc kubenswrapper[4700]: I0227 17:14:26.386936 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbl9q\" (UniqueName: \"kubernetes.io/projected/8c4faed5-1a11-4e87-ab27-544d8ec596e0-kube-api-access-jbl9q\") pod \"community-operators-clzsx\" (UID: \"8c4faed5-1a11-4e87-ab27-544d8ec596e0\") " pod="openshift-marketplace/community-operators-clzsx" Feb 27 17:14:26 crc kubenswrapper[4700]: I0227 17:14:26.502080 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-clzsx" Feb 27 17:14:28 crc kubenswrapper[4700]: I0227 17:14:28.371700 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-clzsx"] Feb 27 17:14:28 crc kubenswrapper[4700]: I0227 17:14:28.652145 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6894b8c5db-82ptt" event={"ID":"97b66bd1-71dd-493c-a8cf-86b371409c8c","Type":"ContainerStarted","Data":"25f3cd1569418963891b04007391a95dc367ca3de059be7c62480df8b838a951"} Feb 27 17:14:28 crc kubenswrapper[4700]: I0227 17:14:28.655203 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-mw89d" event={"ID":"530391a5-04d2-480a-8a3d-99d343188028","Type":"ContainerStarted","Data":"eeefa09c834cd2b56acecf278f51c9d862889049481b41984002371e20998e8d"} Feb 27 17:14:28 crc kubenswrapper[4700]: I0227 17:14:28.658906 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6894b8c5db-7gblc" event={"ID":"fd6f868c-c0c2-4d1f-ac0a-8f81da9db16f","Type":"ContainerStarted","Data":"df9032f89e2534900064f485c0c40238853b22bd8e1f810e538d1dde87db1bdb"} Feb 27 17:14:28 crc kubenswrapper[4700]: I0227 17:14:28.662542 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-z6mk4" event={"ID":"c4b67ce7-9585-4c61-a8d7-bf37151fb6ea","Type":"ContainerStarted","Data":"d357f547ff9c3b29c09e71592e8854c6d0247caa80e0968ecad69f33f8c502cc"} Feb 27 17:14:28 crc kubenswrapper[4700]: I0227 17:14:28.663230 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5bf474d74f-z6mk4" Feb 27 17:14:28 crc kubenswrapper[4700]: I0227 17:14:28.664822 4700 generic.go:334] "Generic (PLEG): container finished" podID="8c4faed5-1a11-4e87-ab27-544d8ec596e0" containerID="5025df38757817edfbb941340afa94defab1191403b07138e647fe8922c53396" exitCode=0 Feb 27 17:14:28 crc kubenswrapper[4700]: I0227 17:14:28.664927 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-clzsx" event={"ID":"8c4faed5-1a11-4e87-ab27-544d8ec596e0","Type":"ContainerDied","Data":"5025df38757817edfbb941340afa94defab1191403b07138e647fe8922c53396"} Feb 27 17:14:28 crc kubenswrapper[4700]: I0227 17:14:28.664959 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-clzsx" event={"ID":"8c4faed5-1a11-4e87-ab27-544d8ec596e0","Type":"ContainerStarted","Data":"cc2323ebd7c223a75e99adf5e5e35f17f141d6343b2c4e251822e8987c85d69b"} Feb 27 17:14:28 crc kubenswrapper[4700]: I0227 17:14:28.666798 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-hc2rv" event={"ID":"eaaadf56-7fe6-4df6-a59b-7cdcab1dc47b","Type":"ContainerStarted","Data":"a95f3d927ddb4c277357b42f246b80302a6849e427dc438ffcdf5f0d96dcf7bc"} Feb 27 17:14:28 crc kubenswrapper[4700]: I0227 17:14:28.667695 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-59bdc8b94-hc2rv" Feb 27 17:14:28 crc kubenswrapper[4700]: I0227 17:14:28.671262 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-59bdc8b94-hc2rv" Feb 27 17:14:28 crc kubenswrapper[4700]: I0227 17:14:28.692650 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6894b8c5db-82ptt" podStartSLOduration=1.9488261489999998 podStartE2EDuration="11.692616458s" podCreationTimestamp="2026-02-27 17:14:17 +0000 UTC" firstStartedPulling="2026-02-27 17:14:18.334529008 +0000 UTC m=+818.319841755" lastFinishedPulling="2026-02-27 17:14:28.078319317 +0000 UTC m=+828.063632064" observedRunningTime="2026-02-27 17:14:28.688078717 +0000 UTC m=+828.673391504" watchObservedRunningTime="2026-02-27 17:14:28.692616458 +0000 UTC m=+828.677929275" Feb 27 17:14:28 crc kubenswrapper[4700]: I0227 17:14:28.726611 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-mw89d" podStartSLOduration=1.714592133 podStartE2EDuration="11.726590792s" podCreationTimestamp="2026-02-27 17:14:17 +0000 UTC" firstStartedPulling="2026-02-27 17:14:18.067989703 +0000 UTC m=+818.053302470" lastFinishedPulling="2026-02-27 17:14:28.079988382 +0000 UTC m=+828.065301129" observedRunningTime="2026-02-27 17:14:28.723390157 +0000 UTC m=+828.708702914" watchObservedRunningTime="2026-02-27 17:14:28.726590792 +0000 UTC m=+828.711903549" Feb 27 17:14:28 crc kubenswrapper[4700]: I0227 17:14:28.753060 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-59bdc8b94-hc2rv" podStartSLOduration=2.101556364 podStartE2EDuration="11.753038356s" podCreationTimestamp="2026-02-27 17:14:17 +0000 UTC" firstStartedPulling="2026-02-27 17:14:18.492968065 +0000 UTC m=+818.478280812" lastFinishedPulling="2026-02-27 17:14:28.144450057 +0000 UTC m=+828.129762804" observedRunningTime="2026-02-27 17:14:28.752146562 +0000 UTC m=+828.737459319" watchObservedRunningTime="2026-02-27 17:14:28.753038356 +0000 UTC m=+828.738351103" Feb 27 17:14:28 crc kubenswrapper[4700]: I0227 17:14:28.795489 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6894b8c5db-7gblc" podStartSLOduration=2.040312573 podStartE2EDuration="11.795448215s" podCreationTimestamp="2026-02-27 17:14:17 +0000 UTC" firstStartedPulling="2026-02-27 17:14:18.335035181 +0000 UTC m=+818.320347928" lastFinishedPulling="2026-02-27 17:14:28.090170823 +0000 UTC m=+828.075483570" observedRunningTime="2026-02-27 17:14:28.791744927 +0000 UTC m=+828.777057674" watchObservedRunningTime="2026-02-27 17:14:28.795448215 +0000 UTC m=+828.780760982" Feb 27 17:14:28 crc kubenswrapper[4700]: I0227 17:14:28.837361 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5bf474d74f-z6mk4" podStartSLOduration=2.529058542 podStartE2EDuration="11.83734659s" podCreationTimestamp="2026-02-27 17:14:17 +0000 UTC" firstStartedPulling="2026-02-27 17:14:18.770545833 +0000 UTC m=+818.755858580" lastFinishedPulling="2026-02-27 17:14:28.078833881 +0000 UTC m=+828.064146628" observedRunningTime="2026-02-27 17:14:28.8362204 +0000 UTC m=+828.821533167" watchObservedRunningTime="2026-02-27 17:14:28.83734659 +0000 UTC m=+828.822659337" Feb 27 17:14:29 crc kubenswrapper[4700]: I0227 17:14:29.673060 4700 generic.go:334] "Generic (PLEG): container finished" podID="8c4faed5-1a11-4e87-ab27-544d8ec596e0" containerID="27f12beb506793e9b280999197c8b9a5694446f8c4ee342b8d9c97ca6993ceed" exitCode=0 Feb 27 17:14:29 crc kubenswrapper[4700]: I0227 17:14:29.673249 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-clzsx" event={"ID":"8c4faed5-1a11-4e87-ab27-544d8ec596e0","Type":"ContainerDied","Data":"27f12beb506793e9b280999197c8b9a5694446f8c4ee342b8d9c97ca6993ceed"} Feb 27 17:14:31 crc kubenswrapper[4700]: I0227 17:14:31.686538 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-clzsx" event={"ID":"8c4faed5-1a11-4e87-ab27-544d8ec596e0","Type":"ContainerStarted","Data":"16e059723365325373df1d45fea4a83b90f2f5ca99b1c7efde971b45da29ed44"} Feb 27 17:14:31 crc kubenswrapper[4700]: I0227 17:14:31.731243 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-clzsx" podStartSLOduration=3.506522291 podStartE2EDuration="5.731227506s" podCreationTimestamp="2026-02-27 17:14:26 +0000 UTC" firstStartedPulling="2026-02-27 17:14:28.666303128 +0000 UTC m=+828.651615875" lastFinishedPulling="2026-02-27 17:14:30.891008333 +0000 UTC m=+830.876321090" observedRunningTime="2026-02-27 17:14:31.727166308 +0000 UTC m=+831.712479055" watchObservedRunningTime="2026-02-27 17:14:31.731227506 +0000 UTC m=+831.716540253" Feb 27 17:14:36 crc kubenswrapper[4700]: I0227 17:14:36.410674 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:14:36 crc kubenswrapper[4700]: I0227 17:14:36.411315 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:14:36 crc kubenswrapper[4700]: I0227 17:14:36.503496 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-clzsx" Feb 27 17:14:36 crc kubenswrapper[4700]: I0227 17:14:36.503544 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-clzsx" Feb 27 17:14:36 crc kubenswrapper[4700]: I0227 17:14:36.567493 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-clzsx" Feb 27 17:14:36 crc kubenswrapper[4700]: I0227 17:14:36.748919 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-clzsx" Feb 27 17:14:38 crc kubenswrapper[4700]: I0227 17:14:38.284747 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5bf474d74f-z6mk4" Feb 27 17:14:38 crc kubenswrapper[4700]: I0227 17:14:38.953634 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-clzsx"] Feb 27 17:14:38 crc kubenswrapper[4700]: I0227 17:14:38.953839 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-clzsx" podUID="8c4faed5-1a11-4e87-ab27-544d8ec596e0" containerName="registry-server" containerID="cri-o://16e059723365325373df1d45fea4a83b90f2f5ca99b1c7efde971b45da29ed44" gracePeriod=2 Feb 27 17:14:39 crc kubenswrapper[4700]: I0227 17:14:39.351732 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-clzsx" Feb 27 17:14:39 crc kubenswrapper[4700]: I0227 17:14:39.435851 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c4faed5-1a11-4e87-ab27-544d8ec596e0-utilities\") pod \"8c4faed5-1a11-4e87-ab27-544d8ec596e0\" (UID: \"8c4faed5-1a11-4e87-ab27-544d8ec596e0\") " Feb 27 17:14:39 crc kubenswrapper[4700]: I0227 17:14:39.435964 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c4faed5-1a11-4e87-ab27-544d8ec596e0-catalog-content\") pod \"8c4faed5-1a11-4e87-ab27-544d8ec596e0\" (UID: \"8c4faed5-1a11-4e87-ab27-544d8ec596e0\") " Feb 27 17:14:39 crc kubenswrapper[4700]: I0227 17:14:39.436008 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbl9q\" (UniqueName: \"kubernetes.io/projected/8c4faed5-1a11-4e87-ab27-544d8ec596e0-kube-api-access-jbl9q\") pod \"8c4faed5-1a11-4e87-ab27-544d8ec596e0\" (UID: \"8c4faed5-1a11-4e87-ab27-544d8ec596e0\") " Feb 27 17:14:39 crc kubenswrapper[4700]: I0227 17:14:39.437907 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c4faed5-1a11-4e87-ab27-544d8ec596e0-utilities" (OuterVolumeSpecName: "utilities") pod "8c4faed5-1a11-4e87-ab27-544d8ec596e0" (UID: "8c4faed5-1a11-4e87-ab27-544d8ec596e0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:14:39 crc kubenswrapper[4700]: I0227 17:14:39.444177 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c4faed5-1a11-4e87-ab27-544d8ec596e0-kube-api-access-jbl9q" (OuterVolumeSpecName: "kube-api-access-jbl9q") pod "8c4faed5-1a11-4e87-ab27-544d8ec596e0" (UID: "8c4faed5-1a11-4e87-ab27-544d8ec596e0"). InnerVolumeSpecName "kube-api-access-jbl9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:14:39 crc kubenswrapper[4700]: I0227 17:14:39.526437 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c4faed5-1a11-4e87-ab27-544d8ec596e0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8c4faed5-1a11-4e87-ab27-544d8ec596e0" (UID: "8c4faed5-1a11-4e87-ab27-544d8ec596e0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:14:39 crc kubenswrapper[4700]: I0227 17:14:39.536978 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c4faed5-1a11-4e87-ab27-544d8ec596e0-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:14:39 crc kubenswrapper[4700]: I0227 17:14:39.537011 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbl9q\" (UniqueName: \"kubernetes.io/projected/8c4faed5-1a11-4e87-ab27-544d8ec596e0-kube-api-access-jbl9q\") on node \"crc\" DevicePath \"\"" Feb 27 17:14:39 crc kubenswrapper[4700]: I0227 17:14:39.537022 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c4faed5-1a11-4e87-ab27-544d8ec596e0-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:14:39 crc kubenswrapper[4700]: I0227 17:14:39.731395 4700 generic.go:334] "Generic (PLEG): container finished" podID="8c4faed5-1a11-4e87-ab27-544d8ec596e0" containerID="16e059723365325373df1d45fea4a83b90f2f5ca99b1c7efde971b45da29ed44" exitCode=0 Feb 27 17:14:39 crc kubenswrapper[4700]: I0227 17:14:39.731446 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-clzsx" event={"ID":"8c4faed5-1a11-4e87-ab27-544d8ec596e0","Type":"ContainerDied","Data":"16e059723365325373df1d45fea4a83b90f2f5ca99b1c7efde971b45da29ed44"} Feb 27 17:14:39 crc kubenswrapper[4700]: I0227 17:14:39.731538 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-clzsx" event={"ID":"8c4faed5-1a11-4e87-ab27-544d8ec596e0","Type":"ContainerDied","Data":"cc2323ebd7c223a75e99adf5e5e35f17f141d6343b2c4e251822e8987c85d69b"} Feb 27 17:14:39 crc kubenswrapper[4700]: I0227 17:14:39.731558 4700 scope.go:117] "RemoveContainer" containerID="16e059723365325373df1d45fea4a83b90f2f5ca99b1c7efde971b45da29ed44" Feb 27 17:14:39 crc kubenswrapper[4700]: I0227 17:14:39.731483 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-clzsx" Feb 27 17:14:39 crc kubenswrapper[4700]: I0227 17:14:39.766490 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-clzsx"] Feb 27 17:14:39 crc kubenswrapper[4700]: I0227 17:14:39.766676 4700 scope.go:117] "RemoveContainer" containerID="27f12beb506793e9b280999197c8b9a5694446f8c4ee342b8d9c97ca6993ceed" Feb 27 17:14:39 crc kubenswrapper[4700]: I0227 17:14:39.771075 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-clzsx"] Feb 27 17:14:39 crc kubenswrapper[4700]: I0227 17:14:39.785643 4700 scope.go:117] "RemoveContainer" containerID="5025df38757817edfbb941340afa94defab1191403b07138e647fe8922c53396" Feb 27 17:14:39 crc kubenswrapper[4700]: I0227 17:14:39.804290 4700 scope.go:117] "RemoveContainer" containerID="16e059723365325373df1d45fea4a83b90f2f5ca99b1c7efde971b45da29ed44" Feb 27 17:14:39 crc kubenswrapper[4700]: E0227 17:14:39.804855 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16e059723365325373df1d45fea4a83b90f2f5ca99b1c7efde971b45da29ed44\": container with ID starting with 16e059723365325373df1d45fea4a83b90f2f5ca99b1c7efde971b45da29ed44 not found: ID does not exist" containerID="16e059723365325373df1d45fea4a83b90f2f5ca99b1c7efde971b45da29ed44" Feb 27 17:14:39 crc kubenswrapper[4700]: I0227 17:14:39.804898 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16e059723365325373df1d45fea4a83b90f2f5ca99b1c7efde971b45da29ed44"} err="failed to get container status \"16e059723365325373df1d45fea4a83b90f2f5ca99b1c7efde971b45da29ed44\": rpc error: code = NotFound desc = could not find container \"16e059723365325373df1d45fea4a83b90f2f5ca99b1c7efde971b45da29ed44\": container with ID starting with 16e059723365325373df1d45fea4a83b90f2f5ca99b1c7efde971b45da29ed44 not found: ID does not exist" Feb 27 17:14:39 crc kubenswrapper[4700]: I0227 17:14:39.804925 4700 scope.go:117] "RemoveContainer" containerID="27f12beb506793e9b280999197c8b9a5694446f8c4ee342b8d9c97ca6993ceed" Feb 27 17:14:39 crc kubenswrapper[4700]: E0227 17:14:39.805324 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27f12beb506793e9b280999197c8b9a5694446f8c4ee342b8d9c97ca6993ceed\": container with ID starting with 27f12beb506793e9b280999197c8b9a5694446f8c4ee342b8d9c97ca6993ceed not found: ID does not exist" containerID="27f12beb506793e9b280999197c8b9a5694446f8c4ee342b8d9c97ca6993ceed" Feb 27 17:14:39 crc kubenswrapper[4700]: I0227 17:14:39.805356 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27f12beb506793e9b280999197c8b9a5694446f8c4ee342b8d9c97ca6993ceed"} err="failed to get container status \"27f12beb506793e9b280999197c8b9a5694446f8c4ee342b8d9c97ca6993ceed\": rpc error: code = NotFound desc = could not find container \"27f12beb506793e9b280999197c8b9a5694446f8c4ee342b8d9c97ca6993ceed\": container with ID starting with 27f12beb506793e9b280999197c8b9a5694446f8c4ee342b8d9c97ca6993ceed not found: ID does not exist" Feb 27 17:14:39 crc kubenswrapper[4700]: I0227 17:14:39.805384 4700 scope.go:117] "RemoveContainer" containerID="5025df38757817edfbb941340afa94defab1191403b07138e647fe8922c53396" Feb 27 17:14:39 crc kubenswrapper[4700]: E0227 17:14:39.805651 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5025df38757817edfbb941340afa94defab1191403b07138e647fe8922c53396\": container with ID starting with 5025df38757817edfbb941340afa94defab1191403b07138e647fe8922c53396 not found: ID does not exist" containerID="5025df38757817edfbb941340afa94defab1191403b07138e647fe8922c53396" Feb 27 17:14:39 crc kubenswrapper[4700]: I0227 17:14:39.805673 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5025df38757817edfbb941340afa94defab1191403b07138e647fe8922c53396"} err="failed to get container status \"5025df38757817edfbb941340afa94defab1191403b07138e647fe8922c53396\": rpc error: code = NotFound desc = could not find container \"5025df38757817edfbb941340afa94defab1191403b07138e647fe8922c53396\": container with ID starting with 5025df38757817edfbb941340afa94defab1191403b07138e647fe8922c53396 not found: ID does not exist" Feb 27 17:14:40 crc kubenswrapper[4700]: I0227 17:14:40.988948 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c4faed5-1a11-4e87-ab27-544d8ec596e0" path="/var/lib/kubelet/pods/8c4faed5-1a11-4e87-ab27-544d8ec596e0/volumes" Feb 27 17:14:43 crc kubenswrapper[4700]: I0227 17:14:43.269269 4700 scope.go:117] "RemoveContainer" containerID="7f88d980ae85d9c36365402938b1ecfccd20e654a5be3321bb42802a5863fcea" Feb 27 17:14:55 crc kubenswrapper[4700]: I0227 17:14:55.618858 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp"] Feb 27 17:14:55 crc kubenswrapper[4700]: E0227 17:14:55.619956 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c4faed5-1a11-4e87-ab27-544d8ec596e0" containerName="registry-server" Feb 27 17:14:55 crc kubenswrapper[4700]: I0227 17:14:55.619980 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c4faed5-1a11-4e87-ab27-544d8ec596e0" containerName="registry-server" Feb 27 17:14:55 crc kubenswrapper[4700]: E0227 17:14:55.620021 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c4faed5-1a11-4e87-ab27-544d8ec596e0" containerName="extract-content" Feb 27 17:14:55 crc kubenswrapper[4700]: I0227 17:14:55.620033 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c4faed5-1a11-4e87-ab27-544d8ec596e0" containerName="extract-content" Feb 27 17:14:55 crc kubenswrapper[4700]: E0227 17:14:55.620054 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c4faed5-1a11-4e87-ab27-544d8ec596e0" containerName="extract-utilities" Feb 27 17:14:55 crc kubenswrapper[4700]: I0227 17:14:55.620066 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c4faed5-1a11-4e87-ab27-544d8ec596e0" containerName="extract-utilities" Feb 27 17:14:55 crc kubenswrapper[4700]: I0227 17:14:55.620247 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c4faed5-1a11-4e87-ab27-544d8ec596e0" containerName="registry-server" Feb 27 17:14:55 crc kubenswrapper[4700]: I0227 17:14:55.621610 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp" Feb 27 17:14:55 crc kubenswrapper[4700]: I0227 17:14:55.623872 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 27 17:14:55 crc kubenswrapper[4700]: I0227 17:14:55.638721 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp"] Feb 27 17:14:55 crc kubenswrapper[4700]: I0227 17:14:55.751578 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tmx9\" (UniqueName: \"kubernetes.io/projected/e8d807a9-fc26-4417-94db-051e3e890be4-kube-api-access-7tmx9\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp\" (UID: \"e8d807a9-fc26-4417-94db-051e3e890be4\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp" Feb 27 17:14:55 crc kubenswrapper[4700]: I0227 17:14:55.751654 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e8d807a9-fc26-4417-94db-051e3e890be4-bundle\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp\" (UID: \"e8d807a9-fc26-4417-94db-051e3e890be4\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp" Feb 27 17:14:55 crc kubenswrapper[4700]: I0227 17:14:55.751727 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e8d807a9-fc26-4417-94db-051e3e890be4-util\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp\" (UID: \"e8d807a9-fc26-4417-94db-051e3e890be4\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp" Feb 27 17:14:55 crc kubenswrapper[4700]: I0227 17:14:55.852822 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tmx9\" (UniqueName: \"kubernetes.io/projected/e8d807a9-fc26-4417-94db-051e3e890be4-kube-api-access-7tmx9\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp\" (UID: \"e8d807a9-fc26-4417-94db-051e3e890be4\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp" Feb 27 17:14:55 crc kubenswrapper[4700]: I0227 17:14:55.852914 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e8d807a9-fc26-4417-94db-051e3e890be4-bundle\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp\" (UID: \"e8d807a9-fc26-4417-94db-051e3e890be4\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp" Feb 27 17:14:55 crc kubenswrapper[4700]: I0227 17:14:55.852981 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e8d807a9-fc26-4417-94db-051e3e890be4-util\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp\" (UID: \"e8d807a9-fc26-4417-94db-051e3e890be4\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp" Feb 27 17:14:55 crc kubenswrapper[4700]: I0227 17:14:55.853601 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e8d807a9-fc26-4417-94db-051e3e890be4-bundle\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp\" (UID: \"e8d807a9-fc26-4417-94db-051e3e890be4\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp" Feb 27 17:14:55 crc kubenswrapper[4700]: I0227 17:14:55.853846 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e8d807a9-fc26-4417-94db-051e3e890be4-util\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp\" (UID: \"e8d807a9-fc26-4417-94db-051e3e890be4\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp" Feb 27 17:14:55 crc kubenswrapper[4700]: I0227 17:14:55.895607 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tmx9\" (UniqueName: \"kubernetes.io/projected/e8d807a9-fc26-4417-94db-051e3e890be4-kube-api-access-7tmx9\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp\" (UID: \"e8d807a9-fc26-4417-94db-051e3e890be4\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp" Feb 27 17:14:55 crc kubenswrapper[4700]: I0227 17:14:55.938058 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp" Feb 27 17:14:56 crc kubenswrapper[4700]: I0227 17:14:56.222567 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp"] Feb 27 17:14:56 crc kubenswrapper[4700]: I0227 17:14:56.840540 4700 generic.go:334] "Generic (PLEG): container finished" podID="e8d807a9-fc26-4417-94db-051e3e890be4" containerID="3026a5bd6271a7b41fe7eb2569f366b25e3f6f8ed7ad8626f1e275b248744eba" exitCode=0 Feb 27 17:14:56 crc kubenswrapper[4700]: I0227 17:14:56.840599 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp" event={"ID":"e8d807a9-fc26-4417-94db-051e3e890be4","Type":"ContainerDied","Data":"3026a5bd6271a7b41fe7eb2569f366b25e3f6f8ed7ad8626f1e275b248744eba"} Feb 27 17:14:56 crc kubenswrapper[4700]: I0227 17:14:56.840637 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp" event={"ID":"e8d807a9-fc26-4417-94db-051e3e890be4","Type":"ContainerStarted","Data":"c54c4a82779b14910a7a1fd9075fe323ebc9cd7c8d23bbb2fada339098825ab8"} Feb 27 17:14:58 crc kubenswrapper[4700]: I0227 17:14:58.861607 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp" event={"ID":"e8d807a9-fc26-4417-94db-051e3e890be4","Type":"ContainerStarted","Data":"8f1c4456501968424cb4a0df5e670307c512049e8fb2a5a2901c16491d6124c2"} Feb 27 17:14:59 crc kubenswrapper[4700]: I0227 17:14:59.872693 4700 generic.go:334] "Generic (PLEG): container finished" podID="e8d807a9-fc26-4417-94db-051e3e890be4" containerID="8f1c4456501968424cb4a0df5e670307c512049e8fb2a5a2901c16491d6124c2" exitCode=0 Feb 27 17:14:59 crc kubenswrapper[4700]: I0227 17:14:59.872729 4700 generic.go:334] "Generic (PLEG): container finished" podID="e8d807a9-fc26-4417-94db-051e3e890be4" containerID="0817337dea9a1877b4ed427d7403900a209a1595b9e28fff08d1ca5d60b7a69d" exitCode=0 Feb 27 17:14:59 crc kubenswrapper[4700]: I0227 17:14:59.872754 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp" event={"ID":"e8d807a9-fc26-4417-94db-051e3e890be4","Type":"ContainerDied","Data":"8f1c4456501968424cb4a0df5e670307c512049e8fb2a5a2901c16491d6124c2"} Feb 27 17:14:59 crc kubenswrapper[4700]: I0227 17:14:59.872786 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp" event={"ID":"e8d807a9-fc26-4417-94db-051e3e890be4","Type":"ContainerDied","Data":"0817337dea9a1877b4ed427d7403900a209a1595b9e28fff08d1ca5d60b7a69d"} Feb 27 17:14:59 crc kubenswrapper[4700]: I0227 17:14:59.970443 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qclmh"] Feb 27 17:14:59 crc kubenswrapper[4700]: I0227 17:14:59.971797 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qclmh" Feb 27 17:14:59 crc kubenswrapper[4700]: I0227 17:14:59.997214 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qclmh"] Feb 27 17:15:00 crc kubenswrapper[4700]: I0227 17:15:00.118530 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w54d9\" (UniqueName: \"kubernetes.io/projected/0282bde1-f6fd-4d4f-a8d9-0342a9b1243d-kube-api-access-w54d9\") pod \"redhat-operators-qclmh\" (UID: \"0282bde1-f6fd-4d4f-a8d9-0342a9b1243d\") " pod="openshift-marketplace/redhat-operators-qclmh" Feb 27 17:15:00 crc kubenswrapper[4700]: I0227 17:15:00.118666 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0282bde1-f6fd-4d4f-a8d9-0342a9b1243d-utilities\") pod \"redhat-operators-qclmh\" (UID: \"0282bde1-f6fd-4d4f-a8d9-0342a9b1243d\") " pod="openshift-marketplace/redhat-operators-qclmh" Feb 27 17:15:00 crc kubenswrapper[4700]: I0227 17:15:00.118698 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0282bde1-f6fd-4d4f-a8d9-0342a9b1243d-catalog-content\") pod \"redhat-operators-qclmh\" (UID: \"0282bde1-f6fd-4d4f-a8d9-0342a9b1243d\") " pod="openshift-marketplace/redhat-operators-qclmh" Feb 27 17:15:00 crc kubenswrapper[4700]: I0227 17:15:00.145732 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536875-hzbmd"] Feb 27 17:15:00 crc kubenswrapper[4700]: I0227 17:15:00.147265 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536875-hzbmd" Feb 27 17:15:00 crc kubenswrapper[4700]: I0227 17:15:00.150900 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 27 17:15:00 crc kubenswrapper[4700]: I0227 17:15:00.150900 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 27 17:15:00 crc kubenswrapper[4700]: I0227 17:15:00.162986 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536875-hzbmd"] Feb 27 17:15:00 crc kubenswrapper[4700]: I0227 17:15:00.219611 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0282bde1-f6fd-4d4f-a8d9-0342a9b1243d-catalog-content\") pod \"redhat-operators-qclmh\" (UID: \"0282bde1-f6fd-4d4f-a8d9-0342a9b1243d\") " pod="openshift-marketplace/redhat-operators-qclmh" Feb 27 17:15:00 crc kubenswrapper[4700]: I0227 17:15:00.219675 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w54d9\" (UniqueName: \"kubernetes.io/projected/0282bde1-f6fd-4d4f-a8d9-0342a9b1243d-kube-api-access-w54d9\") pod \"redhat-operators-qclmh\" (UID: \"0282bde1-f6fd-4d4f-a8d9-0342a9b1243d\") " pod="openshift-marketplace/redhat-operators-qclmh" Feb 27 17:15:00 crc kubenswrapper[4700]: I0227 17:15:00.219738 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sw42t\" (UniqueName: \"kubernetes.io/projected/31b14208-847f-485b-90c9-fb6bb3332143-kube-api-access-sw42t\") pod \"collect-profiles-29536875-hzbmd\" (UID: \"31b14208-847f-485b-90c9-fb6bb3332143\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536875-hzbmd" Feb 27 17:15:00 crc kubenswrapper[4700]: I0227 17:15:00.219792 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/31b14208-847f-485b-90c9-fb6bb3332143-secret-volume\") pod \"collect-profiles-29536875-hzbmd\" (UID: \"31b14208-847f-485b-90c9-fb6bb3332143\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536875-hzbmd" Feb 27 17:15:00 crc kubenswrapper[4700]: I0227 17:15:00.219821 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/31b14208-847f-485b-90c9-fb6bb3332143-config-volume\") pod \"collect-profiles-29536875-hzbmd\" (UID: \"31b14208-847f-485b-90c9-fb6bb3332143\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536875-hzbmd" Feb 27 17:15:00 crc kubenswrapper[4700]: I0227 17:15:00.219868 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0282bde1-f6fd-4d4f-a8d9-0342a9b1243d-utilities\") pod \"redhat-operators-qclmh\" (UID: \"0282bde1-f6fd-4d4f-a8d9-0342a9b1243d\") " pod="openshift-marketplace/redhat-operators-qclmh" Feb 27 17:15:00 crc kubenswrapper[4700]: I0227 17:15:00.220249 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0282bde1-f6fd-4d4f-a8d9-0342a9b1243d-catalog-content\") pod \"redhat-operators-qclmh\" (UID: \"0282bde1-f6fd-4d4f-a8d9-0342a9b1243d\") " pod="openshift-marketplace/redhat-operators-qclmh" Feb 27 17:15:00 crc kubenswrapper[4700]: I0227 17:15:00.220352 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0282bde1-f6fd-4d4f-a8d9-0342a9b1243d-utilities\") pod \"redhat-operators-qclmh\" (UID: \"0282bde1-f6fd-4d4f-a8d9-0342a9b1243d\") " pod="openshift-marketplace/redhat-operators-qclmh" Feb 27 17:15:00 crc kubenswrapper[4700]: I0227 17:15:00.245692 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w54d9\" (UniqueName: \"kubernetes.io/projected/0282bde1-f6fd-4d4f-a8d9-0342a9b1243d-kube-api-access-w54d9\") pod \"redhat-operators-qclmh\" (UID: \"0282bde1-f6fd-4d4f-a8d9-0342a9b1243d\") " pod="openshift-marketplace/redhat-operators-qclmh" Feb 27 17:15:00 crc kubenswrapper[4700]: I0227 17:15:00.298776 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qclmh" Feb 27 17:15:00 crc kubenswrapper[4700]: I0227 17:15:00.321838 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sw42t\" (UniqueName: \"kubernetes.io/projected/31b14208-847f-485b-90c9-fb6bb3332143-kube-api-access-sw42t\") pod \"collect-profiles-29536875-hzbmd\" (UID: \"31b14208-847f-485b-90c9-fb6bb3332143\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536875-hzbmd" Feb 27 17:15:00 crc kubenswrapper[4700]: I0227 17:15:00.321938 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/31b14208-847f-485b-90c9-fb6bb3332143-secret-volume\") pod \"collect-profiles-29536875-hzbmd\" (UID: \"31b14208-847f-485b-90c9-fb6bb3332143\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536875-hzbmd" Feb 27 17:15:00 crc kubenswrapper[4700]: I0227 17:15:00.321980 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/31b14208-847f-485b-90c9-fb6bb3332143-config-volume\") pod \"collect-profiles-29536875-hzbmd\" (UID: \"31b14208-847f-485b-90c9-fb6bb3332143\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536875-hzbmd" Feb 27 17:15:00 crc kubenswrapper[4700]: I0227 17:15:00.323120 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/31b14208-847f-485b-90c9-fb6bb3332143-config-volume\") pod \"collect-profiles-29536875-hzbmd\" (UID: \"31b14208-847f-485b-90c9-fb6bb3332143\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536875-hzbmd" Feb 27 17:15:00 crc kubenswrapper[4700]: I0227 17:15:00.327708 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/31b14208-847f-485b-90c9-fb6bb3332143-secret-volume\") pod \"collect-profiles-29536875-hzbmd\" (UID: \"31b14208-847f-485b-90c9-fb6bb3332143\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536875-hzbmd" Feb 27 17:15:00 crc kubenswrapper[4700]: I0227 17:15:00.339841 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sw42t\" (UniqueName: \"kubernetes.io/projected/31b14208-847f-485b-90c9-fb6bb3332143-kube-api-access-sw42t\") pod \"collect-profiles-29536875-hzbmd\" (UID: \"31b14208-847f-485b-90c9-fb6bb3332143\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536875-hzbmd" Feb 27 17:15:00 crc kubenswrapper[4700]: I0227 17:15:00.472977 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536875-hzbmd" Feb 27 17:15:00 crc kubenswrapper[4700]: I0227 17:15:00.544536 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qclmh"] Feb 27 17:15:00 crc kubenswrapper[4700]: W0227 17:15:00.547610 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0282bde1_f6fd_4d4f_a8d9_0342a9b1243d.slice/crio-2949f6687fa0d99f6253fb02b0e53948f3e793f52200d8cdf23e5e9b9861bc4a WatchSource:0}: Error finding container 2949f6687fa0d99f6253fb02b0e53948f3e793f52200d8cdf23e5e9b9861bc4a: Status 404 returned error can't find the container with id 2949f6687fa0d99f6253fb02b0e53948f3e793f52200d8cdf23e5e9b9861bc4a Feb 27 17:15:00 crc kubenswrapper[4700]: I0227 17:15:00.708945 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536875-hzbmd"] Feb 27 17:15:00 crc kubenswrapper[4700]: W0227 17:15:00.768816 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod31b14208_847f_485b_90c9_fb6bb3332143.slice/crio-62747eb2f7c9a67192be1c23fa60d0d02be834a0ffe533f8ef341c83cd7e8cc2 WatchSource:0}: Error finding container 62747eb2f7c9a67192be1c23fa60d0d02be834a0ffe533f8ef341c83cd7e8cc2: Status 404 returned error can't find the container with id 62747eb2f7c9a67192be1c23fa60d0d02be834a0ffe533f8ef341c83cd7e8cc2 Feb 27 17:15:00 crc kubenswrapper[4700]: I0227 17:15:00.893584 4700 generic.go:334] "Generic (PLEG): container finished" podID="0282bde1-f6fd-4d4f-a8d9-0342a9b1243d" containerID="1801e3857b90bc90aa3d362b761d292b8f7f4d4c1ffad8d1c05a7348b734c0b6" exitCode=0 Feb 27 17:15:00 crc kubenswrapper[4700]: I0227 17:15:00.893676 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qclmh" event={"ID":"0282bde1-f6fd-4d4f-a8d9-0342a9b1243d","Type":"ContainerDied","Data":"1801e3857b90bc90aa3d362b761d292b8f7f4d4c1ffad8d1c05a7348b734c0b6"} Feb 27 17:15:00 crc kubenswrapper[4700]: I0227 17:15:00.893735 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qclmh" event={"ID":"0282bde1-f6fd-4d4f-a8d9-0342a9b1243d","Type":"ContainerStarted","Data":"2949f6687fa0d99f6253fb02b0e53948f3e793f52200d8cdf23e5e9b9861bc4a"} Feb 27 17:15:00 crc kubenswrapper[4700]: I0227 17:15:00.895456 4700 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 17:15:00 crc kubenswrapper[4700]: I0227 17:15:00.896161 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536875-hzbmd" event={"ID":"31b14208-847f-485b-90c9-fb6bb3332143","Type":"ContainerStarted","Data":"62747eb2f7c9a67192be1c23fa60d0d02be834a0ffe533f8ef341c83cd7e8cc2"} Feb 27 17:15:00 crc kubenswrapper[4700]: I0227 17:15:00.933780 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29536875-hzbmd" podStartSLOduration=0.933757498 podStartE2EDuration="933.757498ms" podCreationTimestamp="2026-02-27 17:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:15:00.928812666 +0000 UTC m=+860.914125423" watchObservedRunningTime="2026-02-27 17:15:00.933757498 +0000 UTC m=+860.919070265" Feb 27 17:15:01 crc kubenswrapper[4700]: I0227 17:15:01.107452 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp" Feb 27 17:15:01 crc kubenswrapper[4700]: I0227 17:15:01.235583 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e8d807a9-fc26-4417-94db-051e3e890be4-util\") pod \"e8d807a9-fc26-4417-94db-051e3e890be4\" (UID: \"e8d807a9-fc26-4417-94db-051e3e890be4\") " Feb 27 17:15:01 crc kubenswrapper[4700]: I0227 17:15:01.235641 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e8d807a9-fc26-4417-94db-051e3e890be4-bundle\") pod \"e8d807a9-fc26-4417-94db-051e3e890be4\" (UID: \"e8d807a9-fc26-4417-94db-051e3e890be4\") " Feb 27 17:15:01 crc kubenswrapper[4700]: I0227 17:15:01.235717 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7tmx9\" (UniqueName: \"kubernetes.io/projected/e8d807a9-fc26-4417-94db-051e3e890be4-kube-api-access-7tmx9\") pod \"e8d807a9-fc26-4417-94db-051e3e890be4\" (UID: \"e8d807a9-fc26-4417-94db-051e3e890be4\") " Feb 27 17:15:01 crc kubenswrapper[4700]: I0227 17:15:01.236242 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8d807a9-fc26-4417-94db-051e3e890be4-bundle" (OuterVolumeSpecName: "bundle") pod "e8d807a9-fc26-4417-94db-051e3e890be4" (UID: "e8d807a9-fc26-4417-94db-051e3e890be4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:15:01 crc kubenswrapper[4700]: I0227 17:15:01.244654 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8d807a9-fc26-4417-94db-051e3e890be4-kube-api-access-7tmx9" (OuterVolumeSpecName: "kube-api-access-7tmx9") pod "e8d807a9-fc26-4417-94db-051e3e890be4" (UID: "e8d807a9-fc26-4417-94db-051e3e890be4"). InnerVolumeSpecName "kube-api-access-7tmx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:15:01 crc kubenswrapper[4700]: I0227 17:15:01.248810 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8d807a9-fc26-4417-94db-051e3e890be4-util" (OuterVolumeSpecName: "util") pod "e8d807a9-fc26-4417-94db-051e3e890be4" (UID: "e8d807a9-fc26-4417-94db-051e3e890be4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:15:01 crc kubenswrapper[4700]: I0227 17:15:01.337151 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7tmx9\" (UniqueName: \"kubernetes.io/projected/e8d807a9-fc26-4417-94db-051e3e890be4-kube-api-access-7tmx9\") on node \"crc\" DevicePath \"\"" Feb 27 17:15:01 crc kubenswrapper[4700]: I0227 17:15:01.337563 4700 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e8d807a9-fc26-4417-94db-051e3e890be4-util\") on node \"crc\" DevicePath \"\"" Feb 27 17:15:01 crc kubenswrapper[4700]: I0227 17:15:01.337581 4700 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e8d807a9-fc26-4417-94db-051e3e890be4-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:15:01 crc kubenswrapper[4700]: I0227 17:15:01.903940 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp" event={"ID":"e8d807a9-fc26-4417-94db-051e3e890be4","Type":"ContainerDied","Data":"c54c4a82779b14910a7a1fd9075fe323ebc9cd7c8d23bbb2fada339098825ab8"} Feb 27 17:15:01 crc kubenswrapper[4700]: I0227 17:15:01.903979 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c54c4a82779b14910a7a1fd9075fe323ebc9cd7c8d23bbb2fada339098825ab8" Feb 27 17:15:01 crc kubenswrapper[4700]: I0227 17:15:01.903996 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp" Feb 27 17:15:01 crc kubenswrapper[4700]: I0227 17:15:01.908409 4700 generic.go:334] "Generic (PLEG): container finished" podID="31b14208-847f-485b-90c9-fb6bb3332143" containerID="219100e53b01503afacb5172c0c80e3e9b408ba296258bd707778732d4f9a196" exitCode=0 Feb 27 17:15:01 crc kubenswrapper[4700]: I0227 17:15:01.908440 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536875-hzbmd" event={"ID":"31b14208-847f-485b-90c9-fb6bb3332143","Type":"ContainerDied","Data":"219100e53b01503afacb5172c0c80e3e9b408ba296258bd707778732d4f9a196"} Feb 27 17:15:02 crc kubenswrapper[4700]: I0227 17:15:02.921154 4700 generic.go:334] "Generic (PLEG): container finished" podID="0282bde1-f6fd-4d4f-a8d9-0342a9b1243d" containerID="f7f3dcec8e488f21c91463bc5f997d65dbc134fb2145474cd6926537b5db0e34" exitCode=0 Feb 27 17:15:02 crc kubenswrapper[4700]: I0227 17:15:02.921213 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qclmh" event={"ID":"0282bde1-f6fd-4d4f-a8d9-0342a9b1243d","Type":"ContainerDied","Data":"f7f3dcec8e488f21c91463bc5f997d65dbc134fb2145474cd6926537b5db0e34"} Feb 27 17:15:03 crc kubenswrapper[4700]: I0227 17:15:03.244516 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536875-hzbmd" Feb 27 17:15:03 crc kubenswrapper[4700]: I0227 17:15:03.364425 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/31b14208-847f-485b-90c9-fb6bb3332143-config-volume\") pod \"31b14208-847f-485b-90c9-fb6bb3332143\" (UID: \"31b14208-847f-485b-90c9-fb6bb3332143\") " Feb 27 17:15:03 crc kubenswrapper[4700]: I0227 17:15:03.364553 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sw42t\" (UniqueName: \"kubernetes.io/projected/31b14208-847f-485b-90c9-fb6bb3332143-kube-api-access-sw42t\") pod \"31b14208-847f-485b-90c9-fb6bb3332143\" (UID: \"31b14208-847f-485b-90c9-fb6bb3332143\") " Feb 27 17:15:03 crc kubenswrapper[4700]: I0227 17:15:03.364595 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/31b14208-847f-485b-90c9-fb6bb3332143-secret-volume\") pod \"31b14208-847f-485b-90c9-fb6bb3332143\" (UID: \"31b14208-847f-485b-90c9-fb6bb3332143\") " Feb 27 17:15:03 crc kubenswrapper[4700]: I0227 17:15:03.366721 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31b14208-847f-485b-90c9-fb6bb3332143-config-volume" (OuterVolumeSpecName: "config-volume") pod "31b14208-847f-485b-90c9-fb6bb3332143" (UID: "31b14208-847f-485b-90c9-fb6bb3332143"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:15:03 crc kubenswrapper[4700]: I0227 17:15:03.373339 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31b14208-847f-485b-90c9-fb6bb3332143-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "31b14208-847f-485b-90c9-fb6bb3332143" (UID: "31b14208-847f-485b-90c9-fb6bb3332143"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:15:03 crc kubenswrapper[4700]: I0227 17:15:03.375253 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31b14208-847f-485b-90c9-fb6bb3332143-kube-api-access-sw42t" (OuterVolumeSpecName: "kube-api-access-sw42t") pod "31b14208-847f-485b-90c9-fb6bb3332143" (UID: "31b14208-847f-485b-90c9-fb6bb3332143"). InnerVolumeSpecName "kube-api-access-sw42t". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:15:03 crc kubenswrapper[4700]: I0227 17:15:03.465959 4700 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/31b14208-847f-485b-90c9-fb6bb3332143-config-volume\") on node \"crc\" DevicePath \"\"" Feb 27 17:15:03 crc kubenswrapper[4700]: I0227 17:15:03.466012 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sw42t\" (UniqueName: \"kubernetes.io/projected/31b14208-847f-485b-90c9-fb6bb3332143-kube-api-access-sw42t\") on node \"crc\" DevicePath \"\"" Feb 27 17:15:03 crc kubenswrapper[4700]: I0227 17:15:03.466036 4700 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/31b14208-847f-485b-90c9-fb6bb3332143-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 27 17:15:03 crc kubenswrapper[4700]: I0227 17:15:03.933076 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qclmh" event={"ID":"0282bde1-f6fd-4d4f-a8d9-0342a9b1243d","Type":"ContainerStarted","Data":"b228710907b61dce4cce3ae5711b13c5baa71797250746f541e0fa3d68b84aeb"} Feb 27 17:15:03 crc kubenswrapper[4700]: I0227 17:15:03.936847 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536875-hzbmd" event={"ID":"31b14208-847f-485b-90c9-fb6bb3332143","Type":"ContainerDied","Data":"62747eb2f7c9a67192be1c23fa60d0d02be834a0ffe533f8ef341c83cd7e8cc2"} Feb 27 17:15:03 crc kubenswrapper[4700]: I0227 17:15:03.936896 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="62747eb2f7c9a67192be1c23fa60d0d02be834a0ffe533f8ef341c83cd7e8cc2" Feb 27 17:15:03 crc kubenswrapper[4700]: I0227 17:15:03.937221 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536875-hzbmd" Feb 27 17:15:03 crc kubenswrapper[4700]: I0227 17:15:03.967563 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qclmh" podStartSLOduration=2.517220238 podStartE2EDuration="4.967534348s" podCreationTimestamp="2026-02-27 17:14:59 +0000 UTC" firstStartedPulling="2026-02-27 17:15:00.895053728 +0000 UTC m=+860.880366495" lastFinishedPulling="2026-02-27 17:15:03.345367828 +0000 UTC m=+863.330680605" observedRunningTime="2026-02-27 17:15:03.965077412 +0000 UTC m=+863.950390199" watchObservedRunningTime="2026-02-27 17:15:03.967534348 +0000 UTC m=+863.952847125" Feb 27 17:15:06 crc kubenswrapper[4700]: I0227 17:15:06.410018 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:15:06 crc kubenswrapper[4700]: I0227 17:15:06.410079 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:15:06 crc kubenswrapper[4700]: I0227 17:15:06.410125 4700 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" Feb 27 17:15:06 crc kubenswrapper[4700]: I0227 17:15:06.410614 4700 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"510d6e176ab30bcaf35acfa6f85490773219083ba09de332e6a13153619b2fac"} pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 17:15:06 crc kubenswrapper[4700]: I0227 17:15:06.410667 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" containerID="cri-o://510d6e176ab30bcaf35acfa6f85490773219083ba09de332e6a13153619b2fac" gracePeriod=600 Feb 27 17:15:06 crc kubenswrapper[4700]: I0227 17:15:06.510759 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-75c5dccd6c-6rrfb"] Feb 27 17:15:06 crc kubenswrapper[4700]: E0227 17:15:06.511011 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8d807a9-fc26-4417-94db-051e3e890be4" containerName="pull" Feb 27 17:15:06 crc kubenswrapper[4700]: I0227 17:15:06.511032 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8d807a9-fc26-4417-94db-051e3e890be4" containerName="pull" Feb 27 17:15:06 crc kubenswrapper[4700]: E0227 17:15:06.511046 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8d807a9-fc26-4417-94db-051e3e890be4" containerName="extract" Feb 27 17:15:06 crc kubenswrapper[4700]: I0227 17:15:06.511054 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8d807a9-fc26-4417-94db-051e3e890be4" containerName="extract" Feb 27 17:15:06 crc kubenswrapper[4700]: E0227 17:15:06.511067 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8d807a9-fc26-4417-94db-051e3e890be4" containerName="util" Feb 27 17:15:06 crc kubenswrapper[4700]: I0227 17:15:06.511076 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8d807a9-fc26-4417-94db-051e3e890be4" containerName="util" Feb 27 17:15:06 crc kubenswrapper[4700]: E0227 17:15:06.511091 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31b14208-847f-485b-90c9-fb6bb3332143" containerName="collect-profiles" Feb 27 17:15:06 crc kubenswrapper[4700]: I0227 17:15:06.511099 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="31b14208-847f-485b-90c9-fb6bb3332143" containerName="collect-profiles" Feb 27 17:15:06 crc kubenswrapper[4700]: I0227 17:15:06.511223 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="31b14208-847f-485b-90c9-fb6bb3332143" containerName="collect-profiles" Feb 27 17:15:06 crc kubenswrapper[4700]: I0227 17:15:06.511238 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8d807a9-fc26-4417-94db-051e3e890be4" containerName="extract" Feb 27 17:15:06 crc kubenswrapper[4700]: I0227 17:15:06.511759 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-75c5dccd6c-6rrfb" Feb 27 17:15:06 crc kubenswrapper[4700]: I0227 17:15:06.514619 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-zgp2m" Feb 27 17:15:06 crc kubenswrapper[4700]: I0227 17:15:06.514902 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Feb 27 17:15:06 crc kubenswrapper[4700]: I0227 17:15:06.515048 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Feb 27 17:15:06 crc kubenswrapper[4700]: I0227 17:15:06.540988 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-75c5dccd6c-6rrfb"] Feb 27 17:15:06 crc kubenswrapper[4700]: I0227 17:15:06.604075 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f29xf\" (UniqueName: \"kubernetes.io/projected/3dc864b3-a8d5-4587-9356-ecffe805f637-kube-api-access-f29xf\") pod \"nmstate-operator-75c5dccd6c-6rrfb\" (UID: \"3dc864b3-a8d5-4587-9356-ecffe805f637\") " pod="openshift-nmstate/nmstate-operator-75c5dccd6c-6rrfb" Feb 27 17:15:06 crc kubenswrapper[4700]: I0227 17:15:06.705791 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f29xf\" (UniqueName: \"kubernetes.io/projected/3dc864b3-a8d5-4587-9356-ecffe805f637-kube-api-access-f29xf\") pod \"nmstate-operator-75c5dccd6c-6rrfb\" (UID: \"3dc864b3-a8d5-4587-9356-ecffe805f637\") " pod="openshift-nmstate/nmstate-operator-75c5dccd6c-6rrfb" Feb 27 17:15:06 crc kubenswrapper[4700]: I0227 17:15:06.734678 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f29xf\" (UniqueName: \"kubernetes.io/projected/3dc864b3-a8d5-4587-9356-ecffe805f637-kube-api-access-f29xf\") pod \"nmstate-operator-75c5dccd6c-6rrfb\" (UID: \"3dc864b3-a8d5-4587-9356-ecffe805f637\") " pod="openshift-nmstate/nmstate-operator-75c5dccd6c-6rrfb" Feb 27 17:15:06 crc kubenswrapper[4700]: I0227 17:15:06.875280 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-75c5dccd6c-6rrfb" Feb 27 17:15:06 crc kubenswrapper[4700]: I0227 17:15:06.959002 4700 generic.go:334] "Generic (PLEG): container finished" podID="84b45511-b94c-479f-98db-fd2c4eceec46" containerID="510d6e176ab30bcaf35acfa6f85490773219083ba09de332e6a13153619b2fac" exitCode=0 Feb 27 17:15:06 crc kubenswrapper[4700]: I0227 17:15:06.959063 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerDied","Data":"510d6e176ab30bcaf35acfa6f85490773219083ba09de332e6a13153619b2fac"} Feb 27 17:15:06 crc kubenswrapper[4700]: I0227 17:15:06.959301 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerStarted","Data":"6c88e3b04f19e6791ce8b218282ad3546f21efb28cfed9c5676d0da60d8b308b"} Feb 27 17:15:06 crc kubenswrapper[4700]: I0227 17:15:06.959332 4700 scope.go:117] "RemoveContainer" containerID="e1a1a6356e2245cab05c4d1addb6b5e53e3e02d22a816739c9a355a221dd2d7b" Feb 27 17:15:07 crc kubenswrapper[4700]: W0227 17:15:07.108458 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dc864b3_a8d5_4587_9356_ecffe805f637.slice/crio-ea70a8e2649cc4fa02a5199f967105d19bf2e6447b2937e0d2803261cf417402 WatchSource:0}: Error finding container ea70a8e2649cc4fa02a5199f967105d19bf2e6447b2937e0d2803261cf417402: Status 404 returned error can't find the container with id ea70a8e2649cc4fa02a5199f967105d19bf2e6447b2937e0d2803261cf417402 Feb 27 17:15:07 crc kubenswrapper[4700]: I0227 17:15:07.108626 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-75c5dccd6c-6rrfb"] Feb 27 17:15:07 crc kubenswrapper[4700]: I0227 17:15:07.982326 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-75c5dccd6c-6rrfb" event={"ID":"3dc864b3-a8d5-4587-9356-ecffe805f637","Type":"ContainerStarted","Data":"ea70a8e2649cc4fa02a5199f967105d19bf2e6447b2937e0d2803261cf417402"} Feb 27 17:15:10 crc kubenswrapper[4700]: I0227 17:15:10.299958 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qclmh" Feb 27 17:15:10 crc kubenswrapper[4700]: I0227 17:15:10.302185 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qclmh" Feb 27 17:15:11 crc kubenswrapper[4700]: I0227 17:15:11.012687 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-75c5dccd6c-6rrfb" event={"ID":"3dc864b3-a8d5-4587-9356-ecffe805f637","Type":"ContainerStarted","Data":"9a4b1d8d702afda06db69c6d586a87f23d38845ce9c1b0906a8a7507d0bbe58a"} Feb 27 17:15:11 crc kubenswrapper[4700]: I0227 17:15:11.039032 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-75c5dccd6c-6rrfb" podStartSLOduration=1.681290865 podStartE2EDuration="5.038999878s" podCreationTimestamp="2026-02-27 17:15:06 +0000 UTC" firstStartedPulling="2026-02-27 17:15:07.112286571 +0000 UTC m=+867.097599318" lastFinishedPulling="2026-02-27 17:15:10.469995584 +0000 UTC m=+870.455308331" observedRunningTime="2026-02-27 17:15:11.034406446 +0000 UTC m=+871.019719233" watchObservedRunningTime="2026-02-27 17:15:11.038999878 +0000 UTC m=+871.024312655" Feb 27 17:15:11 crc kubenswrapper[4700]: I0227 17:15:11.353563 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qclmh" podUID="0282bde1-f6fd-4d4f-a8d9-0342a9b1243d" containerName="registry-server" probeResult="failure" output=< Feb 27 17:15:11 crc kubenswrapper[4700]: timeout: failed to connect service ":50051" within 1s Feb 27 17:15:11 crc kubenswrapper[4700]: > Feb 27 17:15:16 crc kubenswrapper[4700]: I0227 17:15:16.766756 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-69594cc75-znbzd"] Feb 27 17:15:16 crc kubenswrapper[4700]: I0227 17:15:16.768931 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-69594cc75-znbzd" Feb 27 17:15:16 crc kubenswrapper[4700]: I0227 17:15:16.773226 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-fw95v" Feb 27 17:15:16 crc kubenswrapper[4700]: I0227 17:15:16.780705 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-786f45cff4-gg78f"] Feb 27 17:15:16 crc kubenswrapper[4700]: I0227 17:15:16.782870 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-786f45cff4-gg78f" Feb 27 17:15:16 crc kubenswrapper[4700]: I0227 17:15:16.785172 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Feb 27 17:15:16 crc kubenswrapper[4700]: I0227 17:15:16.802936 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-69594cc75-znbzd"] Feb 27 17:15:16 crc kubenswrapper[4700]: I0227 17:15:16.814657 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-786f45cff4-gg78f"] Feb 27 17:15:16 crc kubenswrapper[4700]: I0227 17:15:16.819391 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-85tkk"] Feb 27 17:15:16 crc kubenswrapper[4700]: I0227 17:15:16.820297 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-85tkk" Feb 27 17:15:16 crc kubenswrapper[4700]: I0227 17:15:16.836917 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/97a282f2-3d9c-4911-bb64-4593b586bb4e-tls-key-pair\") pod \"nmstate-webhook-786f45cff4-gg78f\" (UID: \"97a282f2-3d9c-4911-bb64-4593b586bb4e\") " pod="openshift-nmstate/nmstate-webhook-786f45cff4-gg78f" Feb 27 17:15:16 crc kubenswrapper[4700]: I0227 17:15:16.836953 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4s9j\" (UniqueName: \"kubernetes.io/projected/47c2c9bd-d6c8-4b33-974c-ab0e57ea05fc-kube-api-access-r4s9j\") pod \"nmstate-metrics-69594cc75-znbzd\" (UID: \"47c2c9bd-d6c8-4b33-974c-ab0e57ea05fc\") " pod="openshift-nmstate/nmstate-metrics-69594cc75-znbzd" Feb 27 17:15:16 crc kubenswrapper[4700]: I0227 17:15:16.836986 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ww4q\" (UniqueName: \"kubernetes.io/projected/97a282f2-3d9c-4911-bb64-4593b586bb4e-kube-api-access-2ww4q\") pod \"nmstate-webhook-786f45cff4-gg78f\" (UID: \"97a282f2-3d9c-4911-bb64-4593b586bb4e\") " pod="openshift-nmstate/nmstate-webhook-786f45cff4-gg78f" Feb 27 17:15:16 crc kubenswrapper[4700]: I0227 17:15:16.893402 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-m9hcq"] Feb 27 17:15:16 crc kubenswrapper[4700]: I0227 17:15:16.894355 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-m9hcq" Feb 27 17:15:16 crc kubenswrapper[4700]: I0227 17:15:16.896559 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Feb 27 17:15:16 crc kubenswrapper[4700]: I0227 17:15:16.896772 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Feb 27 17:15:16 crc kubenswrapper[4700]: I0227 17:15:16.901283 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-m9hcq"] Feb 27 17:15:16 crc kubenswrapper[4700]: I0227 17:15:16.904551 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-scrgg" Feb 27 17:15:16 crc kubenswrapper[4700]: I0227 17:15:16.938368 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/97a282f2-3d9c-4911-bb64-4593b586bb4e-tls-key-pair\") pod \"nmstate-webhook-786f45cff4-gg78f\" (UID: \"97a282f2-3d9c-4911-bb64-4593b586bb4e\") " pod="openshift-nmstate/nmstate-webhook-786f45cff4-gg78f" Feb 27 17:15:16 crc kubenswrapper[4700]: I0227 17:15:16.938410 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4s9j\" (UniqueName: \"kubernetes.io/projected/47c2c9bd-d6c8-4b33-974c-ab0e57ea05fc-kube-api-access-r4s9j\") pod \"nmstate-metrics-69594cc75-znbzd\" (UID: \"47c2c9bd-d6c8-4b33-974c-ab0e57ea05fc\") " pod="openshift-nmstate/nmstate-metrics-69594cc75-znbzd" Feb 27 17:15:16 crc kubenswrapper[4700]: I0227 17:15:16.938437 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/0686b736-3cb6-4fb9-8710-8359e2066637-nmstate-lock\") pod \"nmstate-handler-85tkk\" (UID: \"0686b736-3cb6-4fb9-8710-8359e2066637\") " pod="openshift-nmstate/nmstate-handler-85tkk" Feb 27 17:15:16 crc kubenswrapper[4700]: I0227 17:15:16.938500 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/0686b736-3cb6-4fb9-8710-8359e2066637-ovs-socket\") pod \"nmstate-handler-85tkk\" (UID: \"0686b736-3cb6-4fb9-8710-8359e2066637\") " pod="openshift-nmstate/nmstate-handler-85tkk" Feb 27 17:15:16 crc kubenswrapper[4700]: I0227 17:15:16.938521 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ww4q\" (UniqueName: \"kubernetes.io/projected/97a282f2-3d9c-4911-bb64-4593b586bb4e-kube-api-access-2ww4q\") pod \"nmstate-webhook-786f45cff4-gg78f\" (UID: \"97a282f2-3d9c-4911-bb64-4593b586bb4e\") " pod="openshift-nmstate/nmstate-webhook-786f45cff4-gg78f" Feb 27 17:15:16 crc kubenswrapper[4700]: E0227 17:15:16.938595 4700 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Feb 27 17:15:16 crc kubenswrapper[4700]: E0227 17:15:16.938686 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/97a282f2-3d9c-4911-bb64-4593b586bb4e-tls-key-pair podName:97a282f2-3d9c-4911-bb64-4593b586bb4e nodeName:}" failed. No retries permitted until 2026-02-27 17:15:17.438663651 +0000 UTC m=+877.423976498 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/97a282f2-3d9c-4911-bb64-4593b586bb4e-tls-key-pair") pod "nmstate-webhook-786f45cff4-gg78f" (UID: "97a282f2-3d9c-4911-bb64-4593b586bb4e") : secret "openshift-nmstate-webhook" not found Feb 27 17:15:16 crc kubenswrapper[4700]: I0227 17:15:16.938746 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbf5t\" (UniqueName: \"kubernetes.io/projected/0686b736-3cb6-4fb9-8710-8359e2066637-kube-api-access-sbf5t\") pod \"nmstate-handler-85tkk\" (UID: \"0686b736-3cb6-4fb9-8710-8359e2066637\") " pod="openshift-nmstate/nmstate-handler-85tkk" Feb 27 17:15:16 crc kubenswrapper[4700]: I0227 17:15:16.938808 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/0686b736-3cb6-4fb9-8710-8359e2066637-dbus-socket\") pod \"nmstate-handler-85tkk\" (UID: \"0686b736-3cb6-4fb9-8710-8359e2066637\") " pod="openshift-nmstate/nmstate-handler-85tkk" Feb 27 17:15:16 crc kubenswrapper[4700]: I0227 17:15:16.959347 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ww4q\" (UniqueName: \"kubernetes.io/projected/97a282f2-3d9c-4911-bb64-4593b586bb4e-kube-api-access-2ww4q\") pod \"nmstate-webhook-786f45cff4-gg78f\" (UID: \"97a282f2-3d9c-4911-bb64-4593b586bb4e\") " pod="openshift-nmstate/nmstate-webhook-786f45cff4-gg78f" Feb 27 17:15:16 crc kubenswrapper[4700]: I0227 17:15:16.959951 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4s9j\" (UniqueName: \"kubernetes.io/projected/47c2c9bd-d6c8-4b33-974c-ab0e57ea05fc-kube-api-access-r4s9j\") pod \"nmstate-metrics-69594cc75-znbzd\" (UID: \"47c2c9bd-d6c8-4b33-974c-ab0e57ea05fc\") " pod="openshift-nmstate/nmstate-metrics-69594cc75-znbzd" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.039728 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbf5t\" (UniqueName: \"kubernetes.io/projected/0686b736-3cb6-4fb9-8710-8359e2066637-kube-api-access-sbf5t\") pod \"nmstate-handler-85tkk\" (UID: \"0686b736-3cb6-4fb9-8710-8359e2066637\") " pod="openshift-nmstate/nmstate-handler-85tkk" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.039772 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/0686b736-3cb6-4fb9-8710-8359e2066637-dbus-socket\") pod \"nmstate-handler-85tkk\" (UID: \"0686b736-3cb6-4fb9-8710-8359e2066637\") " pod="openshift-nmstate/nmstate-handler-85tkk" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.039818 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9njx5\" (UniqueName: \"kubernetes.io/projected/fbb064be-36df-4a47-8c0e-9ffdd43d1123-kube-api-access-9njx5\") pod \"nmstate-console-plugin-5dcbbd79cf-m9hcq\" (UID: \"fbb064be-36df-4a47-8c0e-9ffdd43d1123\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-m9hcq" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.039842 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/0686b736-3cb6-4fb9-8710-8359e2066637-nmstate-lock\") pod \"nmstate-handler-85tkk\" (UID: \"0686b736-3cb6-4fb9-8710-8359e2066637\") " pod="openshift-nmstate/nmstate-handler-85tkk" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.039872 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/0686b736-3cb6-4fb9-8710-8359e2066637-ovs-socket\") pod \"nmstate-handler-85tkk\" (UID: \"0686b736-3cb6-4fb9-8710-8359e2066637\") " pod="openshift-nmstate/nmstate-handler-85tkk" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.039897 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/fbb064be-36df-4a47-8c0e-9ffdd43d1123-nginx-conf\") pod \"nmstate-console-plugin-5dcbbd79cf-m9hcq\" (UID: \"fbb064be-36df-4a47-8c0e-9ffdd43d1123\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-m9hcq" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.039935 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/fbb064be-36df-4a47-8c0e-9ffdd43d1123-plugin-serving-cert\") pod \"nmstate-console-plugin-5dcbbd79cf-m9hcq\" (UID: \"fbb064be-36df-4a47-8c0e-9ffdd43d1123\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-m9hcq" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.039972 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/0686b736-3cb6-4fb9-8710-8359e2066637-nmstate-lock\") pod \"nmstate-handler-85tkk\" (UID: \"0686b736-3cb6-4fb9-8710-8359e2066637\") " pod="openshift-nmstate/nmstate-handler-85tkk" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.040027 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/0686b736-3cb6-4fb9-8710-8359e2066637-ovs-socket\") pod \"nmstate-handler-85tkk\" (UID: \"0686b736-3cb6-4fb9-8710-8359e2066637\") " pod="openshift-nmstate/nmstate-handler-85tkk" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.040153 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/0686b736-3cb6-4fb9-8710-8359e2066637-dbus-socket\") pod \"nmstate-handler-85tkk\" (UID: \"0686b736-3cb6-4fb9-8710-8359e2066637\") " pod="openshift-nmstate/nmstate-handler-85tkk" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.055725 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbf5t\" (UniqueName: \"kubernetes.io/projected/0686b736-3cb6-4fb9-8710-8359e2066637-kube-api-access-sbf5t\") pod \"nmstate-handler-85tkk\" (UID: \"0686b736-3cb6-4fb9-8710-8359e2066637\") " pod="openshift-nmstate/nmstate-handler-85tkk" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.088441 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-69594cc75-znbzd" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.094099 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-54477f4d4-6cvcc"] Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.094924 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-54477f4d4-6cvcc" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.115223 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-54477f4d4-6cvcc"] Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.136769 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-85tkk" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.141748 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/fbb064be-36df-4a47-8c0e-9ffdd43d1123-nginx-conf\") pod \"nmstate-console-plugin-5dcbbd79cf-m9hcq\" (UID: \"fbb064be-36df-4a47-8c0e-9ffdd43d1123\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-m9hcq" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.141824 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/fbb064be-36df-4a47-8c0e-9ffdd43d1123-plugin-serving-cert\") pod \"nmstate-console-plugin-5dcbbd79cf-m9hcq\" (UID: \"fbb064be-36df-4a47-8c0e-9ffdd43d1123\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-m9hcq" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.141891 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9njx5\" (UniqueName: \"kubernetes.io/projected/fbb064be-36df-4a47-8c0e-9ffdd43d1123-kube-api-access-9njx5\") pod \"nmstate-console-plugin-5dcbbd79cf-m9hcq\" (UID: \"fbb064be-36df-4a47-8c0e-9ffdd43d1123\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-m9hcq" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.143164 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/fbb064be-36df-4a47-8c0e-9ffdd43d1123-nginx-conf\") pod \"nmstate-console-plugin-5dcbbd79cf-m9hcq\" (UID: \"fbb064be-36df-4a47-8c0e-9ffdd43d1123\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-m9hcq" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.145937 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/fbb064be-36df-4a47-8c0e-9ffdd43d1123-plugin-serving-cert\") pod \"nmstate-console-plugin-5dcbbd79cf-m9hcq\" (UID: \"fbb064be-36df-4a47-8c0e-9ffdd43d1123\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-m9hcq" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.160796 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9njx5\" (UniqueName: \"kubernetes.io/projected/fbb064be-36df-4a47-8c0e-9ffdd43d1123-kube-api-access-9njx5\") pod \"nmstate-console-plugin-5dcbbd79cf-m9hcq\" (UID: \"fbb064be-36df-4a47-8c0e-9ffdd43d1123\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-m9hcq" Feb 27 17:15:17 crc kubenswrapper[4700]: W0227 17:15:17.171854 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0686b736_3cb6_4fb9_8710_8359e2066637.slice/crio-d60744db0312b1df98f6f9a820dad4948ce7310ee9acb3c84f018a3c9e09d1ab WatchSource:0}: Error finding container d60744db0312b1df98f6f9a820dad4948ce7310ee9acb3c84f018a3c9e09d1ab: Status 404 returned error can't find the container with id d60744db0312b1df98f6f9a820dad4948ce7310ee9acb3c84f018a3c9e09d1ab Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.207844 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-m9hcq" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.242966 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/383c4b7d-e18d-4973-a279-191f7800c32c-oauth-serving-cert\") pod \"console-54477f4d4-6cvcc\" (UID: \"383c4b7d-e18d-4973-a279-191f7800c32c\") " pod="openshift-console/console-54477f4d4-6cvcc" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.243016 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/383c4b7d-e18d-4973-a279-191f7800c32c-service-ca\") pod \"console-54477f4d4-6cvcc\" (UID: \"383c4b7d-e18d-4973-a279-191f7800c32c\") " pod="openshift-console/console-54477f4d4-6cvcc" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.243046 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/383c4b7d-e18d-4973-a279-191f7800c32c-console-oauth-config\") pod \"console-54477f4d4-6cvcc\" (UID: \"383c4b7d-e18d-4973-a279-191f7800c32c\") " pod="openshift-console/console-54477f4d4-6cvcc" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.243083 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/383c4b7d-e18d-4973-a279-191f7800c32c-console-serving-cert\") pod \"console-54477f4d4-6cvcc\" (UID: \"383c4b7d-e18d-4973-a279-191f7800c32c\") " pod="openshift-console/console-54477f4d4-6cvcc" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.243171 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/383c4b7d-e18d-4973-a279-191f7800c32c-console-config\") pod \"console-54477f4d4-6cvcc\" (UID: \"383c4b7d-e18d-4973-a279-191f7800c32c\") " pod="openshift-console/console-54477f4d4-6cvcc" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.243274 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhb7x\" (UniqueName: \"kubernetes.io/projected/383c4b7d-e18d-4973-a279-191f7800c32c-kube-api-access-qhb7x\") pod \"console-54477f4d4-6cvcc\" (UID: \"383c4b7d-e18d-4973-a279-191f7800c32c\") " pod="openshift-console/console-54477f4d4-6cvcc" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.243314 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/383c4b7d-e18d-4973-a279-191f7800c32c-trusted-ca-bundle\") pod \"console-54477f4d4-6cvcc\" (UID: \"383c4b7d-e18d-4973-a279-191f7800c32c\") " pod="openshift-console/console-54477f4d4-6cvcc" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.344276 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/383c4b7d-e18d-4973-a279-191f7800c32c-console-config\") pod \"console-54477f4d4-6cvcc\" (UID: \"383c4b7d-e18d-4973-a279-191f7800c32c\") " pod="openshift-console/console-54477f4d4-6cvcc" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.344336 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhb7x\" (UniqueName: \"kubernetes.io/projected/383c4b7d-e18d-4973-a279-191f7800c32c-kube-api-access-qhb7x\") pod \"console-54477f4d4-6cvcc\" (UID: \"383c4b7d-e18d-4973-a279-191f7800c32c\") " pod="openshift-console/console-54477f4d4-6cvcc" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.344356 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/383c4b7d-e18d-4973-a279-191f7800c32c-trusted-ca-bundle\") pod \"console-54477f4d4-6cvcc\" (UID: \"383c4b7d-e18d-4973-a279-191f7800c32c\") " pod="openshift-console/console-54477f4d4-6cvcc" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.344390 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/383c4b7d-e18d-4973-a279-191f7800c32c-oauth-serving-cert\") pod \"console-54477f4d4-6cvcc\" (UID: \"383c4b7d-e18d-4973-a279-191f7800c32c\") " pod="openshift-console/console-54477f4d4-6cvcc" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.344408 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/383c4b7d-e18d-4973-a279-191f7800c32c-service-ca\") pod \"console-54477f4d4-6cvcc\" (UID: \"383c4b7d-e18d-4973-a279-191f7800c32c\") " pod="openshift-console/console-54477f4d4-6cvcc" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.344427 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/383c4b7d-e18d-4973-a279-191f7800c32c-console-oauth-config\") pod \"console-54477f4d4-6cvcc\" (UID: \"383c4b7d-e18d-4973-a279-191f7800c32c\") " pod="openshift-console/console-54477f4d4-6cvcc" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.344481 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/383c4b7d-e18d-4973-a279-191f7800c32c-console-serving-cert\") pod \"console-54477f4d4-6cvcc\" (UID: \"383c4b7d-e18d-4973-a279-191f7800c32c\") " pod="openshift-console/console-54477f4d4-6cvcc" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.346593 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/383c4b7d-e18d-4973-a279-191f7800c32c-trusted-ca-bundle\") pod \"console-54477f4d4-6cvcc\" (UID: \"383c4b7d-e18d-4973-a279-191f7800c32c\") " pod="openshift-console/console-54477f4d4-6cvcc" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.347290 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/383c4b7d-e18d-4973-a279-191f7800c32c-service-ca\") pod \"console-54477f4d4-6cvcc\" (UID: \"383c4b7d-e18d-4973-a279-191f7800c32c\") " pod="openshift-console/console-54477f4d4-6cvcc" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.347773 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/383c4b7d-e18d-4973-a279-191f7800c32c-console-config\") pod \"console-54477f4d4-6cvcc\" (UID: \"383c4b7d-e18d-4973-a279-191f7800c32c\") " pod="openshift-console/console-54477f4d4-6cvcc" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.350649 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/383c4b7d-e18d-4973-a279-191f7800c32c-oauth-serving-cert\") pod \"console-54477f4d4-6cvcc\" (UID: \"383c4b7d-e18d-4973-a279-191f7800c32c\") " pod="openshift-console/console-54477f4d4-6cvcc" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.350662 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/383c4b7d-e18d-4973-a279-191f7800c32c-console-oauth-config\") pod \"console-54477f4d4-6cvcc\" (UID: \"383c4b7d-e18d-4973-a279-191f7800c32c\") " pod="openshift-console/console-54477f4d4-6cvcc" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.350746 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/383c4b7d-e18d-4973-a279-191f7800c32c-console-serving-cert\") pod \"console-54477f4d4-6cvcc\" (UID: \"383c4b7d-e18d-4973-a279-191f7800c32c\") " pod="openshift-console/console-54477f4d4-6cvcc" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.362981 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhb7x\" (UniqueName: \"kubernetes.io/projected/383c4b7d-e18d-4973-a279-191f7800c32c-kube-api-access-qhb7x\") pod \"console-54477f4d4-6cvcc\" (UID: \"383c4b7d-e18d-4973-a279-191f7800c32c\") " pod="openshift-console/console-54477f4d4-6cvcc" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.413201 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-54477f4d4-6cvcc" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.445924 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/97a282f2-3d9c-4911-bb64-4593b586bb4e-tls-key-pair\") pod \"nmstate-webhook-786f45cff4-gg78f\" (UID: \"97a282f2-3d9c-4911-bb64-4593b586bb4e\") " pod="openshift-nmstate/nmstate-webhook-786f45cff4-gg78f" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.452619 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/97a282f2-3d9c-4911-bb64-4593b586bb4e-tls-key-pair\") pod \"nmstate-webhook-786f45cff4-gg78f\" (UID: \"97a282f2-3d9c-4911-bb64-4593b586bb4e\") " pod="openshift-nmstate/nmstate-webhook-786f45cff4-gg78f" Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.532769 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-69594cc75-znbzd"] Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.587544 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-54477f4d4-6cvcc"] Feb 27 17:15:17 crc kubenswrapper[4700]: W0227 17:15:17.590357 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod383c4b7d_e18d_4973_a279_191f7800c32c.slice/crio-175ccfc4741f9b3efe2b669c6f9ad2bb71d3c53048cdeeab6b869166eb88b32e WatchSource:0}: Error finding container 175ccfc4741f9b3efe2b669c6f9ad2bb71d3c53048cdeeab6b869166eb88b32e: Status 404 returned error can't find the container with id 175ccfc4741f9b3efe2b669c6f9ad2bb71d3c53048cdeeab6b869166eb88b32e Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.623839 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-m9hcq"] Feb 27 17:15:17 crc kubenswrapper[4700]: W0227 17:15:17.628404 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfbb064be_36df_4a47_8c0e_9ffdd43d1123.slice/crio-ed7191cd059c9e3db55e24ed332edd95c86f8b55849a2bc632b6a82466ae2465 WatchSource:0}: Error finding container ed7191cd059c9e3db55e24ed332edd95c86f8b55849a2bc632b6a82466ae2465: Status 404 returned error can't find the container with id ed7191cd059c9e3db55e24ed332edd95c86f8b55849a2bc632b6a82466ae2465 Feb 27 17:15:17 crc kubenswrapper[4700]: I0227 17:15:17.697440 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-786f45cff4-gg78f" Feb 27 17:15:18 crc kubenswrapper[4700]: I0227 17:15:18.059172 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-m9hcq" event={"ID":"fbb064be-36df-4a47-8c0e-9ffdd43d1123","Type":"ContainerStarted","Data":"ed7191cd059c9e3db55e24ed332edd95c86f8b55849a2bc632b6a82466ae2465"} Feb 27 17:15:18 crc kubenswrapper[4700]: I0227 17:15:18.060524 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-69594cc75-znbzd" event={"ID":"47c2c9bd-d6c8-4b33-974c-ab0e57ea05fc","Type":"ContainerStarted","Data":"9584a87a5db8ee5d7815ad61e157685e14fe6ab1f0df2058f72b154f4dee2c57"} Feb 27 17:15:18 crc kubenswrapper[4700]: I0227 17:15:18.065539 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-54477f4d4-6cvcc" event={"ID":"383c4b7d-e18d-4973-a279-191f7800c32c","Type":"ContainerStarted","Data":"9d8849dbf609c96ca8e7af8b41961a9e57118d257bfc74d43654006d7b9834a9"} Feb 27 17:15:18 crc kubenswrapper[4700]: I0227 17:15:18.065631 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-54477f4d4-6cvcc" event={"ID":"383c4b7d-e18d-4973-a279-191f7800c32c","Type":"ContainerStarted","Data":"175ccfc4741f9b3efe2b669c6f9ad2bb71d3c53048cdeeab6b869166eb88b32e"} Feb 27 17:15:18 crc kubenswrapper[4700]: I0227 17:15:18.067138 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-85tkk" event={"ID":"0686b736-3cb6-4fb9-8710-8359e2066637","Type":"ContainerStarted","Data":"d60744db0312b1df98f6f9a820dad4948ce7310ee9acb3c84f018a3c9e09d1ab"} Feb 27 17:15:18 crc kubenswrapper[4700]: I0227 17:15:18.092423 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-54477f4d4-6cvcc" podStartSLOduration=1.092399769 podStartE2EDuration="1.092399769s" podCreationTimestamp="2026-02-27 17:15:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:15:18.087336394 +0000 UTC m=+878.072649171" watchObservedRunningTime="2026-02-27 17:15:18.092399769 +0000 UTC m=+878.077712536" Feb 27 17:15:18 crc kubenswrapper[4700]: I0227 17:15:18.128751 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-786f45cff4-gg78f"] Feb 27 17:15:19 crc kubenswrapper[4700]: I0227 17:15:19.080512 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-786f45cff4-gg78f" event={"ID":"97a282f2-3d9c-4911-bb64-4593b586bb4e","Type":"ContainerStarted","Data":"65dbf742ac860658e4355c7bed1517d022cd187503e4921dc3492881626b986c"} Feb 27 17:15:20 crc kubenswrapper[4700]: I0227 17:15:20.348507 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qclmh" Feb 27 17:15:20 crc kubenswrapper[4700]: I0227 17:15:20.401009 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qclmh" Feb 27 17:15:20 crc kubenswrapper[4700]: I0227 17:15:20.585829 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qclmh"] Feb 27 17:15:21 crc kubenswrapper[4700]: I0227 17:15:21.099251 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-69594cc75-znbzd" event={"ID":"47c2c9bd-d6c8-4b33-974c-ab0e57ea05fc","Type":"ContainerStarted","Data":"9626fe86ea9a7e39298d5a05b4078573f4fce753942de672b0142f5bde3afcc0"} Feb 27 17:15:21 crc kubenswrapper[4700]: I0227 17:15:21.104622 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-85tkk" event={"ID":"0686b736-3cb6-4fb9-8710-8359e2066637","Type":"ContainerStarted","Data":"aad4d012a81b4aa711c4a341694151b5edd8d27838081619853083baee9a095b"} Feb 27 17:15:21 crc kubenswrapper[4700]: I0227 17:15:21.104963 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-85tkk" Feb 27 17:15:21 crc kubenswrapper[4700]: I0227 17:15:21.105893 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-786f45cff4-gg78f" event={"ID":"97a282f2-3d9c-4911-bb64-4593b586bb4e","Type":"ContainerStarted","Data":"f96766a908bae6ecff11987e65422b4d16c97d825f6cfeb84defa2471bc027a1"} Feb 27 17:15:21 crc kubenswrapper[4700]: I0227 17:15:21.106352 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-786f45cff4-gg78f" Feb 27 17:15:21 crc kubenswrapper[4700]: I0227 17:15:21.109030 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-m9hcq" event={"ID":"fbb064be-36df-4a47-8c0e-9ffdd43d1123","Type":"ContainerStarted","Data":"ca87e74a9306b051eb9132042883699360f57ecd6851b745602ab93a8b6e0d9c"} Feb 27 17:15:21 crc kubenswrapper[4700]: I0227 17:15:21.124972 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-85tkk" podStartSLOduration=1.866731073 podStartE2EDuration="5.124953757s" podCreationTimestamp="2026-02-27 17:15:16 +0000 UTC" firstStartedPulling="2026-02-27 17:15:17.173784099 +0000 UTC m=+877.159096846" lastFinishedPulling="2026-02-27 17:15:20.432006783 +0000 UTC m=+880.417319530" observedRunningTime="2026-02-27 17:15:21.120698904 +0000 UTC m=+881.106011671" watchObservedRunningTime="2026-02-27 17:15:21.124953757 +0000 UTC m=+881.110266514" Feb 27 17:15:21 crc kubenswrapper[4700]: I0227 17:15:21.138225 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-m9hcq" podStartSLOduration=2.341128531 podStartE2EDuration="5.13820395s" podCreationTimestamp="2026-02-27 17:15:16 +0000 UTC" firstStartedPulling="2026-02-27 17:15:17.632652353 +0000 UTC m=+877.617965100" lastFinishedPulling="2026-02-27 17:15:20.429727772 +0000 UTC m=+880.415040519" observedRunningTime="2026-02-27 17:15:21.136187096 +0000 UTC m=+881.121499853" watchObservedRunningTime="2026-02-27 17:15:21.13820395 +0000 UTC m=+881.123516697" Feb 27 17:15:21 crc kubenswrapper[4700]: I0227 17:15:21.157061 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-786f45cff4-gg78f" podStartSLOduration=2.865144007 podStartE2EDuration="5.157039211s" podCreationTimestamp="2026-02-27 17:15:16 +0000 UTC" firstStartedPulling="2026-02-27 17:15:18.137265843 +0000 UTC m=+878.122578590" lastFinishedPulling="2026-02-27 17:15:20.429161037 +0000 UTC m=+880.414473794" observedRunningTime="2026-02-27 17:15:21.151305568 +0000 UTC m=+881.136618355" watchObservedRunningTime="2026-02-27 17:15:21.157039211 +0000 UTC m=+881.142351968" Feb 27 17:15:22 crc kubenswrapper[4700]: I0227 17:15:22.117235 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qclmh" podUID="0282bde1-f6fd-4d4f-a8d9-0342a9b1243d" containerName="registry-server" containerID="cri-o://b228710907b61dce4cce3ae5711b13c5baa71797250746f541e0fa3d68b84aeb" gracePeriod=2 Feb 27 17:15:22 crc kubenswrapper[4700]: I0227 17:15:22.921236 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qclmh" Feb 27 17:15:23 crc kubenswrapper[4700]: I0227 17:15:23.045067 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0282bde1-f6fd-4d4f-a8d9-0342a9b1243d-catalog-content\") pod \"0282bde1-f6fd-4d4f-a8d9-0342a9b1243d\" (UID: \"0282bde1-f6fd-4d4f-a8d9-0342a9b1243d\") " Feb 27 17:15:23 crc kubenswrapper[4700]: I0227 17:15:23.045336 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w54d9\" (UniqueName: \"kubernetes.io/projected/0282bde1-f6fd-4d4f-a8d9-0342a9b1243d-kube-api-access-w54d9\") pod \"0282bde1-f6fd-4d4f-a8d9-0342a9b1243d\" (UID: \"0282bde1-f6fd-4d4f-a8d9-0342a9b1243d\") " Feb 27 17:15:23 crc kubenswrapper[4700]: I0227 17:15:23.046255 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0282bde1-f6fd-4d4f-a8d9-0342a9b1243d-utilities" (OuterVolumeSpecName: "utilities") pod "0282bde1-f6fd-4d4f-a8d9-0342a9b1243d" (UID: "0282bde1-f6fd-4d4f-a8d9-0342a9b1243d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:15:23 crc kubenswrapper[4700]: I0227 17:15:23.046387 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0282bde1-f6fd-4d4f-a8d9-0342a9b1243d-utilities\") pod \"0282bde1-f6fd-4d4f-a8d9-0342a9b1243d\" (UID: \"0282bde1-f6fd-4d4f-a8d9-0342a9b1243d\") " Feb 27 17:15:23 crc kubenswrapper[4700]: I0227 17:15:23.046910 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0282bde1-f6fd-4d4f-a8d9-0342a9b1243d-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:15:23 crc kubenswrapper[4700]: I0227 17:15:23.050439 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0282bde1-f6fd-4d4f-a8d9-0342a9b1243d-kube-api-access-w54d9" (OuterVolumeSpecName: "kube-api-access-w54d9") pod "0282bde1-f6fd-4d4f-a8d9-0342a9b1243d" (UID: "0282bde1-f6fd-4d4f-a8d9-0342a9b1243d"). InnerVolumeSpecName "kube-api-access-w54d9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:15:23 crc kubenswrapper[4700]: I0227 17:15:23.125107 4700 generic.go:334] "Generic (PLEG): container finished" podID="0282bde1-f6fd-4d4f-a8d9-0342a9b1243d" containerID="b228710907b61dce4cce3ae5711b13c5baa71797250746f541e0fa3d68b84aeb" exitCode=0 Feb 27 17:15:23 crc kubenswrapper[4700]: I0227 17:15:23.125619 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qclmh" event={"ID":"0282bde1-f6fd-4d4f-a8d9-0342a9b1243d","Type":"ContainerDied","Data":"b228710907b61dce4cce3ae5711b13c5baa71797250746f541e0fa3d68b84aeb"} Feb 27 17:15:23 crc kubenswrapper[4700]: I0227 17:15:23.126420 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qclmh" event={"ID":"0282bde1-f6fd-4d4f-a8d9-0342a9b1243d","Type":"ContainerDied","Data":"2949f6687fa0d99f6253fb02b0e53948f3e793f52200d8cdf23e5e9b9861bc4a"} Feb 27 17:15:23 crc kubenswrapper[4700]: I0227 17:15:23.125689 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qclmh" Feb 27 17:15:23 crc kubenswrapper[4700]: I0227 17:15:23.126508 4700 scope.go:117] "RemoveContainer" containerID="b228710907b61dce4cce3ae5711b13c5baa71797250746f541e0fa3d68b84aeb" Feb 27 17:15:23 crc kubenswrapper[4700]: I0227 17:15:23.128164 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-69594cc75-znbzd" event={"ID":"47c2c9bd-d6c8-4b33-974c-ab0e57ea05fc","Type":"ContainerStarted","Data":"ada4b2460df82c6ab6edafe9b60b9853cea88836deaca255bd719537f6011c2b"} Feb 27 17:15:23 crc kubenswrapper[4700]: I0227 17:15:23.142355 4700 scope.go:117] "RemoveContainer" containerID="f7f3dcec8e488f21c91463bc5f997d65dbc134fb2145474cd6926537b5db0e34" Feb 27 17:15:23 crc kubenswrapper[4700]: I0227 17:15:23.150229 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w54d9\" (UniqueName: \"kubernetes.io/projected/0282bde1-f6fd-4d4f-a8d9-0342a9b1243d-kube-api-access-w54d9\") on node \"crc\" DevicePath \"\"" Feb 27 17:15:23 crc kubenswrapper[4700]: I0227 17:15:23.163608 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-69594cc75-znbzd" podStartSLOduration=1.749577776 podStartE2EDuration="7.163583559s" podCreationTimestamp="2026-02-27 17:15:16 +0000 UTC" firstStartedPulling="2026-02-27 17:15:17.542176315 +0000 UTC m=+877.527489062" lastFinishedPulling="2026-02-27 17:15:22.956182098 +0000 UTC m=+882.941494845" observedRunningTime="2026-02-27 17:15:23.149431632 +0000 UTC m=+883.134744409" watchObservedRunningTime="2026-02-27 17:15:23.163583559 +0000 UTC m=+883.148896306" Feb 27 17:15:23 crc kubenswrapper[4700]: I0227 17:15:23.176936 4700 scope.go:117] "RemoveContainer" containerID="1801e3857b90bc90aa3d362b761d292b8f7f4d4c1ffad8d1c05a7348b734c0b6" Feb 27 17:15:23 crc kubenswrapper[4700]: I0227 17:15:23.185803 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0282bde1-f6fd-4d4f-a8d9-0342a9b1243d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0282bde1-f6fd-4d4f-a8d9-0342a9b1243d" (UID: "0282bde1-f6fd-4d4f-a8d9-0342a9b1243d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:15:23 crc kubenswrapper[4700]: I0227 17:15:23.195778 4700 scope.go:117] "RemoveContainer" containerID="b228710907b61dce4cce3ae5711b13c5baa71797250746f541e0fa3d68b84aeb" Feb 27 17:15:23 crc kubenswrapper[4700]: E0227 17:15:23.196736 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b228710907b61dce4cce3ae5711b13c5baa71797250746f541e0fa3d68b84aeb\": container with ID starting with b228710907b61dce4cce3ae5711b13c5baa71797250746f541e0fa3d68b84aeb not found: ID does not exist" containerID="b228710907b61dce4cce3ae5711b13c5baa71797250746f541e0fa3d68b84aeb" Feb 27 17:15:23 crc kubenswrapper[4700]: I0227 17:15:23.196772 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b228710907b61dce4cce3ae5711b13c5baa71797250746f541e0fa3d68b84aeb"} err="failed to get container status \"b228710907b61dce4cce3ae5711b13c5baa71797250746f541e0fa3d68b84aeb\": rpc error: code = NotFound desc = could not find container \"b228710907b61dce4cce3ae5711b13c5baa71797250746f541e0fa3d68b84aeb\": container with ID starting with b228710907b61dce4cce3ae5711b13c5baa71797250746f541e0fa3d68b84aeb not found: ID does not exist" Feb 27 17:15:23 crc kubenswrapper[4700]: I0227 17:15:23.196795 4700 scope.go:117] "RemoveContainer" containerID="f7f3dcec8e488f21c91463bc5f997d65dbc134fb2145474cd6926537b5db0e34" Feb 27 17:15:23 crc kubenswrapper[4700]: E0227 17:15:23.197168 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7f3dcec8e488f21c91463bc5f997d65dbc134fb2145474cd6926537b5db0e34\": container with ID starting with f7f3dcec8e488f21c91463bc5f997d65dbc134fb2145474cd6926537b5db0e34 not found: ID does not exist" containerID="f7f3dcec8e488f21c91463bc5f997d65dbc134fb2145474cd6926537b5db0e34" Feb 27 17:15:23 crc kubenswrapper[4700]: I0227 17:15:23.197232 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7f3dcec8e488f21c91463bc5f997d65dbc134fb2145474cd6926537b5db0e34"} err="failed to get container status \"f7f3dcec8e488f21c91463bc5f997d65dbc134fb2145474cd6926537b5db0e34\": rpc error: code = NotFound desc = could not find container \"f7f3dcec8e488f21c91463bc5f997d65dbc134fb2145474cd6926537b5db0e34\": container with ID starting with f7f3dcec8e488f21c91463bc5f997d65dbc134fb2145474cd6926537b5db0e34 not found: ID does not exist" Feb 27 17:15:23 crc kubenswrapper[4700]: I0227 17:15:23.197281 4700 scope.go:117] "RemoveContainer" containerID="1801e3857b90bc90aa3d362b761d292b8f7f4d4c1ffad8d1c05a7348b734c0b6" Feb 27 17:15:23 crc kubenswrapper[4700]: E0227 17:15:23.197673 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1801e3857b90bc90aa3d362b761d292b8f7f4d4c1ffad8d1c05a7348b734c0b6\": container with ID starting with 1801e3857b90bc90aa3d362b761d292b8f7f4d4c1ffad8d1c05a7348b734c0b6 not found: ID does not exist" containerID="1801e3857b90bc90aa3d362b761d292b8f7f4d4c1ffad8d1c05a7348b734c0b6" Feb 27 17:15:23 crc kubenswrapper[4700]: I0227 17:15:23.197729 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1801e3857b90bc90aa3d362b761d292b8f7f4d4c1ffad8d1c05a7348b734c0b6"} err="failed to get container status \"1801e3857b90bc90aa3d362b761d292b8f7f4d4c1ffad8d1c05a7348b734c0b6\": rpc error: code = NotFound desc = could not find container \"1801e3857b90bc90aa3d362b761d292b8f7f4d4c1ffad8d1c05a7348b734c0b6\": container with ID starting with 1801e3857b90bc90aa3d362b761d292b8f7f4d4c1ffad8d1c05a7348b734c0b6 not found: ID does not exist" Feb 27 17:15:23 crc kubenswrapper[4700]: I0227 17:15:23.251324 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0282bde1-f6fd-4d4f-a8d9-0342a9b1243d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:15:23 crc kubenswrapper[4700]: I0227 17:15:23.471539 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qclmh"] Feb 27 17:15:23 crc kubenswrapper[4700]: I0227 17:15:23.482439 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qclmh"] Feb 27 17:15:24 crc kubenswrapper[4700]: I0227 17:15:24.994946 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0282bde1-f6fd-4d4f-a8d9-0342a9b1243d" path="/var/lib/kubelet/pods/0282bde1-f6fd-4d4f-a8d9-0342a9b1243d/volumes" Feb 27 17:15:27 crc kubenswrapper[4700]: I0227 17:15:27.156845 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-85tkk" Feb 27 17:15:27 crc kubenswrapper[4700]: I0227 17:15:27.414022 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-54477f4d4-6cvcc" Feb 27 17:15:27 crc kubenswrapper[4700]: I0227 17:15:27.414111 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-54477f4d4-6cvcc" Feb 27 17:15:27 crc kubenswrapper[4700]: I0227 17:15:27.420543 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-54477f4d4-6cvcc" Feb 27 17:15:28 crc kubenswrapper[4700]: I0227 17:15:28.168877 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-54477f4d4-6cvcc" Feb 27 17:15:28 crc kubenswrapper[4700]: I0227 17:15:28.219794 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-kq9sb"] Feb 27 17:15:37 crc kubenswrapper[4700]: I0227 17:15:37.704388 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-786f45cff4-gg78f" Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.267616 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-kq9sb" podUID="facc8439-0b4d-4228-95b2-d01d48ee1c53" containerName="console" containerID="cri-o://40567104d50ad623de5ae2b8117b2c4f946f7c9d2cd7b29ef18d788df2360bb2" gracePeriod=15 Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.610563 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d"] Feb 27 17:15:53 crc kubenswrapper[4700]: E0227 17:15:53.611071 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0282bde1-f6fd-4d4f-a8d9-0342a9b1243d" containerName="extract-utilities" Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.611084 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="0282bde1-f6fd-4d4f-a8d9-0342a9b1243d" containerName="extract-utilities" Feb 27 17:15:53 crc kubenswrapper[4700]: E0227 17:15:53.611094 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0282bde1-f6fd-4d4f-a8d9-0342a9b1243d" containerName="registry-server" Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.611100 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="0282bde1-f6fd-4d4f-a8d9-0342a9b1243d" containerName="registry-server" Feb 27 17:15:53 crc kubenswrapper[4700]: E0227 17:15:53.611111 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0282bde1-f6fd-4d4f-a8d9-0342a9b1243d" containerName="extract-content" Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.611116 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="0282bde1-f6fd-4d4f-a8d9-0342a9b1243d" containerName="extract-content" Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.611247 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="0282bde1-f6fd-4d4f-a8d9-0342a9b1243d" containerName="registry-server" Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.612020 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d" Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.615067 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.626577 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d"] Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.709103 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-kq9sb_facc8439-0b4d-4228-95b2-d01d48ee1c53/console/0.log" Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.709175 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-kq9sb" Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.798879 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/facc8439-0b4d-4228-95b2-d01d48ee1c53-console-oauth-config\") pod \"facc8439-0b4d-4228-95b2-d01d48ee1c53\" (UID: \"facc8439-0b4d-4228-95b2-d01d48ee1c53\") " Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.798953 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/facc8439-0b4d-4228-95b2-d01d48ee1c53-console-config\") pod \"facc8439-0b4d-4228-95b2-d01d48ee1c53\" (UID: \"facc8439-0b4d-4228-95b2-d01d48ee1c53\") " Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.799061 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/facc8439-0b4d-4228-95b2-d01d48ee1c53-trusted-ca-bundle\") pod \"facc8439-0b4d-4228-95b2-d01d48ee1c53\" (UID: \"facc8439-0b4d-4228-95b2-d01d48ee1c53\") " Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.799108 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/facc8439-0b4d-4228-95b2-d01d48ee1c53-console-serving-cert\") pod \"facc8439-0b4d-4228-95b2-d01d48ee1c53\" (UID: \"facc8439-0b4d-4228-95b2-d01d48ee1c53\") " Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.799167 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/facc8439-0b4d-4228-95b2-d01d48ee1c53-service-ca\") pod \"facc8439-0b4d-4228-95b2-d01d48ee1c53\" (UID: \"facc8439-0b4d-4228-95b2-d01d48ee1c53\") " Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.799198 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/facc8439-0b4d-4228-95b2-d01d48ee1c53-oauth-serving-cert\") pod \"facc8439-0b4d-4228-95b2-d01d48ee1c53\" (UID: \"facc8439-0b4d-4228-95b2-d01d48ee1c53\") " Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.799277 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6f4c\" (UniqueName: \"kubernetes.io/projected/facc8439-0b4d-4228-95b2-d01d48ee1c53-kube-api-access-d6f4c\") pod \"facc8439-0b4d-4228-95b2-d01d48ee1c53\" (UID: \"facc8439-0b4d-4228-95b2-d01d48ee1c53\") " Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.799512 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ccc245e7-023f-4d49-99d0-9ebc992a9ccc-bundle\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d\" (UID: \"ccc245e7-023f-4d49-99d0-9ebc992a9ccc\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d" Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.799558 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ccc245e7-023f-4d49-99d0-9ebc992a9ccc-util\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d\" (UID: \"ccc245e7-023f-4d49-99d0-9ebc992a9ccc\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d" Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.799622 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55kzl\" (UniqueName: \"kubernetes.io/projected/ccc245e7-023f-4d49-99d0-9ebc992a9ccc-kube-api-access-55kzl\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d\" (UID: \"ccc245e7-023f-4d49-99d0-9ebc992a9ccc\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d" Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.800186 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/facc8439-0b4d-4228-95b2-d01d48ee1c53-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "facc8439-0b4d-4228-95b2-d01d48ee1c53" (UID: "facc8439-0b4d-4228-95b2-d01d48ee1c53"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.800400 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/facc8439-0b4d-4228-95b2-d01d48ee1c53-console-config" (OuterVolumeSpecName: "console-config") pod "facc8439-0b4d-4228-95b2-d01d48ee1c53" (UID: "facc8439-0b4d-4228-95b2-d01d48ee1c53"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.800811 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/facc8439-0b4d-4228-95b2-d01d48ee1c53-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "facc8439-0b4d-4228-95b2-d01d48ee1c53" (UID: "facc8439-0b4d-4228-95b2-d01d48ee1c53"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.801121 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/facc8439-0b4d-4228-95b2-d01d48ee1c53-service-ca" (OuterVolumeSpecName: "service-ca") pod "facc8439-0b4d-4228-95b2-d01d48ee1c53" (UID: "facc8439-0b4d-4228-95b2-d01d48ee1c53"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.808998 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/facc8439-0b4d-4228-95b2-d01d48ee1c53-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "facc8439-0b4d-4228-95b2-d01d48ee1c53" (UID: "facc8439-0b4d-4228-95b2-d01d48ee1c53"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.809048 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/facc8439-0b4d-4228-95b2-d01d48ee1c53-kube-api-access-d6f4c" (OuterVolumeSpecName: "kube-api-access-d6f4c") pod "facc8439-0b4d-4228-95b2-d01d48ee1c53" (UID: "facc8439-0b4d-4228-95b2-d01d48ee1c53"). InnerVolumeSpecName "kube-api-access-d6f4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.809493 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/facc8439-0b4d-4228-95b2-d01d48ee1c53-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "facc8439-0b4d-4228-95b2-d01d48ee1c53" (UID: "facc8439-0b4d-4228-95b2-d01d48ee1c53"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.900554 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55kzl\" (UniqueName: \"kubernetes.io/projected/ccc245e7-023f-4d49-99d0-9ebc992a9ccc-kube-api-access-55kzl\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d\" (UID: \"ccc245e7-023f-4d49-99d0-9ebc992a9ccc\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d" Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.900859 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ccc245e7-023f-4d49-99d0-9ebc992a9ccc-bundle\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d\" (UID: \"ccc245e7-023f-4d49-99d0-9ebc992a9ccc\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d" Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.900928 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ccc245e7-023f-4d49-99d0-9ebc992a9ccc-util\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d\" (UID: \"ccc245e7-023f-4d49-99d0-9ebc992a9ccc\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d" Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.901043 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6f4c\" (UniqueName: \"kubernetes.io/projected/facc8439-0b4d-4228-95b2-d01d48ee1c53-kube-api-access-d6f4c\") on node \"crc\" DevicePath \"\"" Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.901102 4700 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/facc8439-0b4d-4228-95b2-d01d48ee1c53-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.901159 4700 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/facc8439-0b4d-4228-95b2-d01d48ee1c53-console-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.901217 4700 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/facc8439-0b4d-4228-95b2-d01d48ee1c53-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.901268 4700 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/facc8439-0b4d-4228-95b2-d01d48ee1c53-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.901331 4700 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/facc8439-0b4d-4228-95b2-d01d48ee1c53-service-ca\") on node \"crc\" DevicePath \"\"" Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.901392 4700 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/facc8439-0b4d-4228-95b2-d01d48ee1c53-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.901893 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ccc245e7-023f-4d49-99d0-9ebc992a9ccc-util\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d\" (UID: \"ccc245e7-023f-4d49-99d0-9ebc992a9ccc\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d" Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.902428 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ccc245e7-023f-4d49-99d0-9ebc992a9ccc-bundle\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d\" (UID: \"ccc245e7-023f-4d49-99d0-9ebc992a9ccc\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d" Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.920090 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55kzl\" (UniqueName: \"kubernetes.io/projected/ccc245e7-023f-4d49-99d0-9ebc992a9ccc-kube-api-access-55kzl\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d\" (UID: \"ccc245e7-023f-4d49-99d0-9ebc992a9ccc\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d" Feb 27 17:15:53 crc kubenswrapper[4700]: I0227 17:15:53.925640 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d" Feb 27 17:15:54 crc kubenswrapper[4700]: I0227 17:15:54.203885 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d"] Feb 27 17:15:54 crc kubenswrapper[4700]: W0227 17:15:54.209854 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podccc245e7_023f_4d49_99d0_9ebc992a9ccc.slice/crio-e7cba9d43cb6da6a781f5f1ace482bf83aeb5bfdc7694b945f2819b90a06673c WatchSource:0}: Error finding container e7cba9d43cb6da6a781f5f1ace482bf83aeb5bfdc7694b945f2819b90a06673c: Status 404 returned error can't find the container with id e7cba9d43cb6da6a781f5f1ace482bf83aeb5bfdc7694b945f2819b90a06673c Feb 27 17:15:54 crc kubenswrapper[4700]: I0227 17:15:54.379853 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d" event={"ID":"ccc245e7-023f-4d49-99d0-9ebc992a9ccc","Type":"ContainerStarted","Data":"12830e70f82a47e2a925a1a9e17e307e2b733c96b1d9dd49506a9c643dd51927"} Feb 27 17:15:54 crc kubenswrapper[4700]: I0227 17:15:54.380729 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d" event={"ID":"ccc245e7-023f-4d49-99d0-9ebc992a9ccc","Type":"ContainerStarted","Data":"e7cba9d43cb6da6a781f5f1ace482bf83aeb5bfdc7694b945f2819b90a06673c"} Feb 27 17:15:54 crc kubenswrapper[4700]: I0227 17:15:54.383789 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-kq9sb_facc8439-0b4d-4228-95b2-d01d48ee1c53/console/0.log" Feb 27 17:15:54 crc kubenswrapper[4700]: I0227 17:15:54.383882 4700 generic.go:334] "Generic (PLEG): container finished" podID="facc8439-0b4d-4228-95b2-d01d48ee1c53" containerID="40567104d50ad623de5ae2b8117b2c4f946f7c9d2cd7b29ef18d788df2360bb2" exitCode=2 Feb 27 17:15:54 crc kubenswrapper[4700]: I0227 17:15:54.383919 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-kq9sb" event={"ID":"facc8439-0b4d-4228-95b2-d01d48ee1c53","Type":"ContainerDied","Data":"40567104d50ad623de5ae2b8117b2c4f946f7c9d2cd7b29ef18d788df2360bb2"} Feb 27 17:15:54 crc kubenswrapper[4700]: I0227 17:15:54.383951 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-kq9sb" event={"ID":"facc8439-0b4d-4228-95b2-d01d48ee1c53","Type":"ContainerDied","Data":"3c147cb863230b875f155f4e04d65e0368c2333d0ea58fd2097cc0b4a8f13f7e"} Feb 27 17:15:54 crc kubenswrapper[4700]: I0227 17:15:54.383949 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-kq9sb" Feb 27 17:15:54 crc kubenswrapper[4700]: I0227 17:15:54.383968 4700 scope.go:117] "RemoveContainer" containerID="40567104d50ad623de5ae2b8117b2c4f946f7c9d2cd7b29ef18d788df2360bb2" Feb 27 17:15:54 crc kubenswrapper[4700]: I0227 17:15:54.442417 4700 scope.go:117] "RemoveContainer" containerID="40567104d50ad623de5ae2b8117b2c4f946f7c9d2cd7b29ef18d788df2360bb2" Feb 27 17:15:54 crc kubenswrapper[4700]: E0227 17:15:54.444216 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40567104d50ad623de5ae2b8117b2c4f946f7c9d2cd7b29ef18d788df2360bb2\": container with ID starting with 40567104d50ad623de5ae2b8117b2c4f946f7c9d2cd7b29ef18d788df2360bb2 not found: ID does not exist" containerID="40567104d50ad623de5ae2b8117b2c4f946f7c9d2cd7b29ef18d788df2360bb2" Feb 27 17:15:54 crc kubenswrapper[4700]: I0227 17:15:54.444263 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40567104d50ad623de5ae2b8117b2c4f946f7c9d2cd7b29ef18d788df2360bb2"} err="failed to get container status \"40567104d50ad623de5ae2b8117b2c4f946f7c9d2cd7b29ef18d788df2360bb2\": rpc error: code = NotFound desc = could not find container \"40567104d50ad623de5ae2b8117b2c4f946f7c9d2cd7b29ef18d788df2360bb2\": container with ID starting with 40567104d50ad623de5ae2b8117b2c4f946f7c9d2cd7b29ef18d788df2360bb2 not found: ID does not exist" Feb 27 17:15:54 crc kubenswrapper[4700]: I0227 17:15:54.445694 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-kq9sb"] Feb 27 17:15:54 crc kubenswrapper[4700]: I0227 17:15:54.455590 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-kq9sb"] Feb 27 17:15:54 crc kubenswrapper[4700]: I0227 17:15:54.993891 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="facc8439-0b4d-4228-95b2-d01d48ee1c53" path="/var/lib/kubelet/pods/facc8439-0b4d-4228-95b2-d01d48ee1c53/volumes" Feb 27 17:15:55 crc kubenswrapper[4700]: I0227 17:15:55.396987 4700 generic.go:334] "Generic (PLEG): container finished" podID="ccc245e7-023f-4d49-99d0-9ebc992a9ccc" containerID="12830e70f82a47e2a925a1a9e17e307e2b733c96b1d9dd49506a9c643dd51927" exitCode=0 Feb 27 17:15:55 crc kubenswrapper[4700]: I0227 17:15:55.397076 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d" event={"ID":"ccc245e7-023f-4d49-99d0-9ebc992a9ccc","Type":"ContainerDied","Data":"12830e70f82a47e2a925a1a9e17e307e2b733c96b1d9dd49506a9c643dd51927"} Feb 27 17:15:58 crc kubenswrapper[4700]: I0227 17:15:58.424758 4700 generic.go:334] "Generic (PLEG): container finished" podID="ccc245e7-023f-4d49-99d0-9ebc992a9ccc" containerID="d6937e7cca3fc929f3827c0584022ebc922f3e5ef9961115a9a04e22092bfb75" exitCode=0 Feb 27 17:15:58 crc kubenswrapper[4700]: I0227 17:15:58.424823 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d" event={"ID":"ccc245e7-023f-4d49-99d0-9ebc992a9ccc","Type":"ContainerDied","Data":"d6937e7cca3fc929f3827c0584022ebc922f3e5ef9961115a9a04e22092bfb75"} Feb 27 17:15:59 crc kubenswrapper[4700]: I0227 17:15:59.439693 4700 generic.go:334] "Generic (PLEG): container finished" podID="ccc245e7-023f-4d49-99d0-9ebc992a9ccc" containerID="0e491656c6c1a8d48ee27a45f5e8e0c0bbbe54df7c6bd98bb14a4432b16ac289" exitCode=0 Feb 27 17:15:59 crc kubenswrapper[4700]: I0227 17:15:59.439768 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d" event={"ID":"ccc245e7-023f-4d49-99d0-9ebc992a9ccc","Type":"ContainerDied","Data":"0e491656c6c1a8d48ee27a45f5e8e0c0bbbe54df7c6bd98bb14a4432b16ac289"} Feb 27 17:16:00 crc kubenswrapper[4700]: I0227 17:16:00.143835 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536876-pxhvd"] Feb 27 17:16:00 crc kubenswrapper[4700]: E0227 17:16:00.144109 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="facc8439-0b4d-4228-95b2-d01d48ee1c53" containerName="console" Feb 27 17:16:00 crc kubenswrapper[4700]: I0227 17:16:00.144131 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="facc8439-0b4d-4228-95b2-d01d48ee1c53" containerName="console" Feb 27 17:16:00 crc kubenswrapper[4700]: I0227 17:16:00.144258 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="facc8439-0b4d-4228-95b2-d01d48ee1c53" containerName="console" Feb 27 17:16:00 crc kubenswrapper[4700]: I0227 17:16:00.144752 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536876-pxhvd" Feb 27 17:16:00 crc kubenswrapper[4700]: I0227 17:16:00.149363 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:16:00 crc kubenswrapper[4700]: I0227 17:16:00.156000 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d62kc" Feb 27 17:16:00 crc kubenswrapper[4700]: I0227 17:16:00.156178 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:16:00 crc kubenswrapper[4700]: I0227 17:16:00.171424 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536876-pxhvd"] Feb 27 17:16:00 crc kubenswrapper[4700]: I0227 17:16:00.329622 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcncs\" (UniqueName: \"kubernetes.io/projected/d1f9992d-c533-4184-9474-1cfe9b3e9819-kube-api-access-rcncs\") pod \"auto-csr-approver-29536876-pxhvd\" (UID: \"d1f9992d-c533-4184-9474-1cfe9b3e9819\") " pod="openshift-infra/auto-csr-approver-29536876-pxhvd" Feb 27 17:16:00 crc kubenswrapper[4700]: I0227 17:16:00.430919 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcncs\" (UniqueName: \"kubernetes.io/projected/d1f9992d-c533-4184-9474-1cfe9b3e9819-kube-api-access-rcncs\") pod \"auto-csr-approver-29536876-pxhvd\" (UID: \"d1f9992d-c533-4184-9474-1cfe9b3e9819\") " pod="openshift-infra/auto-csr-approver-29536876-pxhvd" Feb 27 17:16:00 crc kubenswrapper[4700]: I0227 17:16:00.457029 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcncs\" (UniqueName: \"kubernetes.io/projected/d1f9992d-c533-4184-9474-1cfe9b3e9819-kube-api-access-rcncs\") pod \"auto-csr-approver-29536876-pxhvd\" (UID: \"d1f9992d-c533-4184-9474-1cfe9b3e9819\") " pod="openshift-infra/auto-csr-approver-29536876-pxhvd" Feb 27 17:16:00 crc kubenswrapper[4700]: I0227 17:16:00.482843 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536876-pxhvd" Feb 27 17:16:00 crc kubenswrapper[4700]: I0227 17:16:00.700424 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536876-pxhvd"] Feb 27 17:16:00 crc kubenswrapper[4700]: W0227 17:16:00.705507 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd1f9992d_c533_4184_9474_1cfe9b3e9819.slice/crio-1aded29e16d8282dd37a01b03ffa070d41353431016adf4fcc2ccf85a33491d8 WatchSource:0}: Error finding container 1aded29e16d8282dd37a01b03ffa070d41353431016adf4fcc2ccf85a33491d8: Status 404 returned error can't find the container with id 1aded29e16d8282dd37a01b03ffa070d41353431016adf4fcc2ccf85a33491d8 Feb 27 17:16:00 crc kubenswrapper[4700]: I0227 17:16:00.713932 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d" Feb 27 17:16:00 crc kubenswrapper[4700]: I0227 17:16:00.836728 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ccc245e7-023f-4d49-99d0-9ebc992a9ccc-util\") pod \"ccc245e7-023f-4d49-99d0-9ebc992a9ccc\" (UID: \"ccc245e7-023f-4d49-99d0-9ebc992a9ccc\") " Feb 27 17:16:00 crc kubenswrapper[4700]: I0227 17:16:00.836838 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ccc245e7-023f-4d49-99d0-9ebc992a9ccc-bundle\") pod \"ccc245e7-023f-4d49-99d0-9ebc992a9ccc\" (UID: \"ccc245e7-023f-4d49-99d0-9ebc992a9ccc\") " Feb 27 17:16:00 crc kubenswrapper[4700]: I0227 17:16:00.836994 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55kzl\" (UniqueName: \"kubernetes.io/projected/ccc245e7-023f-4d49-99d0-9ebc992a9ccc-kube-api-access-55kzl\") pod \"ccc245e7-023f-4d49-99d0-9ebc992a9ccc\" (UID: \"ccc245e7-023f-4d49-99d0-9ebc992a9ccc\") " Feb 27 17:16:00 crc kubenswrapper[4700]: I0227 17:16:00.840313 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccc245e7-023f-4d49-99d0-9ebc992a9ccc-bundle" (OuterVolumeSpecName: "bundle") pod "ccc245e7-023f-4d49-99d0-9ebc992a9ccc" (UID: "ccc245e7-023f-4d49-99d0-9ebc992a9ccc"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:16:00 crc kubenswrapper[4700]: I0227 17:16:00.842339 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccc245e7-023f-4d49-99d0-9ebc992a9ccc-kube-api-access-55kzl" (OuterVolumeSpecName: "kube-api-access-55kzl") pod "ccc245e7-023f-4d49-99d0-9ebc992a9ccc" (UID: "ccc245e7-023f-4d49-99d0-9ebc992a9ccc"). InnerVolumeSpecName "kube-api-access-55kzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:16:00 crc kubenswrapper[4700]: I0227 17:16:00.859400 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccc245e7-023f-4d49-99d0-9ebc992a9ccc-util" (OuterVolumeSpecName: "util") pod "ccc245e7-023f-4d49-99d0-9ebc992a9ccc" (UID: "ccc245e7-023f-4d49-99d0-9ebc992a9ccc"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:16:00 crc kubenswrapper[4700]: I0227 17:16:00.939856 4700 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ccc245e7-023f-4d49-99d0-9ebc992a9ccc-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:16:00 crc kubenswrapper[4700]: I0227 17:16:00.939893 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55kzl\" (UniqueName: \"kubernetes.io/projected/ccc245e7-023f-4d49-99d0-9ebc992a9ccc-kube-api-access-55kzl\") on node \"crc\" DevicePath \"\"" Feb 27 17:16:00 crc kubenswrapper[4700]: I0227 17:16:00.939910 4700 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ccc245e7-023f-4d49-99d0-9ebc992a9ccc-util\") on node \"crc\" DevicePath \"\"" Feb 27 17:16:01 crc kubenswrapper[4700]: I0227 17:16:01.454668 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536876-pxhvd" event={"ID":"d1f9992d-c533-4184-9474-1cfe9b3e9819","Type":"ContainerStarted","Data":"1aded29e16d8282dd37a01b03ffa070d41353431016adf4fcc2ccf85a33491d8"} Feb 27 17:16:01 crc kubenswrapper[4700]: I0227 17:16:01.468563 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d" event={"ID":"ccc245e7-023f-4d49-99d0-9ebc992a9ccc","Type":"ContainerDied","Data":"e7cba9d43cb6da6a781f5f1ace482bf83aeb5bfdc7694b945f2819b90a06673c"} Feb 27 17:16:01 crc kubenswrapper[4700]: I0227 17:16:01.468612 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e7cba9d43cb6da6a781f5f1ace482bf83aeb5bfdc7694b945f2819b90a06673c" Feb 27 17:16:01 crc kubenswrapper[4700]: I0227 17:16:01.468714 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d" Feb 27 17:16:02 crc kubenswrapper[4700]: I0227 17:16:02.479917 4700 generic.go:334] "Generic (PLEG): container finished" podID="d1f9992d-c533-4184-9474-1cfe9b3e9819" containerID="cb8af2e94530ddd5327d653df36b6c92e24ad0db68ed6c6a9b9d9f95eb6a5e71" exitCode=0 Feb 27 17:16:02 crc kubenswrapper[4700]: I0227 17:16:02.480149 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536876-pxhvd" event={"ID":"d1f9992d-c533-4184-9474-1cfe9b3e9819","Type":"ContainerDied","Data":"cb8af2e94530ddd5327d653df36b6c92e24ad0db68ed6c6a9b9d9f95eb6a5e71"} Feb 27 17:16:03 crc kubenswrapper[4700]: I0227 17:16:03.770084 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536876-pxhvd" Feb 27 17:16:03 crc kubenswrapper[4700]: I0227 17:16:03.878424 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rcncs\" (UniqueName: \"kubernetes.io/projected/d1f9992d-c533-4184-9474-1cfe9b3e9819-kube-api-access-rcncs\") pod \"d1f9992d-c533-4184-9474-1cfe9b3e9819\" (UID: \"d1f9992d-c533-4184-9474-1cfe9b3e9819\") " Feb 27 17:16:03 crc kubenswrapper[4700]: I0227 17:16:03.885374 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1f9992d-c533-4184-9474-1cfe9b3e9819-kube-api-access-rcncs" (OuterVolumeSpecName: "kube-api-access-rcncs") pod "d1f9992d-c533-4184-9474-1cfe9b3e9819" (UID: "d1f9992d-c533-4184-9474-1cfe9b3e9819"). InnerVolumeSpecName "kube-api-access-rcncs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:16:03 crc kubenswrapper[4700]: I0227 17:16:03.980128 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcncs\" (UniqueName: \"kubernetes.io/projected/d1f9992d-c533-4184-9474-1cfe9b3e9819-kube-api-access-rcncs\") on node \"crc\" DevicePath \"\"" Feb 27 17:16:04 crc kubenswrapper[4700]: I0227 17:16:04.498788 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536876-pxhvd" event={"ID":"d1f9992d-c533-4184-9474-1cfe9b3e9819","Type":"ContainerDied","Data":"1aded29e16d8282dd37a01b03ffa070d41353431016adf4fcc2ccf85a33491d8"} Feb 27 17:16:04 crc kubenswrapper[4700]: I0227 17:16:04.499143 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1aded29e16d8282dd37a01b03ffa070d41353431016adf4fcc2ccf85a33491d8" Feb 27 17:16:04 crc kubenswrapper[4700]: I0227 17:16:04.498837 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536876-pxhvd" Feb 27 17:16:04 crc kubenswrapper[4700]: I0227 17:16:04.822691 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536870-ss2vg"] Feb 27 17:16:04 crc kubenswrapper[4700]: I0227 17:16:04.827321 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536870-ss2vg"] Feb 27 17:16:04 crc kubenswrapper[4700]: I0227 17:16:04.994093 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c8bc912-83a1-4ed4-8df7-ec5060b47700" path="/var/lib/kubelet/pods/1c8bc912-83a1-4ed4-8df7-ec5060b47700/volumes" Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:11.999776 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-cc5d6c7cf-x6qzc"] Feb 27 17:16:12 crc kubenswrapper[4700]: E0227 17:16:12.000444 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccc245e7-023f-4d49-99d0-9ebc992a9ccc" containerName="extract" Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.000461 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccc245e7-023f-4d49-99d0-9ebc992a9ccc" containerName="extract" Feb 27 17:16:12 crc kubenswrapper[4700]: E0227 17:16:12.000475 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccc245e7-023f-4d49-99d0-9ebc992a9ccc" containerName="pull" Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.000484 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccc245e7-023f-4d49-99d0-9ebc992a9ccc" containerName="pull" Feb 27 17:16:12 crc kubenswrapper[4700]: E0227 17:16:12.000516 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1f9992d-c533-4184-9474-1cfe9b3e9819" containerName="oc" Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.000526 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1f9992d-c533-4184-9474-1cfe9b3e9819" containerName="oc" Feb 27 17:16:12 crc kubenswrapper[4700]: E0227 17:16:12.000536 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccc245e7-023f-4d49-99d0-9ebc992a9ccc" containerName="util" Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.000543 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccc245e7-023f-4d49-99d0-9ebc992a9ccc" containerName="util" Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.000673 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1f9992d-c533-4184-9474-1cfe9b3e9819" containerName="oc" Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.000694 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccc245e7-023f-4d49-99d0-9ebc992a9ccc" containerName="extract" Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.001144 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-cc5d6c7cf-x6qzc" Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.005174 4700 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.005857 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.006328 4700 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.007293 4700 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-qvdft" Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.007508 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.024415 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-cc5d6c7cf-x6qzc"] Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.093364 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1d823d65-036b-4040-af8a-9b6698071d32-webhook-cert\") pod \"metallb-operator-controller-manager-cc5d6c7cf-x6qzc\" (UID: \"1d823d65-036b-4040-af8a-9b6698071d32\") " pod="metallb-system/metallb-operator-controller-manager-cc5d6c7cf-x6qzc" Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.093439 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1d823d65-036b-4040-af8a-9b6698071d32-apiservice-cert\") pod \"metallb-operator-controller-manager-cc5d6c7cf-x6qzc\" (UID: \"1d823d65-036b-4040-af8a-9b6698071d32\") " pod="metallb-system/metallb-operator-controller-manager-cc5d6c7cf-x6qzc" Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.093612 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nv7w7\" (UniqueName: \"kubernetes.io/projected/1d823d65-036b-4040-af8a-9b6698071d32-kube-api-access-nv7w7\") pod \"metallb-operator-controller-manager-cc5d6c7cf-x6qzc\" (UID: \"1d823d65-036b-4040-af8a-9b6698071d32\") " pod="metallb-system/metallb-operator-controller-manager-cc5d6c7cf-x6qzc" Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.194591 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nv7w7\" (UniqueName: \"kubernetes.io/projected/1d823d65-036b-4040-af8a-9b6698071d32-kube-api-access-nv7w7\") pod \"metallb-operator-controller-manager-cc5d6c7cf-x6qzc\" (UID: \"1d823d65-036b-4040-af8a-9b6698071d32\") " pod="metallb-system/metallb-operator-controller-manager-cc5d6c7cf-x6qzc" Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.194689 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1d823d65-036b-4040-af8a-9b6698071d32-webhook-cert\") pod \"metallb-operator-controller-manager-cc5d6c7cf-x6qzc\" (UID: \"1d823d65-036b-4040-af8a-9b6698071d32\") " pod="metallb-system/metallb-operator-controller-manager-cc5d6c7cf-x6qzc" Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.194716 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1d823d65-036b-4040-af8a-9b6698071d32-apiservice-cert\") pod \"metallb-operator-controller-manager-cc5d6c7cf-x6qzc\" (UID: \"1d823d65-036b-4040-af8a-9b6698071d32\") " pod="metallb-system/metallb-operator-controller-manager-cc5d6c7cf-x6qzc" Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.200857 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1d823d65-036b-4040-af8a-9b6698071d32-webhook-cert\") pod \"metallb-operator-controller-manager-cc5d6c7cf-x6qzc\" (UID: \"1d823d65-036b-4040-af8a-9b6698071d32\") " pod="metallb-system/metallb-operator-controller-manager-cc5d6c7cf-x6qzc" Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.203101 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1d823d65-036b-4040-af8a-9b6698071d32-apiservice-cert\") pod \"metallb-operator-controller-manager-cc5d6c7cf-x6qzc\" (UID: \"1d823d65-036b-4040-af8a-9b6698071d32\") " pod="metallb-system/metallb-operator-controller-manager-cc5d6c7cf-x6qzc" Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.216159 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nv7w7\" (UniqueName: \"kubernetes.io/projected/1d823d65-036b-4040-af8a-9b6698071d32-kube-api-access-nv7w7\") pod \"metallb-operator-controller-manager-cc5d6c7cf-x6qzc\" (UID: \"1d823d65-036b-4040-af8a-9b6698071d32\") " pod="metallb-system/metallb-operator-controller-manager-cc5d6c7cf-x6qzc" Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.316332 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-cc5d6c7cf-x6qzc" Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.445192 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-f4987db88-t66kg"] Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.445888 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-f4987db88-t66kg" Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.447859 4700 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.448045 4700 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-qw8zz" Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.448770 4700 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.456638 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-f4987db88-t66kg"] Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.614212 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/520ca54b-383f-4ca4-819a-7922a704404e-apiservice-cert\") pod \"metallb-operator-webhook-server-f4987db88-t66kg\" (UID: \"520ca54b-383f-4ca4-819a-7922a704404e\") " pod="metallb-system/metallb-operator-webhook-server-f4987db88-t66kg" Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.614467 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/520ca54b-383f-4ca4-819a-7922a704404e-webhook-cert\") pod \"metallb-operator-webhook-server-f4987db88-t66kg\" (UID: \"520ca54b-383f-4ca4-819a-7922a704404e\") " pod="metallb-system/metallb-operator-webhook-server-f4987db88-t66kg" Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.614593 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvd4b\" (UniqueName: \"kubernetes.io/projected/520ca54b-383f-4ca4-819a-7922a704404e-kube-api-access-lvd4b\") pod \"metallb-operator-webhook-server-f4987db88-t66kg\" (UID: \"520ca54b-383f-4ca4-819a-7922a704404e\") " pod="metallb-system/metallb-operator-webhook-server-f4987db88-t66kg" Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.699223 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-cc5d6c7cf-x6qzc"] Feb 27 17:16:12 crc kubenswrapper[4700]: W0227 17:16:12.706904 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d823d65_036b_4040_af8a_9b6698071d32.slice/crio-9c65808ac4be94d57a1c407c3c03cfc602b72bd68407d2ffb65fd982768a8991 WatchSource:0}: Error finding container 9c65808ac4be94d57a1c407c3c03cfc602b72bd68407d2ffb65fd982768a8991: Status 404 returned error can't find the container with id 9c65808ac4be94d57a1c407c3c03cfc602b72bd68407d2ffb65fd982768a8991 Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.715120 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvd4b\" (UniqueName: \"kubernetes.io/projected/520ca54b-383f-4ca4-819a-7922a704404e-kube-api-access-lvd4b\") pod \"metallb-operator-webhook-server-f4987db88-t66kg\" (UID: \"520ca54b-383f-4ca4-819a-7922a704404e\") " pod="metallb-system/metallb-operator-webhook-server-f4987db88-t66kg" Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.715157 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/520ca54b-383f-4ca4-819a-7922a704404e-apiservice-cert\") pod \"metallb-operator-webhook-server-f4987db88-t66kg\" (UID: \"520ca54b-383f-4ca4-819a-7922a704404e\") " pod="metallb-system/metallb-operator-webhook-server-f4987db88-t66kg" Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.715181 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/520ca54b-383f-4ca4-819a-7922a704404e-webhook-cert\") pod \"metallb-operator-webhook-server-f4987db88-t66kg\" (UID: \"520ca54b-383f-4ca4-819a-7922a704404e\") " pod="metallb-system/metallb-operator-webhook-server-f4987db88-t66kg" Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.719699 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/520ca54b-383f-4ca4-819a-7922a704404e-webhook-cert\") pod \"metallb-operator-webhook-server-f4987db88-t66kg\" (UID: \"520ca54b-383f-4ca4-819a-7922a704404e\") " pod="metallb-system/metallb-operator-webhook-server-f4987db88-t66kg" Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.732737 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvd4b\" (UniqueName: \"kubernetes.io/projected/520ca54b-383f-4ca4-819a-7922a704404e-kube-api-access-lvd4b\") pod \"metallb-operator-webhook-server-f4987db88-t66kg\" (UID: \"520ca54b-383f-4ca4-819a-7922a704404e\") " pod="metallb-system/metallb-operator-webhook-server-f4987db88-t66kg" Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.733476 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/520ca54b-383f-4ca4-819a-7922a704404e-apiservice-cert\") pod \"metallb-operator-webhook-server-f4987db88-t66kg\" (UID: \"520ca54b-383f-4ca4-819a-7922a704404e\") " pod="metallb-system/metallb-operator-webhook-server-f4987db88-t66kg" Feb 27 17:16:12 crc kubenswrapper[4700]: I0227 17:16:12.834308 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-f4987db88-t66kg" Feb 27 17:16:13 crc kubenswrapper[4700]: I0227 17:16:13.102032 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-f4987db88-t66kg"] Feb 27 17:16:13 crc kubenswrapper[4700]: W0227 17:16:13.108397 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod520ca54b_383f_4ca4_819a_7922a704404e.slice/crio-706856f28dd72b1dbb896751feb9192dd3440851852561d4cc240f29dbf2b9ad WatchSource:0}: Error finding container 706856f28dd72b1dbb896751feb9192dd3440851852561d4cc240f29dbf2b9ad: Status 404 returned error can't find the container with id 706856f28dd72b1dbb896751feb9192dd3440851852561d4cc240f29dbf2b9ad Feb 27 17:16:13 crc kubenswrapper[4700]: I0227 17:16:13.569214 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-f4987db88-t66kg" event={"ID":"520ca54b-383f-4ca4-819a-7922a704404e","Type":"ContainerStarted","Data":"706856f28dd72b1dbb896751feb9192dd3440851852561d4cc240f29dbf2b9ad"} Feb 27 17:16:13 crc kubenswrapper[4700]: I0227 17:16:13.570974 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-cc5d6c7cf-x6qzc" event={"ID":"1d823d65-036b-4040-af8a-9b6698071d32","Type":"ContainerStarted","Data":"9c65808ac4be94d57a1c407c3c03cfc602b72bd68407d2ffb65fd982768a8991"} Feb 27 17:16:16 crc kubenswrapper[4700]: I0227 17:16:16.589899 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-cc5d6c7cf-x6qzc" event={"ID":"1d823d65-036b-4040-af8a-9b6698071d32","Type":"ContainerStarted","Data":"fce40df3dc9d2d61893154a6aaf99d4b24e2acd4d1a0f9ed79a57e8f15f12f4e"} Feb 27 17:16:16 crc kubenswrapper[4700]: I0227 17:16:16.590456 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-cc5d6c7cf-x6qzc" Feb 27 17:16:16 crc kubenswrapper[4700]: I0227 17:16:16.613671 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-cc5d6c7cf-x6qzc" podStartSLOduration=1.973349594 podStartE2EDuration="5.613650026s" podCreationTimestamp="2026-02-27 17:16:11 +0000 UTC" firstStartedPulling="2026-02-27 17:16:12.70896053 +0000 UTC m=+932.694273277" lastFinishedPulling="2026-02-27 17:16:16.349260962 +0000 UTC m=+936.334573709" observedRunningTime="2026-02-27 17:16:16.609106095 +0000 UTC m=+936.594418852" watchObservedRunningTime="2026-02-27 17:16:16.613650026 +0000 UTC m=+936.598962773" Feb 27 17:16:18 crc kubenswrapper[4700]: I0227 17:16:18.604694 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-f4987db88-t66kg" event={"ID":"520ca54b-383f-4ca4-819a-7922a704404e","Type":"ContainerStarted","Data":"f3b112f527047f96a94274ce962e0282821965b96fdfe5dcc4df09718a5cb555"} Feb 27 17:16:18 crc kubenswrapper[4700]: I0227 17:16:18.605013 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-f4987db88-t66kg" Feb 27 17:16:18 crc kubenswrapper[4700]: I0227 17:16:18.627525 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-f4987db88-t66kg" podStartSLOduration=1.409277834 podStartE2EDuration="6.627503883s" podCreationTimestamp="2026-02-27 17:16:12 +0000 UTC" firstStartedPulling="2026-02-27 17:16:13.116363023 +0000 UTC m=+933.101675790" lastFinishedPulling="2026-02-27 17:16:18.334589092 +0000 UTC m=+938.319901839" observedRunningTime="2026-02-27 17:16:18.621549695 +0000 UTC m=+938.606862482" watchObservedRunningTime="2026-02-27 17:16:18.627503883 +0000 UTC m=+938.612816640" Feb 27 17:16:32 crc kubenswrapper[4700]: I0227 17:16:32.839596 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-f4987db88-t66kg" Feb 27 17:16:43 crc kubenswrapper[4700]: I0227 17:16:43.354719 4700 scope.go:117] "RemoveContainer" containerID="a3b8232566af1f16791722b65b5aacf80836bcde9541da4ef5ed04b439a21043" Feb 27 17:16:52 crc kubenswrapper[4700]: I0227 17:16:52.320425 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-cc5d6c7cf-x6qzc" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.109694 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-j65dj"] Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.111773 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-j65dj" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.113212 4700 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.113728 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.114044 4700 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-jglqd" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.123908 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7f989f654f-kf8ck"] Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.124789 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-kf8ck" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.129818 4700 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.147645 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/74b35197-b618-42e5-afba-88be4729f5b9-frr-sockets\") pod \"frr-k8s-j65dj\" (UID: \"74b35197-b618-42e5-afba-88be4729f5b9\") " pod="metallb-system/frr-k8s-j65dj" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.147681 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0aa251ee-47d2-4917-b18b-00ed276a6d24-cert\") pod \"frr-k8s-webhook-server-7f989f654f-kf8ck\" (UID: \"0aa251ee-47d2-4917-b18b-00ed276a6d24\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-kf8ck" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.147710 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zthpg\" (UniqueName: \"kubernetes.io/projected/74b35197-b618-42e5-afba-88be4729f5b9-kube-api-access-zthpg\") pod \"frr-k8s-j65dj\" (UID: \"74b35197-b618-42e5-afba-88be4729f5b9\") " pod="metallb-system/frr-k8s-j65dj" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.147734 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/74b35197-b618-42e5-afba-88be4729f5b9-metrics-certs\") pod \"frr-k8s-j65dj\" (UID: \"74b35197-b618-42e5-afba-88be4729f5b9\") " pod="metallb-system/frr-k8s-j65dj" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.147757 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/74b35197-b618-42e5-afba-88be4729f5b9-reloader\") pod \"frr-k8s-j65dj\" (UID: \"74b35197-b618-42e5-afba-88be4729f5b9\") " pod="metallb-system/frr-k8s-j65dj" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.147968 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/74b35197-b618-42e5-afba-88be4729f5b9-frr-startup\") pod \"frr-k8s-j65dj\" (UID: \"74b35197-b618-42e5-afba-88be4729f5b9\") " pod="metallb-system/frr-k8s-j65dj" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.148021 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hwp4\" (UniqueName: \"kubernetes.io/projected/0aa251ee-47d2-4917-b18b-00ed276a6d24-kube-api-access-6hwp4\") pod \"frr-k8s-webhook-server-7f989f654f-kf8ck\" (UID: \"0aa251ee-47d2-4917-b18b-00ed276a6d24\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-kf8ck" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.148073 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/74b35197-b618-42e5-afba-88be4729f5b9-frr-conf\") pod \"frr-k8s-j65dj\" (UID: \"74b35197-b618-42e5-afba-88be4729f5b9\") " pod="metallb-system/frr-k8s-j65dj" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.148220 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/74b35197-b618-42e5-afba-88be4729f5b9-metrics\") pod \"frr-k8s-j65dj\" (UID: \"74b35197-b618-42e5-afba-88be4729f5b9\") " pod="metallb-system/frr-k8s-j65dj" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.150745 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7f989f654f-kf8ck"] Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.215498 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-7ktps"] Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.216359 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-7ktps" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.221043 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.221341 4700 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-7vwrr" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.223331 4700 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.223516 4700 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.236695 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-86ddb6bd46-fvdl8"] Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.237520 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-86ddb6bd46-fvdl8" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.239965 4700 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.247846 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-86ddb6bd46-fvdl8"] Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.249433 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/74b35197-b618-42e5-afba-88be4729f5b9-frr-sockets\") pod \"frr-k8s-j65dj\" (UID: \"74b35197-b618-42e5-afba-88be4729f5b9\") " pod="metallb-system/frr-k8s-j65dj" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.249476 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0aa251ee-47d2-4917-b18b-00ed276a6d24-cert\") pod \"frr-k8s-webhook-server-7f989f654f-kf8ck\" (UID: \"0aa251ee-47d2-4917-b18b-00ed276a6d24\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-kf8ck" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.249506 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6108b3df-5dda-48cd-8cb4-55b36e5843aa-memberlist\") pod \"speaker-7ktps\" (UID: \"6108b3df-5dda-48cd-8cb4-55b36e5843aa\") " pod="metallb-system/speaker-7ktps" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.249527 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zthpg\" (UniqueName: \"kubernetes.io/projected/74b35197-b618-42e5-afba-88be4729f5b9-kube-api-access-zthpg\") pod \"frr-k8s-j65dj\" (UID: \"74b35197-b618-42e5-afba-88be4729f5b9\") " pod="metallb-system/frr-k8s-j65dj" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.249548 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/74b35197-b618-42e5-afba-88be4729f5b9-metrics-certs\") pod \"frr-k8s-j65dj\" (UID: \"74b35197-b618-42e5-afba-88be4729f5b9\") " pod="metallb-system/frr-k8s-j65dj" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.249571 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/74b35197-b618-42e5-afba-88be4729f5b9-reloader\") pod \"frr-k8s-j65dj\" (UID: \"74b35197-b618-42e5-afba-88be4729f5b9\") " pod="metallb-system/frr-k8s-j65dj" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.249589 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6108b3df-5dda-48cd-8cb4-55b36e5843aa-metrics-certs\") pod \"speaker-7ktps\" (UID: \"6108b3df-5dda-48cd-8cb4-55b36e5843aa\") " pod="metallb-system/speaker-7ktps" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.249615 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/6108b3df-5dda-48cd-8cb4-55b36e5843aa-metallb-excludel2\") pod \"speaker-7ktps\" (UID: \"6108b3df-5dda-48cd-8cb4-55b36e5843aa\") " pod="metallb-system/speaker-7ktps" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.249635 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/74b35197-b618-42e5-afba-88be4729f5b9-frr-startup\") pod \"frr-k8s-j65dj\" (UID: \"74b35197-b618-42e5-afba-88be4729f5b9\") " pod="metallb-system/frr-k8s-j65dj" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.249650 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hwp4\" (UniqueName: \"kubernetes.io/projected/0aa251ee-47d2-4917-b18b-00ed276a6d24-kube-api-access-6hwp4\") pod \"frr-k8s-webhook-server-7f989f654f-kf8ck\" (UID: \"0aa251ee-47d2-4917-b18b-00ed276a6d24\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-kf8ck" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.249673 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ks2rw\" (UniqueName: \"kubernetes.io/projected/6108b3df-5dda-48cd-8cb4-55b36e5843aa-kube-api-access-ks2rw\") pod \"speaker-7ktps\" (UID: \"6108b3df-5dda-48cd-8cb4-55b36e5843aa\") " pod="metallb-system/speaker-7ktps" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.249693 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/74b35197-b618-42e5-afba-88be4729f5b9-frr-conf\") pod \"frr-k8s-j65dj\" (UID: \"74b35197-b618-42e5-afba-88be4729f5b9\") " pod="metallb-system/frr-k8s-j65dj" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.249709 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4wct\" (UniqueName: \"kubernetes.io/projected/16cfcb53-eb3f-49d0-9cbc-f83b73527d8b-kube-api-access-v4wct\") pod \"controller-86ddb6bd46-fvdl8\" (UID: \"16cfcb53-eb3f-49d0-9cbc-f83b73527d8b\") " pod="metallb-system/controller-86ddb6bd46-fvdl8" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.249727 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/16cfcb53-eb3f-49d0-9cbc-f83b73527d8b-metrics-certs\") pod \"controller-86ddb6bd46-fvdl8\" (UID: \"16cfcb53-eb3f-49d0-9cbc-f83b73527d8b\") " pod="metallb-system/controller-86ddb6bd46-fvdl8" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.249746 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/16cfcb53-eb3f-49d0-9cbc-f83b73527d8b-cert\") pod \"controller-86ddb6bd46-fvdl8\" (UID: \"16cfcb53-eb3f-49d0-9cbc-f83b73527d8b\") " pod="metallb-system/controller-86ddb6bd46-fvdl8" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.249763 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/74b35197-b618-42e5-afba-88be4729f5b9-metrics\") pod \"frr-k8s-j65dj\" (UID: \"74b35197-b618-42e5-afba-88be4729f5b9\") " pod="metallb-system/frr-k8s-j65dj" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.250118 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/74b35197-b618-42e5-afba-88be4729f5b9-metrics\") pod \"frr-k8s-j65dj\" (UID: \"74b35197-b618-42e5-afba-88be4729f5b9\") " pod="metallb-system/frr-k8s-j65dj" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.250441 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/74b35197-b618-42e5-afba-88be4729f5b9-frr-sockets\") pod \"frr-k8s-j65dj\" (UID: \"74b35197-b618-42e5-afba-88be4729f5b9\") " pod="metallb-system/frr-k8s-j65dj" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.250499 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/74b35197-b618-42e5-afba-88be4729f5b9-reloader\") pod \"frr-k8s-j65dj\" (UID: \"74b35197-b618-42e5-afba-88be4729f5b9\") " pod="metallb-system/frr-k8s-j65dj" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.250846 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/74b35197-b618-42e5-afba-88be4729f5b9-frr-conf\") pod \"frr-k8s-j65dj\" (UID: \"74b35197-b618-42e5-afba-88be4729f5b9\") " pod="metallb-system/frr-k8s-j65dj" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.251866 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/74b35197-b618-42e5-afba-88be4729f5b9-frr-startup\") pod \"frr-k8s-j65dj\" (UID: \"74b35197-b618-42e5-afba-88be4729f5b9\") " pod="metallb-system/frr-k8s-j65dj" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.256902 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0aa251ee-47d2-4917-b18b-00ed276a6d24-cert\") pod \"frr-k8s-webhook-server-7f989f654f-kf8ck\" (UID: \"0aa251ee-47d2-4917-b18b-00ed276a6d24\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-kf8ck" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.274219 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/74b35197-b618-42e5-afba-88be4729f5b9-metrics-certs\") pod \"frr-k8s-j65dj\" (UID: \"74b35197-b618-42e5-afba-88be4729f5b9\") " pod="metallb-system/frr-k8s-j65dj" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.284152 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zthpg\" (UniqueName: \"kubernetes.io/projected/74b35197-b618-42e5-afba-88be4729f5b9-kube-api-access-zthpg\") pod \"frr-k8s-j65dj\" (UID: \"74b35197-b618-42e5-afba-88be4729f5b9\") " pod="metallb-system/frr-k8s-j65dj" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.287616 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hwp4\" (UniqueName: \"kubernetes.io/projected/0aa251ee-47d2-4917-b18b-00ed276a6d24-kube-api-access-6hwp4\") pod \"frr-k8s-webhook-server-7f989f654f-kf8ck\" (UID: \"0aa251ee-47d2-4917-b18b-00ed276a6d24\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-kf8ck" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.350732 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/16cfcb53-eb3f-49d0-9cbc-f83b73527d8b-metrics-certs\") pod \"controller-86ddb6bd46-fvdl8\" (UID: \"16cfcb53-eb3f-49d0-9cbc-f83b73527d8b\") " pod="metallb-system/controller-86ddb6bd46-fvdl8" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.350803 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/16cfcb53-eb3f-49d0-9cbc-f83b73527d8b-cert\") pod \"controller-86ddb6bd46-fvdl8\" (UID: \"16cfcb53-eb3f-49d0-9cbc-f83b73527d8b\") " pod="metallb-system/controller-86ddb6bd46-fvdl8" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.350862 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6108b3df-5dda-48cd-8cb4-55b36e5843aa-memberlist\") pod \"speaker-7ktps\" (UID: \"6108b3df-5dda-48cd-8cb4-55b36e5843aa\") " pod="metallb-system/speaker-7ktps" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.350908 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6108b3df-5dda-48cd-8cb4-55b36e5843aa-metrics-certs\") pod \"speaker-7ktps\" (UID: \"6108b3df-5dda-48cd-8cb4-55b36e5843aa\") " pod="metallb-system/speaker-7ktps" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.350939 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/6108b3df-5dda-48cd-8cb4-55b36e5843aa-metallb-excludel2\") pod \"speaker-7ktps\" (UID: \"6108b3df-5dda-48cd-8cb4-55b36e5843aa\") " pod="metallb-system/speaker-7ktps" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.350969 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ks2rw\" (UniqueName: \"kubernetes.io/projected/6108b3df-5dda-48cd-8cb4-55b36e5843aa-kube-api-access-ks2rw\") pod \"speaker-7ktps\" (UID: \"6108b3df-5dda-48cd-8cb4-55b36e5843aa\") " pod="metallb-system/speaker-7ktps" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.351000 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4wct\" (UniqueName: \"kubernetes.io/projected/16cfcb53-eb3f-49d0-9cbc-f83b73527d8b-kube-api-access-v4wct\") pod \"controller-86ddb6bd46-fvdl8\" (UID: \"16cfcb53-eb3f-49d0-9cbc-f83b73527d8b\") " pod="metallb-system/controller-86ddb6bd46-fvdl8" Feb 27 17:16:53 crc kubenswrapper[4700]: E0227 17:16:53.351098 4700 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Feb 27 17:16:53 crc kubenswrapper[4700]: E0227 17:16:53.351170 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6108b3df-5dda-48cd-8cb4-55b36e5843aa-metrics-certs podName:6108b3df-5dda-48cd-8cb4-55b36e5843aa nodeName:}" failed. No retries permitted until 2026-02-27 17:16:53.851151992 +0000 UTC m=+973.836464739 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6108b3df-5dda-48cd-8cb4-55b36e5843aa-metrics-certs") pod "speaker-7ktps" (UID: "6108b3df-5dda-48cd-8cb4-55b36e5843aa") : secret "speaker-certs-secret" not found Feb 27 17:16:53 crc kubenswrapper[4700]: E0227 17:16:53.351217 4700 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Feb 27 17:16:53 crc kubenswrapper[4700]: E0227 17:16:53.351236 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6108b3df-5dda-48cd-8cb4-55b36e5843aa-memberlist podName:6108b3df-5dda-48cd-8cb4-55b36e5843aa nodeName:}" failed. No retries permitted until 2026-02-27 17:16:53.851230184 +0000 UTC m=+973.836542931 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/6108b3df-5dda-48cd-8cb4-55b36e5843aa-memberlist") pod "speaker-7ktps" (UID: "6108b3df-5dda-48cd-8cb4-55b36e5843aa") : secret "metallb-memberlist" not found Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.351699 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/6108b3df-5dda-48cd-8cb4-55b36e5843aa-metallb-excludel2\") pod \"speaker-7ktps\" (UID: \"6108b3df-5dda-48cd-8cb4-55b36e5843aa\") " pod="metallb-system/speaker-7ktps" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.353578 4700 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.355166 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/16cfcb53-eb3f-49d0-9cbc-f83b73527d8b-metrics-certs\") pod \"controller-86ddb6bd46-fvdl8\" (UID: \"16cfcb53-eb3f-49d0-9cbc-f83b73527d8b\") " pod="metallb-system/controller-86ddb6bd46-fvdl8" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.366527 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/16cfcb53-eb3f-49d0-9cbc-f83b73527d8b-cert\") pod \"controller-86ddb6bd46-fvdl8\" (UID: \"16cfcb53-eb3f-49d0-9cbc-f83b73527d8b\") " pod="metallb-system/controller-86ddb6bd46-fvdl8" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.368174 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4wct\" (UniqueName: \"kubernetes.io/projected/16cfcb53-eb3f-49d0-9cbc-f83b73527d8b-kube-api-access-v4wct\") pod \"controller-86ddb6bd46-fvdl8\" (UID: \"16cfcb53-eb3f-49d0-9cbc-f83b73527d8b\") " pod="metallb-system/controller-86ddb6bd46-fvdl8" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.370963 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ks2rw\" (UniqueName: \"kubernetes.io/projected/6108b3df-5dda-48cd-8cb4-55b36e5843aa-kube-api-access-ks2rw\") pod \"speaker-7ktps\" (UID: \"6108b3df-5dda-48cd-8cb4-55b36e5843aa\") " pod="metallb-system/speaker-7ktps" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.431288 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-j65dj" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.441026 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-kf8ck" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.550982 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-86ddb6bd46-fvdl8" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.665020 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7f989f654f-kf8ck"] Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.754686 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-86ddb6bd46-fvdl8"] Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.855723 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-86ddb6bd46-fvdl8" event={"ID":"16cfcb53-eb3f-49d0-9cbc-f83b73527d8b","Type":"ContainerStarted","Data":"b623ecaefb5455fdb22da1e6d28c17ebf6e05e47edb6ff6b71325a32d85dea3c"} Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.857887 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6108b3df-5dda-48cd-8cb4-55b36e5843aa-memberlist\") pod \"speaker-7ktps\" (UID: \"6108b3df-5dda-48cd-8cb4-55b36e5843aa\") " pod="metallb-system/speaker-7ktps" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.857987 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6108b3df-5dda-48cd-8cb4-55b36e5843aa-metrics-certs\") pod \"speaker-7ktps\" (UID: \"6108b3df-5dda-48cd-8cb4-55b36e5843aa\") " pod="metallb-system/speaker-7ktps" Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.858926 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-kf8ck" event={"ID":"0aa251ee-47d2-4917-b18b-00ed276a6d24","Type":"ContainerStarted","Data":"ca9f84112b768a6a6b79912ce2e83d18462c81aae95607c80e64bfc79cd9cd20"} Feb 27 17:16:53 crc kubenswrapper[4700]: E0227 17:16:53.858974 4700 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Feb 27 17:16:53 crc kubenswrapper[4700]: E0227 17:16:53.859217 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6108b3df-5dda-48cd-8cb4-55b36e5843aa-memberlist podName:6108b3df-5dda-48cd-8cb4-55b36e5843aa nodeName:}" failed. No retries permitted until 2026-02-27 17:16:54.859147627 +0000 UTC m=+974.844460414 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/6108b3df-5dda-48cd-8cb4-55b36e5843aa-memberlist") pod "speaker-7ktps" (UID: "6108b3df-5dda-48cd-8cb4-55b36e5843aa") : secret "metallb-memberlist" not found Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.861152 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j65dj" event={"ID":"74b35197-b618-42e5-afba-88be4729f5b9","Type":"ContainerStarted","Data":"b60560a3bbacb3521869cab9f2c71551dc03afec5f71897e0df2fa1aaafc2a08"} Feb 27 17:16:53 crc kubenswrapper[4700]: I0227 17:16:53.866279 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6108b3df-5dda-48cd-8cb4-55b36e5843aa-metrics-certs\") pod \"speaker-7ktps\" (UID: \"6108b3df-5dda-48cd-8cb4-55b36e5843aa\") " pod="metallb-system/speaker-7ktps" Feb 27 17:16:54 crc kubenswrapper[4700]: I0227 17:16:54.871363 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-86ddb6bd46-fvdl8" event={"ID":"16cfcb53-eb3f-49d0-9cbc-f83b73527d8b","Type":"ContainerStarted","Data":"3bdcd68cf2659e4561c03efaed6c3ffaac969e3195179751ff50e93f3ddf9d0f"} Feb 27 17:16:54 crc kubenswrapper[4700]: I0227 17:16:54.871935 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-86ddb6bd46-fvdl8" event={"ID":"16cfcb53-eb3f-49d0-9cbc-f83b73527d8b","Type":"ContainerStarted","Data":"8e1eeb043c8b7494fcd8af1ad6faf161a404959a23424dff23f99d7f21f38dbf"} Feb 27 17:16:54 crc kubenswrapper[4700]: I0227 17:16:54.871979 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-86ddb6bd46-fvdl8" Feb 27 17:16:54 crc kubenswrapper[4700]: I0227 17:16:54.877009 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6108b3df-5dda-48cd-8cb4-55b36e5843aa-memberlist\") pod \"speaker-7ktps\" (UID: \"6108b3df-5dda-48cd-8cb4-55b36e5843aa\") " pod="metallb-system/speaker-7ktps" Feb 27 17:16:54 crc kubenswrapper[4700]: I0227 17:16:54.884395 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6108b3df-5dda-48cd-8cb4-55b36e5843aa-memberlist\") pod \"speaker-7ktps\" (UID: \"6108b3df-5dda-48cd-8cb4-55b36e5843aa\") " pod="metallb-system/speaker-7ktps" Feb 27 17:16:55 crc kubenswrapper[4700]: I0227 17:16:55.028579 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-7ktps" Feb 27 17:16:55 crc kubenswrapper[4700]: W0227 17:16:55.061109 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6108b3df_5dda_48cd_8cb4_55b36e5843aa.slice/crio-d74b4349d3c0d239162fb7f3698cb36adf54cf76a615024365e081a77ff726e0 WatchSource:0}: Error finding container d74b4349d3c0d239162fb7f3698cb36adf54cf76a615024365e081a77ff726e0: Status 404 returned error can't find the container with id d74b4349d3c0d239162fb7f3698cb36adf54cf76a615024365e081a77ff726e0 Feb 27 17:16:55 crc kubenswrapper[4700]: I0227 17:16:55.878569 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-7ktps" event={"ID":"6108b3df-5dda-48cd-8cb4-55b36e5843aa","Type":"ContainerStarted","Data":"7a599e2d0fd3e3465bb9eeb2da44f591b9054905e088cc98f98f7bf9c8a0157c"} Feb 27 17:16:55 crc kubenswrapper[4700]: I0227 17:16:55.879049 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-7ktps" event={"ID":"6108b3df-5dda-48cd-8cb4-55b36e5843aa","Type":"ContainerStarted","Data":"251377b140470e2e95b9a6206f7bcbd2f1155ec8123c6bc6e682cd2b8d537e71"} Feb 27 17:16:55 crc kubenswrapper[4700]: I0227 17:16:55.879062 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-7ktps" event={"ID":"6108b3df-5dda-48cd-8cb4-55b36e5843aa","Type":"ContainerStarted","Data":"d74b4349d3c0d239162fb7f3698cb36adf54cf76a615024365e081a77ff726e0"} Feb 27 17:16:55 crc kubenswrapper[4700]: I0227 17:16:55.879210 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-7ktps" Feb 27 17:16:55 crc kubenswrapper[4700]: I0227 17:16:55.901274 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-7ktps" podStartSLOduration=2.901259234 podStartE2EDuration="2.901259234s" podCreationTimestamp="2026-02-27 17:16:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:16:55.897570976 +0000 UTC m=+975.882883723" watchObservedRunningTime="2026-02-27 17:16:55.901259234 +0000 UTC m=+975.886571981" Feb 27 17:16:55 crc kubenswrapper[4700]: I0227 17:16:55.902345 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-86ddb6bd46-fvdl8" podStartSLOduration=2.902339253 podStartE2EDuration="2.902339253s" podCreationTimestamp="2026-02-27 17:16:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:16:54.898920477 +0000 UTC m=+974.884233234" watchObservedRunningTime="2026-02-27 17:16:55.902339253 +0000 UTC m=+975.887652000" Feb 27 17:17:01 crc kubenswrapper[4700]: I0227 17:17:01.921150 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-kf8ck" event={"ID":"0aa251ee-47d2-4917-b18b-00ed276a6d24","Type":"ContainerStarted","Data":"da618bad3ba8a5f0c2c0c63ad2f698a2786bf92d85100ca8d2f32805945eb14c"} Feb 27 17:17:01 crc kubenswrapper[4700]: I0227 17:17:01.921654 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-kf8ck" Feb 27 17:17:01 crc kubenswrapper[4700]: I0227 17:17:01.922664 4700 generic.go:334] "Generic (PLEG): container finished" podID="74b35197-b618-42e5-afba-88be4729f5b9" containerID="6a841a178dd2d8c6ced76a6e01a5f3e1f7e4e23d666810abf5aadf806f3e929f" exitCode=0 Feb 27 17:17:01 crc kubenswrapper[4700]: I0227 17:17:01.922698 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j65dj" event={"ID":"74b35197-b618-42e5-afba-88be4729f5b9","Type":"ContainerDied","Data":"6a841a178dd2d8c6ced76a6e01a5f3e1f7e4e23d666810abf5aadf806f3e929f"} Feb 27 17:17:01 crc kubenswrapper[4700]: I0227 17:17:01.938791 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-kf8ck" podStartSLOduration=0.913948393 podStartE2EDuration="8.938773606s" podCreationTimestamp="2026-02-27 17:16:53 +0000 UTC" firstStartedPulling="2026-02-27 17:16:53.674762219 +0000 UTC m=+973.660074986" lastFinishedPulling="2026-02-27 17:17:01.699587452 +0000 UTC m=+981.684900199" observedRunningTime="2026-02-27 17:17:01.93477747 +0000 UTC m=+981.920090257" watchObservedRunningTime="2026-02-27 17:17:01.938773606 +0000 UTC m=+981.924086353" Feb 27 17:17:02 crc kubenswrapper[4700]: I0227 17:17:02.951354 4700 generic.go:334] "Generic (PLEG): container finished" podID="74b35197-b618-42e5-afba-88be4729f5b9" containerID="05ef13100a11bfa3ef011b2bdfd79fabb3f902ffc649e3a09c7cebb831163bd3" exitCode=0 Feb 27 17:17:02 crc kubenswrapper[4700]: I0227 17:17:02.951420 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j65dj" event={"ID":"74b35197-b618-42e5-afba-88be4729f5b9","Type":"ContainerDied","Data":"05ef13100a11bfa3ef011b2bdfd79fabb3f902ffc649e3a09c7cebb831163bd3"} Feb 27 17:17:03 crc kubenswrapper[4700]: I0227 17:17:03.555640 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-86ddb6bd46-fvdl8" Feb 27 17:17:03 crc kubenswrapper[4700]: I0227 17:17:03.960207 4700 generic.go:334] "Generic (PLEG): container finished" podID="74b35197-b618-42e5-afba-88be4729f5b9" containerID="a49ffa8987194af0fa071b53b64fddf0d7b2a6ade3cf3ee74b21f3f9d91d60b3" exitCode=0 Feb 27 17:17:03 crc kubenswrapper[4700]: I0227 17:17:03.960258 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j65dj" event={"ID":"74b35197-b618-42e5-afba-88be4729f5b9","Type":"ContainerDied","Data":"a49ffa8987194af0fa071b53b64fddf0d7b2a6ade3cf3ee74b21f3f9d91d60b3"} Feb 27 17:17:04 crc kubenswrapper[4700]: I0227 17:17:04.973104 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j65dj" event={"ID":"74b35197-b618-42e5-afba-88be4729f5b9","Type":"ContainerStarted","Data":"1bb792a7fdc817764e2a645fa5d1acca5632e783f2918f695011b945b376ed5a"} Feb 27 17:17:04 crc kubenswrapper[4700]: I0227 17:17:04.973447 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j65dj" event={"ID":"74b35197-b618-42e5-afba-88be4729f5b9","Type":"ContainerStarted","Data":"b9217ba7c324d642783d999e708f6e4347eb900cc9b72fcf5da8bdbdec57f3d8"} Feb 27 17:17:04 crc kubenswrapper[4700]: I0227 17:17:04.973490 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j65dj" event={"ID":"74b35197-b618-42e5-afba-88be4729f5b9","Type":"ContainerStarted","Data":"0f69a1ee35fcfe30717ef872195971ae53e2f4c193f423dcbf51a2ab85837bb7"} Feb 27 17:17:04 crc kubenswrapper[4700]: I0227 17:17:04.973503 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j65dj" event={"ID":"74b35197-b618-42e5-afba-88be4729f5b9","Type":"ContainerStarted","Data":"798fa96a538ec5b615b738d9e2fceb7da1d77cfbc26fd63eda7221b24643cd50"} Feb 27 17:17:05 crc kubenswrapper[4700]: I0227 17:17:05.032319 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-7ktps" Feb 27 17:17:05 crc kubenswrapper[4700]: I0227 17:17:05.985062 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j65dj" event={"ID":"74b35197-b618-42e5-afba-88be4729f5b9","Type":"ContainerStarted","Data":"bc2f2bdc23ca0a1ca9487e98a80d3068f6ca36c00a94171a303c6f476da95404"} Feb 27 17:17:05 crc kubenswrapper[4700]: I0227 17:17:05.985134 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j65dj" event={"ID":"74b35197-b618-42e5-afba-88be4729f5b9","Type":"ContainerStarted","Data":"17fe933a194e7bf7807b2c14540506594044e25c899bf33a96d6df50fcf89b63"} Feb 27 17:17:05 crc kubenswrapper[4700]: I0227 17:17:05.985317 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-j65dj" Feb 27 17:17:06 crc kubenswrapper[4700]: I0227 17:17:06.011667 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-j65dj" podStartSLOduration=4.906563694 podStartE2EDuration="13.011650959s" podCreationTimestamp="2026-02-27 17:16:53 +0000 UTC" firstStartedPulling="2026-02-27 17:16:53.570595342 +0000 UTC m=+973.555908109" lastFinishedPulling="2026-02-27 17:17:01.675682627 +0000 UTC m=+981.660995374" observedRunningTime="2026-02-27 17:17:06.006825391 +0000 UTC m=+985.992138138" watchObservedRunningTime="2026-02-27 17:17:06.011650959 +0000 UTC m=+985.996963706" Feb 27 17:17:06 crc kubenswrapper[4700]: I0227 17:17:06.410953 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:17:06 crc kubenswrapper[4700]: I0227 17:17:06.411286 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:17:08 crc kubenswrapper[4700]: I0227 17:17:08.065114 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-625fx"] Feb 27 17:17:08 crc kubenswrapper[4700]: I0227 17:17:08.065830 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-625fx" Feb 27 17:17:08 crc kubenswrapper[4700]: I0227 17:17:08.068286 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Feb 27 17:17:08 crc kubenswrapper[4700]: I0227 17:17:08.068379 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Feb 27 17:17:08 crc kubenswrapper[4700]: I0227 17:17:08.068315 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-5khwf" Feb 27 17:17:08 crc kubenswrapper[4700]: I0227 17:17:08.077890 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-625fx"] Feb 27 17:17:08 crc kubenswrapper[4700]: I0227 17:17:08.131207 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6t29\" (UniqueName: \"kubernetes.io/projected/c6c88872-336d-4953-902b-14fc2bae54ec-kube-api-access-b6t29\") pod \"openstack-operator-index-625fx\" (UID: \"c6c88872-336d-4953-902b-14fc2bae54ec\") " pod="openstack-operators/openstack-operator-index-625fx" Feb 27 17:17:08 crc kubenswrapper[4700]: I0227 17:17:08.232965 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6t29\" (UniqueName: \"kubernetes.io/projected/c6c88872-336d-4953-902b-14fc2bae54ec-kube-api-access-b6t29\") pod \"openstack-operator-index-625fx\" (UID: \"c6c88872-336d-4953-902b-14fc2bae54ec\") " pod="openstack-operators/openstack-operator-index-625fx" Feb 27 17:17:08 crc kubenswrapper[4700]: I0227 17:17:08.255564 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6t29\" (UniqueName: \"kubernetes.io/projected/c6c88872-336d-4953-902b-14fc2bae54ec-kube-api-access-b6t29\") pod \"openstack-operator-index-625fx\" (UID: \"c6c88872-336d-4953-902b-14fc2bae54ec\") " pod="openstack-operators/openstack-operator-index-625fx" Feb 27 17:17:08 crc kubenswrapper[4700]: I0227 17:17:08.385898 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-625fx" Feb 27 17:17:08 crc kubenswrapper[4700]: I0227 17:17:08.431832 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-j65dj" Feb 27 17:17:08 crc kubenswrapper[4700]: I0227 17:17:08.493082 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-j65dj" Feb 27 17:17:08 crc kubenswrapper[4700]: I0227 17:17:08.690146 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-625fx"] Feb 27 17:17:08 crc kubenswrapper[4700]: W0227 17:17:08.695665 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc6c88872_336d_4953_902b_14fc2bae54ec.slice/crio-da295d833b715c62b0cb0b0e64d7da82d98739b6ebb5b82b627ad56edec1b372 WatchSource:0}: Error finding container da295d833b715c62b0cb0b0e64d7da82d98739b6ebb5b82b627ad56edec1b372: Status 404 returned error can't find the container with id da295d833b715c62b0cb0b0e64d7da82d98739b6ebb5b82b627ad56edec1b372 Feb 27 17:17:09 crc kubenswrapper[4700]: I0227 17:17:09.019032 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-625fx" event={"ID":"c6c88872-336d-4953-902b-14fc2bae54ec","Type":"ContainerStarted","Data":"da295d833b715c62b0cb0b0e64d7da82d98739b6ebb5b82b627ad56edec1b372"} Feb 27 17:17:11 crc kubenswrapper[4700]: I0227 17:17:11.249401 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-625fx"] Feb 27 17:17:11 crc kubenswrapper[4700]: I0227 17:17:11.851890 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-tdw6d"] Feb 27 17:17:11 crc kubenswrapper[4700]: I0227 17:17:11.853606 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-tdw6d" Feb 27 17:17:11 crc kubenswrapper[4700]: I0227 17:17:11.864851 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-tdw6d"] Feb 27 17:17:11 crc kubenswrapper[4700]: I0227 17:17:11.885859 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85bl8\" (UniqueName: \"kubernetes.io/projected/16e196e0-ad2c-4e3e-8056-d15d71661a96-kube-api-access-85bl8\") pod \"openstack-operator-index-tdw6d\" (UID: \"16e196e0-ad2c-4e3e-8056-d15d71661a96\") " pod="openstack-operators/openstack-operator-index-tdw6d" Feb 27 17:17:11 crc kubenswrapper[4700]: I0227 17:17:11.987738 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85bl8\" (UniqueName: \"kubernetes.io/projected/16e196e0-ad2c-4e3e-8056-d15d71661a96-kube-api-access-85bl8\") pod \"openstack-operator-index-tdw6d\" (UID: \"16e196e0-ad2c-4e3e-8056-d15d71661a96\") " pod="openstack-operators/openstack-operator-index-tdw6d" Feb 27 17:17:12 crc kubenswrapper[4700]: I0227 17:17:12.010726 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85bl8\" (UniqueName: \"kubernetes.io/projected/16e196e0-ad2c-4e3e-8056-d15d71661a96-kube-api-access-85bl8\") pod \"openstack-operator-index-tdw6d\" (UID: \"16e196e0-ad2c-4e3e-8056-d15d71661a96\") " pod="openstack-operators/openstack-operator-index-tdw6d" Feb 27 17:17:12 crc kubenswrapper[4700]: I0227 17:17:12.057063 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-625fx" event={"ID":"c6c88872-336d-4953-902b-14fc2bae54ec","Type":"ContainerStarted","Data":"4a0feab881b455f73dff6c146461f20aafd2cbd156fc6c315aabb879bd272c69"} Feb 27 17:17:12 crc kubenswrapper[4700]: I0227 17:17:12.057207 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-625fx" podUID="c6c88872-336d-4953-902b-14fc2bae54ec" containerName="registry-server" containerID="cri-o://4a0feab881b455f73dff6c146461f20aafd2cbd156fc6c315aabb879bd272c69" gracePeriod=2 Feb 27 17:17:12 crc kubenswrapper[4700]: I0227 17:17:12.086349 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-625fx" podStartSLOduration=1.701903339 podStartE2EDuration="4.086322149s" podCreationTimestamp="2026-02-27 17:17:08 +0000 UTC" firstStartedPulling="2026-02-27 17:17:08.696950573 +0000 UTC m=+988.682263320" lastFinishedPulling="2026-02-27 17:17:11.081369373 +0000 UTC m=+991.066682130" observedRunningTime="2026-02-27 17:17:12.076267922 +0000 UTC m=+992.061580689" watchObservedRunningTime="2026-02-27 17:17:12.086322149 +0000 UTC m=+992.071634936" Feb 27 17:17:12 crc kubenswrapper[4700]: I0227 17:17:12.173108 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-tdw6d" Feb 27 17:17:12 crc kubenswrapper[4700]: I0227 17:17:12.497013 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-tdw6d"] Feb 27 17:17:12 crc kubenswrapper[4700]: W0227 17:17:12.509126 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod16e196e0_ad2c_4e3e_8056_d15d71661a96.slice/crio-e2dee4c03b215553a75697a012e318db41c53cae6efc3260b03ac3d37133ff91 WatchSource:0}: Error finding container e2dee4c03b215553a75697a012e318db41c53cae6efc3260b03ac3d37133ff91: Status 404 returned error can't find the container with id e2dee4c03b215553a75697a012e318db41c53cae6efc3260b03ac3d37133ff91 Feb 27 17:17:12 crc kubenswrapper[4700]: I0227 17:17:12.541843 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-625fx" Feb 27 17:17:12 crc kubenswrapper[4700]: I0227 17:17:12.595720 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6t29\" (UniqueName: \"kubernetes.io/projected/c6c88872-336d-4953-902b-14fc2bae54ec-kube-api-access-b6t29\") pod \"c6c88872-336d-4953-902b-14fc2bae54ec\" (UID: \"c6c88872-336d-4953-902b-14fc2bae54ec\") " Feb 27 17:17:12 crc kubenswrapper[4700]: I0227 17:17:12.600388 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6c88872-336d-4953-902b-14fc2bae54ec-kube-api-access-b6t29" (OuterVolumeSpecName: "kube-api-access-b6t29") pod "c6c88872-336d-4953-902b-14fc2bae54ec" (UID: "c6c88872-336d-4953-902b-14fc2bae54ec"). InnerVolumeSpecName "kube-api-access-b6t29". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:17:12 crc kubenswrapper[4700]: I0227 17:17:12.698371 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6t29\" (UniqueName: \"kubernetes.io/projected/c6c88872-336d-4953-902b-14fc2bae54ec-kube-api-access-b6t29\") on node \"crc\" DevicePath \"\"" Feb 27 17:17:13 crc kubenswrapper[4700]: I0227 17:17:13.069999 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-tdw6d" event={"ID":"16e196e0-ad2c-4e3e-8056-d15d71661a96","Type":"ContainerStarted","Data":"c145b1a027992f5991e752a375d3a1e6a28f79ed9af6a2f6cf313bb0c87d8476"} Feb 27 17:17:13 crc kubenswrapper[4700]: I0227 17:17:13.070437 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-tdw6d" event={"ID":"16e196e0-ad2c-4e3e-8056-d15d71661a96","Type":"ContainerStarted","Data":"e2dee4c03b215553a75697a012e318db41c53cae6efc3260b03ac3d37133ff91"} Feb 27 17:17:13 crc kubenswrapper[4700]: I0227 17:17:13.073518 4700 generic.go:334] "Generic (PLEG): container finished" podID="c6c88872-336d-4953-902b-14fc2bae54ec" containerID="4a0feab881b455f73dff6c146461f20aafd2cbd156fc6c315aabb879bd272c69" exitCode=0 Feb 27 17:17:13 crc kubenswrapper[4700]: I0227 17:17:13.073586 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-625fx" event={"ID":"c6c88872-336d-4953-902b-14fc2bae54ec","Type":"ContainerDied","Data":"4a0feab881b455f73dff6c146461f20aafd2cbd156fc6c315aabb879bd272c69"} Feb 27 17:17:13 crc kubenswrapper[4700]: I0227 17:17:13.073653 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-625fx" event={"ID":"c6c88872-336d-4953-902b-14fc2bae54ec","Type":"ContainerDied","Data":"da295d833b715c62b0cb0b0e64d7da82d98739b6ebb5b82b627ad56edec1b372"} Feb 27 17:17:13 crc kubenswrapper[4700]: I0227 17:17:13.073683 4700 scope.go:117] "RemoveContainer" containerID="4a0feab881b455f73dff6c146461f20aafd2cbd156fc6c315aabb879bd272c69" Feb 27 17:17:13 crc kubenswrapper[4700]: I0227 17:17:13.073584 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-625fx" Feb 27 17:17:13 crc kubenswrapper[4700]: I0227 17:17:13.110197 4700 scope.go:117] "RemoveContainer" containerID="4a0feab881b455f73dff6c146461f20aafd2cbd156fc6c315aabb879bd272c69" Feb 27 17:17:13 crc kubenswrapper[4700]: I0227 17:17:13.110834 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-tdw6d" podStartSLOduration=2.051113487 podStartE2EDuration="2.110793533s" podCreationTimestamp="2026-02-27 17:17:11 +0000 UTC" firstStartedPulling="2026-02-27 17:17:12.51292422 +0000 UTC m=+992.498236967" lastFinishedPulling="2026-02-27 17:17:12.572604256 +0000 UTC m=+992.557917013" observedRunningTime="2026-02-27 17:17:13.10053704 +0000 UTC m=+993.085849847" watchObservedRunningTime="2026-02-27 17:17:13.110793533 +0000 UTC m=+993.096106320" Feb 27 17:17:13 crc kubenswrapper[4700]: E0227 17:17:13.120156 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a0feab881b455f73dff6c146461f20aafd2cbd156fc6c315aabb879bd272c69\": container with ID starting with 4a0feab881b455f73dff6c146461f20aafd2cbd156fc6c315aabb879bd272c69 not found: ID does not exist" containerID="4a0feab881b455f73dff6c146461f20aafd2cbd156fc6c315aabb879bd272c69" Feb 27 17:17:13 crc kubenswrapper[4700]: I0227 17:17:13.120236 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a0feab881b455f73dff6c146461f20aafd2cbd156fc6c315aabb879bd272c69"} err="failed to get container status \"4a0feab881b455f73dff6c146461f20aafd2cbd156fc6c315aabb879bd272c69\": rpc error: code = NotFound desc = could not find container \"4a0feab881b455f73dff6c146461f20aafd2cbd156fc6c315aabb879bd272c69\": container with ID starting with 4a0feab881b455f73dff6c146461f20aafd2cbd156fc6c315aabb879bd272c69 not found: ID does not exist" Feb 27 17:17:13 crc kubenswrapper[4700]: I0227 17:17:13.130832 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-625fx"] Feb 27 17:17:13 crc kubenswrapper[4700]: I0227 17:17:13.139033 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-625fx"] Feb 27 17:17:13 crc kubenswrapper[4700]: I0227 17:17:13.454757 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-kf8ck" Feb 27 17:17:14 crc kubenswrapper[4700]: I0227 17:17:14.993426 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6c88872-336d-4953-902b-14fc2bae54ec" path="/var/lib/kubelet/pods/c6c88872-336d-4953-902b-14fc2bae54ec/volumes" Feb 27 17:17:22 crc kubenswrapper[4700]: I0227 17:17:22.174199 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-tdw6d" Feb 27 17:17:22 crc kubenswrapper[4700]: I0227 17:17:22.174730 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-tdw6d" Feb 27 17:17:22 crc kubenswrapper[4700]: I0227 17:17:22.215868 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-tdw6d" Feb 27 17:17:23 crc kubenswrapper[4700]: I0227 17:17:23.182689 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-tdw6d" Feb 27 17:17:23 crc kubenswrapper[4700]: I0227 17:17:23.435707 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-j65dj" Feb 27 17:17:28 crc kubenswrapper[4700]: I0227 17:17:28.705963 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/a70a7d17df4824bd764910ad293b109e72da56f2f4bf07c2a6401f9135j2sd8"] Feb 27 17:17:28 crc kubenswrapper[4700]: E0227 17:17:28.706906 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6c88872-336d-4953-902b-14fc2bae54ec" containerName="registry-server" Feb 27 17:17:28 crc kubenswrapper[4700]: I0227 17:17:28.706928 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6c88872-336d-4953-902b-14fc2bae54ec" containerName="registry-server" Feb 27 17:17:28 crc kubenswrapper[4700]: I0227 17:17:28.707138 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6c88872-336d-4953-902b-14fc2bae54ec" containerName="registry-server" Feb 27 17:17:28 crc kubenswrapper[4700]: I0227 17:17:28.708605 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a70a7d17df4824bd764910ad293b109e72da56f2f4bf07c2a6401f9135j2sd8" Feb 27 17:17:28 crc kubenswrapper[4700]: I0227 17:17:28.713911 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/a70a7d17df4824bd764910ad293b109e72da56f2f4bf07c2a6401f9135j2sd8"] Feb 27 17:17:28 crc kubenswrapper[4700]: I0227 17:17:28.715940 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-d24nb" Feb 27 17:17:28 crc kubenswrapper[4700]: I0227 17:17:28.837999 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/670f4e8a-0015-44af-8397-87e1ef84c3e0-bundle\") pod \"a70a7d17df4824bd764910ad293b109e72da56f2f4bf07c2a6401f9135j2sd8\" (UID: \"670f4e8a-0015-44af-8397-87e1ef84c3e0\") " pod="openstack-operators/a70a7d17df4824bd764910ad293b109e72da56f2f4bf07c2a6401f9135j2sd8" Feb 27 17:17:28 crc kubenswrapper[4700]: I0227 17:17:28.838243 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8p69\" (UniqueName: \"kubernetes.io/projected/670f4e8a-0015-44af-8397-87e1ef84c3e0-kube-api-access-s8p69\") pod \"a70a7d17df4824bd764910ad293b109e72da56f2f4bf07c2a6401f9135j2sd8\" (UID: \"670f4e8a-0015-44af-8397-87e1ef84c3e0\") " pod="openstack-operators/a70a7d17df4824bd764910ad293b109e72da56f2f4bf07c2a6401f9135j2sd8" Feb 27 17:17:28 crc kubenswrapper[4700]: I0227 17:17:28.838416 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/670f4e8a-0015-44af-8397-87e1ef84c3e0-util\") pod \"a70a7d17df4824bd764910ad293b109e72da56f2f4bf07c2a6401f9135j2sd8\" (UID: \"670f4e8a-0015-44af-8397-87e1ef84c3e0\") " pod="openstack-operators/a70a7d17df4824bd764910ad293b109e72da56f2f4bf07c2a6401f9135j2sd8" Feb 27 17:17:28 crc kubenswrapper[4700]: I0227 17:17:28.939688 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/670f4e8a-0015-44af-8397-87e1ef84c3e0-bundle\") pod \"a70a7d17df4824bd764910ad293b109e72da56f2f4bf07c2a6401f9135j2sd8\" (UID: \"670f4e8a-0015-44af-8397-87e1ef84c3e0\") " pod="openstack-operators/a70a7d17df4824bd764910ad293b109e72da56f2f4bf07c2a6401f9135j2sd8" Feb 27 17:17:28 crc kubenswrapper[4700]: I0227 17:17:28.939815 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8p69\" (UniqueName: \"kubernetes.io/projected/670f4e8a-0015-44af-8397-87e1ef84c3e0-kube-api-access-s8p69\") pod \"a70a7d17df4824bd764910ad293b109e72da56f2f4bf07c2a6401f9135j2sd8\" (UID: \"670f4e8a-0015-44af-8397-87e1ef84c3e0\") " pod="openstack-operators/a70a7d17df4824bd764910ad293b109e72da56f2f4bf07c2a6401f9135j2sd8" Feb 27 17:17:28 crc kubenswrapper[4700]: I0227 17:17:28.939857 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/670f4e8a-0015-44af-8397-87e1ef84c3e0-util\") pod \"a70a7d17df4824bd764910ad293b109e72da56f2f4bf07c2a6401f9135j2sd8\" (UID: \"670f4e8a-0015-44af-8397-87e1ef84c3e0\") " pod="openstack-operators/a70a7d17df4824bd764910ad293b109e72da56f2f4bf07c2a6401f9135j2sd8" Feb 27 17:17:28 crc kubenswrapper[4700]: I0227 17:17:28.940270 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/670f4e8a-0015-44af-8397-87e1ef84c3e0-bundle\") pod \"a70a7d17df4824bd764910ad293b109e72da56f2f4bf07c2a6401f9135j2sd8\" (UID: \"670f4e8a-0015-44af-8397-87e1ef84c3e0\") " pod="openstack-operators/a70a7d17df4824bd764910ad293b109e72da56f2f4bf07c2a6401f9135j2sd8" Feb 27 17:17:28 crc kubenswrapper[4700]: I0227 17:17:28.940683 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/670f4e8a-0015-44af-8397-87e1ef84c3e0-util\") pod \"a70a7d17df4824bd764910ad293b109e72da56f2f4bf07c2a6401f9135j2sd8\" (UID: \"670f4e8a-0015-44af-8397-87e1ef84c3e0\") " pod="openstack-operators/a70a7d17df4824bd764910ad293b109e72da56f2f4bf07c2a6401f9135j2sd8" Feb 27 17:17:28 crc kubenswrapper[4700]: I0227 17:17:28.963720 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8p69\" (UniqueName: \"kubernetes.io/projected/670f4e8a-0015-44af-8397-87e1ef84c3e0-kube-api-access-s8p69\") pod \"a70a7d17df4824bd764910ad293b109e72da56f2f4bf07c2a6401f9135j2sd8\" (UID: \"670f4e8a-0015-44af-8397-87e1ef84c3e0\") " pod="openstack-operators/a70a7d17df4824bd764910ad293b109e72da56f2f4bf07c2a6401f9135j2sd8" Feb 27 17:17:29 crc kubenswrapper[4700]: I0227 17:17:29.037296 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a70a7d17df4824bd764910ad293b109e72da56f2f4bf07c2a6401f9135j2sd8" Feb 27 17:17:29 crc kubenswrapper[4700]: I0227 17:17:29.519183 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/a70a7d17df4824bd764910ad293b109e72da56f2f4bf07c2a6401f9135j2sd8"] Feb 27 17:17:29 crc kubenswrapper[4700]: W0227 17:17:29.527917 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod670f4e8a_0015_44af_8397_87e1ef84c3e0.slice/crio-a63565f11f349c0f43b6d2d919da0fc4dd9fb36bd7eb488cddfe2e17606bc3b2 WatchSource:0}: Error finding container a63565f11f349c0f43b6d2d919da0fc4dd9fb36bd7eb488cddfe2e17606bc3b2: Status 404 returned error can't find the container with id a63565f11f349c0f43b6d2d919da0fc4dd9fb36bd7eb488cddfe2e17606bc3b2 Feb 27 17:17:30 crc kubenswrapper[4700]: I0227 17:17:30.222821 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a70a7d17df4824bd764910ad293b109e72da56f2f4bf07c2a6401f9135j2sd8" event={"ID":"670f4e8a-0015-44af-8397-87e1ef84c3e0","Type":"ContainerStarted","Data":"a63565f11f349c0f43b6d2d919da0fc4dd9fb36bd7eb488cddfe2e17606bc3b2"} Feb 27 17:17:31 crc kubenswrapper[4700]: I0227 17:17:31.230908 4700 generic.go:334] "Generic (PLEG): container finished" podID="670f4e8a-0015-44af-8397-87e1ef84c3e0" containerID="554c9ab7608d9d2f6d1d155c01f3121b58461f108a4819b5fd891516c648e286" exitCode=0 Feb 27 17:17:31 crc kubenswrapper[4700]: I0227 17:17:31.231008 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a70a7d17df4824bd764910ad293b109e72da56f2f4bf07c2a6401f9135j2sd8" event={"ID":"670f4e8a-0015-44af-8397-87e1ef84c3e0","Type":"ContainerDied","Data":"554c9ab7608d9d2f6d1d155c01f3121b58461f108a4819b5fd891516c648e286"} Feb 27 17:17:32 crc kubenswrapper[4700]: I0227 17:17:32.244150 4700 generic.go:334] "Generic (PLEG): container finished" podID="670f4e8a-0015-44af-8397-87e1ef84c3e0" containerID="8839a068a176c8875c4f4ce564d3e7380a87fb38b84120c356e7932737178fb8" exitCode=0 Feb 27 17:17:32 crc kubenswrapper[4700]: I0227 17:17:32.244253 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a70a7d17df4824bd764910ad293b109e72da56f2f4bf07c2a6401f9135j2sd8" event={"ID":"670f4e8a-0015-44af-8397-87e1ef84c3e0","Type":"ContainerDied","Data":"8839a068a176c8875c4f4ce564d3e7380a87fb38b84120c356e7932737178fb8"} Feb 27 17:17:33 crc kubenswrapper[4700]: I0227 17:17:33.255670 4700 generic.go:334] "Generic (PLEG): container finished" podID="670f4e8a-0015-44af-8397-87e1ef84c3e0" containerID="0c72c0429030ef38df3a44a58b490a29507f2aa3bd0a4df919fb4c13ae4e12a0" exitCode=0 Feb 27 17:17:33 crc kubenswrapper[4700]: I0227 17:17:33.255733 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a70a7d17df4824bd764910ad293b109e72da56f2f4bf07c2a6401f9135j2sd8" event={"ID":"670f4e8a-0015-44af-8397-87e1ef84c3e0","Type":"ContainerDied","Data":"0c72c0429030ef38df3a44a58b490a29507f2aa3bd0a4df919fb4c13ae4e12a0"} Feb 27 17:17:34 crc kubenswrapper[4700]: I0227 17:17:34.597605 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a70a7d17df4824bd764910ad293b109e72da56f2f4bf07c2a6401f9135j2sd8" Feb 27 17:17:34 crc kubenswrapper[4700]: I0227 17:17:34.725377 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8p69\" (UniqueName: \"kubernetes.io/projected/670f4e8a-0015-44af-8397-87e1ef84c3e0-kube-api-access-s8p69\") pod \"670f4e8a-0015-44af-8397-87e1ef84c3e0\" (UID: \"670f4e8a-0015-44af-8397-87e1ef84c3e0\") " Feb 27 17:17:34 crc kubenswrapper[4700]: I0227 17:17:34.725618 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/670f4e8a-0015-44af-8397-87e1ef84c3e0-bundle\") pod \"670f4e8a-0015-44af-8397-87e1ef84c3e0\" (UID: \"670f4e8a-0015-44af-8397-87e1ef84c3e0\") " Feb 27 17:17:34 crc kubenswrapper[4700]: I0227 17:17:34.725714 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/670f4e8a-0015-44af-8397-87e1ef84c3e0-util\") pod \"670f4e8a-0015-44af-8397-87e1ef84c3e0\" (UID: \"670f4e8a-0015-44af-8397-87e1ef84c3e0\") " Feb 27 17:17:34 crc kubenswrapper[4700]: I0227 17:17:34.726779 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/670f4e8a-0015-44af-8397-87e1ef84c3e0-bundle" (OuterVolumeSpecName: "bundle") pod "670f4e8a-0015-44af-8397-87e1ef84c3e0" (UID: "670f4e8a-0015-44af-8397-87e1ef84c3e0"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:17:34 crc kubenswrapper[4700]: I0227 17:17:34.735420 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/670f4e8a-0015-44af-8397-87e1ef84c3e0-kube-api-access-s8p69" (OuterVolumeSpecName: "kube-api-access-s8p69") pod "670f4e8a-0015-44af-8397-87e1ef84c3e0" (UID: "670f4e8a-0015-44af-8397-87e1ef84c3e0"). InnerVolumeSpecName "kube-api-access-s8p69". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:17:34 crc kubenswrapper[4700]: I0227 17:17:34.745563 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/670f4e8a-0015-44af-8397-87e1ef84c3e0-util" (OuterVolumeSpecName: "util") pod "670f4e8a-0015-44af-8397-87e1ef84c3e0" (UID: "670f4e8a-0015-44af-8397-87e1ef84c3e0"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:17:34 crc kubenswrapper[4700]: I0227 17:17:34.828125 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8p69\" (UniqueName: \"kubernetes.io/projected/670f4e8a-0015-44af-8397-87e1ef84c3e0-kube-api-access-s8p69\") on node \"crc\" DevicePath \"\"" Feb 27 17:17:34 crc kubenswrapper[4700]: I0227 17:17:34.828175 4700 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/670f4e8a-0015-44af-8397-87e1ef84c3e0-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:17:34 crc kubenswrapper[4700]: I0227 17:17:34.828194 4700 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/670f4e8a-0015-44af-8397-87e1ef84c3e0-util\") on node \"crc\" DevicePath \"\"" Feb 27 17:17:35 crc kubenswrapper[4700]: I0227 17:17:35.274716 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a70a7d17df4824bd764910ad293b109e72da56f2f4bf07c2a6401f9135j2sd8" event={"ID":"670f4e8a-0015-44af-8397-87e1ef84c3e0","Type":"ContainerDied","Data":"a63565f11f349c0f43b6d2d919da0fc4dd9fb36bd7eb488cddfe2e17606bc3b2"} Feb 27 17:17:35 crc kubenswrapper[4700]: I0227 17:17:35.275311 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a63565f11f349c0f43b6d2d919da0fc4dd9fb36bd7eb488cddfe2e17606bc3b2" Feb 27 17:17:35 crc kubenswrapper[4700]: I0227 17:17:35.274807 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a70a7d17df4824bd764910ad293b109e72da56f2f4bf07c2a6401f9135j2sd8" Feb 27 17:17:36 crc kubenswrapper[4700]: I0227 17:17:36.410635 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:17:36 crc kubenswrapper[4700]: I0227 17:17:36.410727 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:17:40 crc kubenswrapper[4700]: I0227 17:17:40.889266 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-64d4474cb4-mb4nb"] Feb 27 17:17:40 crc kubenswrapper[4700]: E0227 17:17:40.889879 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="670f4e8a-0015-44af-8397-87e1ef84c3e0" containerName="extract" Feb 27 17:17:40 crc kubenswrapper[4700]: I0227 17:17:40.889894 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="670f4e8a-0015-44af-8397-87e1ef84c3e0" containerName="extract" Feb 27 17:17:40 crc kubenswrapper[4700]: E0227 17:17:40.889911 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="670f4e8a-0015-44af-8397-87e1ef84c3e0" containerName="util" Feb 27 17:17:40 crc kubenswrapper[4700]: I0227 17:17:40.889919 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="670f4e8a-0015-44af-8397-87e1ef84c3e0" containerName="util" Feb 27 17:17:40 crc kubenswrapper[4700]: E0227 17:17:40.889935 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="670f4e8a-0015-44af-8397-87e1ef84c3e0" containerName="pull" Feb 27 17:17:40 crc kubenswrapper[4700]: I0227 17:17:40.889943 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="670f4e8a-0015-44af-8397-87e1ef84c3e0" containerName="pull" Feb 27 17:17:40 crc kubenswrapper[4700]: I0227 17:17:40.890088 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="670f4e8a-0015-44af-8397-87e1ef84c3e0" containerName="extract" Feb 27 17:17:40 crc kubenswrapper[4700]: I0227 17:17:40.890590 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-64d4474cb4-mb4nb" Feb 27 17:17:40 crc kubenswrapper[4700]: I0227 17:17:40.896878 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-9g49s" Feb 27 17:17:40 crc kubenswrapper[4700]: I0227 17:17:40.918652 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-64d4474cb4-mb4nb"] Feb 27 17:17:41 crc kubenswrapper[4700]: I0227 17:17:41.020730 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rldv9\" (UniqueName: \"kubernetes.io/projected/d65f20c2-a9a2-43a1-9adf-9925736c38c5-kube-api-access-rldv9\") pod \"openstack-operator-controller-init-64d4474cb4-mb4nb\" (UID: \"d65f20c2-a9a2-43a1-9adf-9925736c38c5\") " pod="openstack-operators/openstack-operator-controller-init-64d4474cb4-mb4nb" Feb 27 17:17:41 crc kubenswrapper[4700]: I0227 17:17:41.121910 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rldv9\" (UniqueName: \"kubernetes.io/projected/d65f20c2-a9a2-43a1-9adf-9925736c38c5-kube-api-access-rldv9\") pod \"openstack-operator-controller-init-64d4474cb4-mb4nb\" (UID: \"d65f20c2-a9a2-43a1-9adf-9925736c38c5\") " pod="openstack-operators/openstack-operator-controller-init-64d4474cb4-mb4nb" Feb 27 17:17:41 crc kubenswrapper[4700]: I0227 17:17:41.138540 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rldv9\" (UniqueName: \"kubernetes.io/projected/d65f20c2-a9a2-43a1-9adf-9925736c38c5-kube-api-access-rldv9\") pod \"openstack-operator-controller-init-64d4474cb4-mb4nb\" (UID: \"d65f20c2-a9a2-43a1-9adf-9925736c38c5\") " pod="openstack-operators/openstack-operator-controller-init-64d4474cb4-mb4nb" Feb 27 17:17:41 crc kubenswrapper[4700]: I0227 17:17:41.209201 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-64d4474cb4-mb4nb" Feb 27 17:17:41 crc kubenswrapper[4700]: I0227 17:17:41.427153 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-64d4474cb4-mb4nb"] Feb 27 17:17:41 crc kubenswrapper[4700]: W0227 17:17:41.443697 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd65f20c2_a9a2_43a1_9adf_9925736c38c5.slice/crio-1b1195f7d465b3978317d899ed60707689894cfd05ff0dfb91507f27352b8aa6 WatchSource:0}: Error finding container 1b1195f7d465b3978317d899ed60707689894cfd05ff0dfb91507f27352b8aa6: Status 404 returned error can't find the container with id 1b1195f7d465b3978317d899ed60707689894cfd05ff0dfb91507f27352b8aa6 Feb 27 17:17:42 crc kubenswrapper[4700]: I0227 17:17:42.349883 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-64d4474cb4-mb4nb" event={"ID":"d65f20c2-a9a2-43a1-9adf-9925736c38c5","Type":"ContainerStarted","Data":"1b1195f7d465b3978317d899ed60707689894cfd05ff0dfb91507f27352b8aa6"} Feb 27 17:17:46 crc kubenswrapper[4700]: I0227 17:17:46.396613 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-64d4474cb4-mb4nb" event={"ID":"d65f20c2-a9a2-43a1-9adf-9925736c38c5","Type":"ContainerStarted","Data":"ae9698cc796ec64e591d7e8925a2d3a081ade9ec503289cba420d2e23605f209"} Feb 27 17:17:46 crc kubenswrapper[4700]: I0227 17:17:46.397267 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-64d4474cb4-mb4nb" Feb 27 17:17:46 crc kubenswrapper[4700]: I0227 17:17:46.432816 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-64d4474cb4-mb4nb" podStartSLOduration=2.070792695 podStartE2EDuration="6.432799649s" podCreationTimestamp="2026-02-27 17:17:40 +0000 UTC" firstStartedPulling="2026-02-27 17:17:41.445351631 +0000 UTC m=+1021.430664378" lastFinishedPulling="2026-02-27 17:17:45.807358585 +0000 UTC m=+1025.792671332" observedRunningTime="2026-02-27 17:17:46.430285732 +0000 UTC m=+1026.415598509" watchObservedRunningTime="2026-02-27 17:17:46.432799649 +0000 UTC m=+1026.418112436" Feb 27 17:17:51 crc kubenswrapper[4700]: I0227 17:17:51.213166 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-64d4474cb4-mb4nb" Feb 27 17:17:57 crc kubenswrapper[4700]: I0227 17:17:57.761406 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qmhw5"] Feb 27 17:17:57 crc kubenswrapper[4700]: I0227 17:17:57.763055 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qmhw5" Feb 27 17:17:57 crc kubenswrapper[4700]: I0227 17:17:57.797656 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qmhw5"] Feb 27 17:17:57 crc kubenswrapper[4700]: I0227 17:17:57.885594 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tgdp\" (UniqueName: \"kubernetes.io/projected/b1a1f4b1-e13f-4b11-bdeb-5d815af7b98b-kube-api-access-5tgdp\") pod \"certified-operators-qmhw5\" (UID: \"b1a1f4b1-e13f-4b11-bdeb-5d815af7b98b\") " pod="openshift-marketplace/certified-operators-qmhw5" Feb 27 17:17:57 crc kubenswrapper[4700]: I0227 17:17:57.885654 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1a1f4b1-e13f-4b11-bdeb-5d815af7b98b-catalog-content\") pod \"certified-operators-qmhw5\" (UID: \"b1a1f4b1-e13f-4b11-bdeb-5d815af7b98b\") " pod="openshift-marketplace/certified-operators-qmhw5" Feb 27 17:17:57 crc kubenswrapper[4700]: I0227 17:17:57.885768 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1a1f4b1-e13f-4b11-bdeb-5d815af7b98b-utilities\") pod \"certified-operators-qmhw5\" (UID: \"b1a1f4b1-e13f-4b11-bdeb-5d815af7b98b\") " pod="openshift-marketplace/certified-operators-qmhw5" Feb 27 17:17:57 crc kubenswrapper[4700]: I0227 17:17:57.987272 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1a1f4b1-e13f-4b11-bdeb-5d815af7b98b-utilities\") pod \"certified-operators-qmhw5\" (UID: \"b1a1f4b1-e13f-4b11-bdeb-5d815af7b98b\") " pod="openshift-marketplace/certified-operators-qmhw5" Feb 27 17:17:57 crc kubenswrapper[4700]: I0227 17:17:57.987349 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tgdp\" (UniqueName: \"kubernetes.io/projected/b1a1f4b1-e13f-4b11-bdeb-5d815af7b98b-kube-api-access-5tgdp\") pod \"certified-operators-qmhw5\" (UID: \"b1a1f4b1-e13f-4b11-bdeb-5d815af7b98b\") " pod="openshift-marketplace/certified-operators-qmhw5" Feb 27 17:17:57 crc kubenswrapper[4700]: I0227 17:17:57.987385 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1a1f4b1-e13f-4b11-bdeb-5d815af7b98b-catalog-content\") pod \"certified-operators-qmhw5\" (UID: \"b1a1f4b1-e13f-4b11-bdeb-5d815af7b98b\") " pod="openshift-marketplace/certified-operators-qmhw5" Feb 27 17:17:57 crc kubenswrapper[4700]: I0227 17:17:57.987822 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1a1f4b1-e13f-4b11-bdeb-5d815af7b98b-utilities\") pod \"certified-operators-qmhw5\" (UID: \"b1a1f4b1-e13f-4b11-bdeb-5d815af7b98b\") " pod="openshift-marketplace/certified-operators-qmhw5" Feb 27 17:17:57 crc kubenswrapper[4700]: I0227 17:17:57.987856 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1a1f4b1-e13f-4b11-bdeb-5d815af7b98b-catalog-content\") pod \"certified-operators-qmhw5\" (UID: \"b1a1f4b1-e13f-4b11-bdeb-5d815af7b98b\") " pod="openshift-marketplace/certified-operators-qmhw5" Feb 27 17:17:58 crc kubenswrapper[4700]: I0227 17:17:58.008279 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tgdp\" (UniqueName: \"kubernetes.io/projected/b1a1f4b1-e13f-4b11-bdeb-5d815af7b98b-kube-api-access-5tgdp\") pod \"certified-operators-qmhw5\" (UID: \"b1a1f4b1-e13f-4b11-bdeb-5d815af7b98b\") " pod="openshift-marketplace/certified-operators-qmhw5" Feb 27 17:17:58 crc kubenswrapper[4700]: I0227 17:17:58.084334 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qmhw5" Feb 27 17:17:58 crc kubenswrapper[4700]: I0227 17:17:58.433847 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qmhw5"] Feb 27 17:17:58 crc kubenswrapper[4700]: I0227 17:17:58.485878 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qmhw5" event={"ID":"b1a1f4b1-e13f-4b11-bdeb-5d815af7b98b","Type":"ContainerStarted","Data":"f659f73e41cb0f024bf84950088e494fb977081ddb14ca8e507fd638f679cf2a"} Feb 27 17:17:59 crc kubenswrapper[4700]: I0227 17:17:59.493354 4700 generic.go:334] "Generic (PLEG): container finished" podID="b1a1f4b1-e13f-4b11-bdeb-5d815af7b98b" containerID="fd07cbbd3663957d4d7627f2432007a3c65c91a537f5d584d39879b9a33a01cc" exitCode=0 Feb 27 17:17:59 crc kubenswrapper[4700]: I0227 17:17:59.493417 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qmhw5" event={"ID":"b1a1f4b1-e13f-4b11-bdeb-5d815af7b98b","Type":"ContainerDied","Data":"fd07cbbd3663957d4d7627f2432007a3c65c91a537f5d584d39879b9a33a01cc"} Feb 27 17:18:00 crc kubenswrapper[4700]: I0227 17:18:00.104051 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wwl7g"] Feb 27 17:18:00 crc kubenswrapper[4700]: I0227 17:18:00.105504 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wwl7g" Feb 27 17:18:00 crc kubenswrapper[4700]: I0227 17:18:00.115870 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wwl7g"] Feb 27 17:18:00 crc kubenswrapper[4700]: I0227 17:18:00.169617 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536878-zp9lh"] Feb 27 17:18:00 crc kubenswrapper[4700]: I0227 17:18:00.171111 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536878-zp9lh" Feb 27 17:18:00 crc kubenswrapper[4700]: I0227 17:18:00.173844 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:18:00 crc kubenswrapper[4700]: I0227 17:18:00.174158 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:18:00 crc kubenswrapper[4700]: I0227 17:18:00.174385 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d62kc" Feb 27 17:18:00 crc kubenswrapper[4700]: I0227 17:18:00.176791 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536878-zp9lh"] Feb 27 17:18:00 crc kubenswrapper[4700]: I0227 17:18:00.217565 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddpll\" (UniqueName: \"kubernetes.io/projected/e548373a-d9bf-46bd-8efa-b4adf2dedc4b-kube-api-access-ddpll\") pod \"redhat-marketplace-wwl7g\" (UID: \"e548373a-d9bf-46bd-8efa-b4adf2dedc4b\") " pod="openshift-marketplace/redhat-marketplace-wwl7g" Feb 27 17:18:00 crc kubenswrapper[4700]: I0227 17:18:00.217644 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e548373a-d9bf-46bd-8efa-b4adf2dedc4b-utilities\") pod \"redhat-marketplace-wwl7g\" (UID: \"e548373a-d9bf-46bd-8efa-b4adf2dedc4b\") " pod="openshift-marketplace/redhat-marketplace-wwl7g" Feb 27 17:18:00 crc kubenswrapper[4700]: I0227 17:18:00.217676 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2z4d\" (UniqueName: \"kubernetes.io/projected/7765df9b-4c2e-4ac9-b862-4c300d69acb5-kube-api-access-j2z4d\") pod \"auto-csr-approver-29536878-zp9lh\" (UID: \"7765df9b-4c2e-4ac9-b862-4c300d69acb5\") " pod="openshift-infra/auto-csr-approver-29536878-zp9lh" Feb 27 17:18:00 crc kubenswrapper[4700]: I0227 17:18:00.217719 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e548373a-d9bf-46bd-8efa-b4adf2dedc4b-catalog-content\") pod \"redhat-marketplace-wwl7g\" (UID: \"e548373a-d9bf-46bd-8efa-b4adf2dedc4b\") " pod="openshift-marketplace/redhat-marketplace-wwl7g" Feb 27 17:18:00 crc kubenswrapper[4700]: I0227 17:18:00.319176 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddpll\" (UniqueName: \"kubernetes.io/projected/e548373a-d9bf-46bd-8efa-b4adf2dedc4b-kube-api-access-ddpll\") pod \"redhat-marketplace-wwl7g\" (UID: \"e548373a-d9bf-46bd-8efa-b4adf2dedc4b\") " pod="openshift-marketplace/redhat-marketplace-wwl7g" Feb 27 17:18:00 crc kubenswrapper[4700]: I0227 17:18:00.319516 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e548373a-d9bf-46bd-8efa-b4adf2dedc4b-utilities\") pod \"redhat-marketplace-wwl7g\" (UID: \"e548373a-d9bf-46bd-8efa-b4adf2dedc4b\") " pod="openshift-marketplace/redhat-marketplace-wwl7g" Feb 27 17:18:00 crc kubenswrapper[4700]: I0227 17:18:00.319639 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2z4d\" (UniqueName: \"kubernetes.io/projected/7765df9b-4c2e-4ac9-b862-4c300d69acb5-kube-api-access-j2z4d\") pod \"auto-csr-approver-29536878-zp9lh\" (UID: \"7765df9b-4c2e-4ac9-b862-4c300d69acb5\") " pod="openshift-infra/auto-csr-approver-29536878-zp9lh" Feb 27 17:18:00 crc kubenswrapper[4700]: I0227 17:18:00.319750 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e548373a-d9bf-46bd-8efa-b4adf2dedc4b-catalog-content\") pod \"redhat-marketplace-wwl7g\" (UID: \"e548373a-d9bf-46bd-8efa-b4adf2dedc4b\") " pod="openshift-marketplace/redhat-marketplace-wwl7g" Feb 27 17:18:00 crc kubenswrapper[4700]: I0227 17:18:00.319913 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e548373a-d9bf-46bd-8efa-b4adf2dedc4b-utilities\") pod \"redhat-marketplace-wwl7g\" (UID: \"e548373a-d9bf-46bd-8efa-b4adf2dedc4b\") " pod="openshift-marketplace/redhat-marketplace-wwl7g" Feb 27 17:18:00 crc kubenswrapper[4700]: I0227 17:18:00.320121 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e548373a-d9bf-46bd-8efa-b4adf2dedc4b-catalog-content\") pod \"redhat-marketplace-wwl7g\" (UID: \"e548373a-d9bf-46bd-8efa-b4adf2dedc4b\") " pod="openshift-marketplace/redhat-marketplace-wwl7g" Feb 27 17:18:00 crc kubenswrapper[4700]: I0227 17:18:00.341843 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddpll\" (UniqueName: \"kubernetes.io/projected/e548373a-d9bf-46bd-8efa-b4adf2dedc4b-kube-api-access-ddpll\") pod \"redhat-marketplace-wwl7g\" (UID: \"e548373a-d9bf-46bd-8efa-b4adf2dedc4b\") " pod="openshift-marketplace/redhat-marketplace-wwl7g" Feb 27 17:18:00 crc kubenswrapper[4700]: I0227 17:18:00.343188 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2z4d\" (UniqueName: \"kubernetes.io/projected/7765df9b-4c2e-4ac9-b862-4c300d69acb5-kube-api-access-j2z4d\") pod \"auto-csr-approver-29536878-zp9lh\" (UID: \"7765df9b-4c2e-4ac9-b862-4c300d69acb5\") " pod="openshift-infra/auto-csr-approver-29536878-zp9lh" Feb 27 17:18:00 crc kubenswrapper[4700]: I0227 17:18:00.423484 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wwl7g" Feb 27 17:18:00 crc kubenswrapper[4700]: I0227 17:18:00.485418 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536878-zp9lh" Feb 27 17:18:00 crc kubenswrapper[4700]: I0227 17:18:00.902214 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wwl7g"] Feb 27 17:18:00 crc kubenswrapper[4700]: I0227 17:18:00.961179 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536878-zp9lh"] Feb 27 17:18:00 crc kubenswrapper[4700]: W0227 17:18:00.970532 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7765df9b_4c2e_4ac9_b862_4c300d69acb5.slice/crio-8d30a6f98833e22430d334595f7d2269f7eb9ee712b992f79df5def0b6273216 WatchSource:0}: Error finding container 8d30a6f98833e22430d334595f7d2269f7eb9ee712b992f79df5def0b6273216: Status 404 returned error can't find the container with id 8d30a6f98833e22430d334595f7d2269f7eb9ee712b992f79df5def0b6273216 Feb 27 17:18:01 crc kubenswrapper[4700]: I0227 17:18:01.512726 4700 generic.go:334] "Generic (PLEG): container finished" podID="e548373a-d9bf-46bd-8efa-b4adf2dedc4b" containerID="84ba22f971daed6b03ac4db6cac8d8ed639abe496d8deb4ca197bde236d699b6" exitCode=0 Feb 27 17:18:01 crc kubenswrapper[4700]: I0227 17:18:01.512834 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wwl7g" event={"ID":"e548373a-d9bf-46bd-8efa-b4adf2dedc4b","Type":"ContainerDied","Data":"84ba22f971daed6b03ac4db6cac8d8ed639abe496d8deb4ca197bde236d699b6"} Feb 27 17:18:01 crc kubenswrapper[4700]: I0227 17:18:01.513104 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wwl7g" event={"ID":"e548373a-d9bf-46bd-8efa-b4adf2dedc4b","Type":"ContainerStarted","Data":"e240c8fabd13b1b1dad1ae559c41c98d2ab7f927d48e1f1204cbcc97f5308a60"} Feb 27 17:18:01 crc kubenswrapper[4700]: I0227 17:18:01.515881 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536878-zp9lh" event={"ID":"7765df9b-4c2e-4ac9-b862-4c300d69acb5","Type":"ContainerStarted","Data":"8d30a6f98833e22430d334595f7d2269f7eb9ee712b992f79df5def0b6273216"} Feb 27 17:18:03 crc kubenswrapper[4700]: I0227 17:18:03.546578 4700 generic.go:334] "Generic (PLEG): container finished" podID="e548373a-d9bf-46bd-8efa-b4adf2dedc4b" containerID="d216a732af5a93b29dbe0c142cc4ac56b9a23e406c28e7b3db1c40ba7c4ca2d2" exitCode=0 Feb 27 17:18:03 crc kubenswrapper[4700]: I0227 17:18:03.546664 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wwl7g" event={"ID":"e548373a-d9bf-46bd-8efa-b4adf2dedc4b","Type":"ContainerDied","Data":"d216a732af5a93b29dbe0c142cc4ac56b9a23e406c28e7b3db1c40ba7c4ca2d2"} Feb 27 17:18:06 crc kubenswrapper[4700]: I0227 17:18:06.410319 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:18:06 crc kubenswrapper[4700]: I0227 17:18:06.410591 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:18:06 crc kubenswrapper[4700]: I0227 17:18:06.410637 4700 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" Feb 27 17:18:06 crc kubenswrapper[4700]: I0227 17:18:06.411266 4700 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6c88e3b04f19e6791ce8b218282ad3546f21efb28cfed9c5676d0da60d8b308b"} pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 17:18:06 crc kubenswrapper[4700]: I0227 17:18:06.411315 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" containerID="cri-o://6c88e3b04f19e6791ce8b218282ad3546f21efb28cfed9c5676d0da60d8b308b" gracePeriod=600 Feb 27 17:18:06 crc kubenswrapper[4700]: I0227 17:18:06.595316 4700 generic.go:334] "Generic (PLEG): container finished" podID="84b45511-b94c-479f-98db-fd2c4eceec46" containerID="6c88e3b04f19e6791ce8b218282ad3546f21efb28cfed9c5676d0da60d8b308b" exitCode=0 Feb 27 17:18:06 crc kubenswrapper[4700]: I0227 17:18:06.595387 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerDied","Data":"6c88e3b04f19e6791ce8b218282ad3546f21efb28cfed9c5676d0da60d8b308b"} Feb 27 17:18:06 crc kubenswrapper[4700]: I0227 17:18:06.595787 4700 scope.go:117] "RemoveContainer" containerID="510d6e176ab30bcaf35acfa6f85490773219083ba09de332e6a13153619b2fac" Feb 27 17:18:08 crc kubenswrapper[4700]: I0227 17:18:08.614789 4700 generic.go:334] "Generic (PLEG): container finished" podID="7765df9b-4c2e-4ac9-b862-4c300d69acb5" containerID="a92bd1eb5131a729ffcb950c63978cef03db723ace5353e6cb1430d02d09d7e9" exitCode=0 Feb 27 17:18:08 crc kubenswrapper[4700]: I0227 17:18:08.614836 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536878-zp9lh" event={"ID":"7765df9b-4c2e-4ac9-b862-4c300d69acb5","Type":"ContainerDied","Data":"a92bd1eb5131a729ffcb950c63978cef03db723ace5353e6cb1430d02d09d7e9"} Feb 27 17:18:08 crc kubenswrapper[4700]: I0227 17:18:08.618219 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerStarted","Data":"c99a71cb46c45356ba1bf62a0adf7979247710d0e51e5eac769c9c248f05463d"} Feb 27 17:18:08 crc kubenswrapper[4700]: I0227 17:18:08.622972 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qmhw5" event={"ID":"b1a1f4b1-e13f-4b11-bdeb-5d815af7b98b","Type":"ContainerStarted","Data":"cdcce2cda8b113b14406bad56743fc404a521b843fdeff66fc57c595565ff66b"} Feb 27 17:18:08 crc kubenswrapper[4700]: I0227 17:18:08.626247 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wwl7g" event={"ID":"e548373a-d9bf-46bd-8efa-b4adf2dedc4b","Type":"ContainerStarted","Data":"1136a57c54af307682805b5db5c45047ff68468150674f891a45786ee0d36bff"} Feb 27 17:18:08 crc kubenswrapper[4700]: I0227 17:18:08.645932 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wwl7g" podStartSLOduration=2.050154141 podStartE2EDuration="8.645911417s" podCreationTimestamp="2026-02-27 17:18:00 +0000 UTC" firstStartedPulling="2026-02-27 17:18:01.514474811 +0000 UTC m=+1041.499787568" lastFinishedPulling="2026-02-27 17:18:08.110232087 +0000 UTC m=+1048.095544844" observedRunningTime="2026-02-27 17:18:08.645506866 +0000 UTC m=+1048.630819633" watchObservedRunningTime="2026-02-27 17:18:08.645911417 +0000 UTC m=+1048.631224164" Feb 27 17:18:09 crc kubenswrapper[4700]: I0227 17:18:09.636062 4700 generic.go:334] "Generic (PLEG): container finished" podID="b1a1f4b1-e13f-4b11-bdeb-5d815af7b98b" containerID="cdcce2cda8b113b14406bad56743fc404a521b843fdeff66fc57c595565ff66b" exitCode=0 Feb 27 17:18:09 crc kubenswrapper[4700]: I0227 17:18:09.636181 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qmhw5" event={"ID":"b1a1f4b1-e13f-4b11-bdeb-5d815af7b98b","Type":"ContainerDied","Data":"cdcce2cda8b113b14406bad56743fc404a521b843fdeff66fc57c595565ff66b"} Feb 27 17:18:09 crc kubenswrapper[4700]: I0227 17:18:09.971247 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536878-zp9lh" Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.064578 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2z4d\" (UniqueName: \"kubernetes.io/projected/7765df9b-4c2e-4ac9-b862-4c300d69acb5-kube-api-access-j2z4d\") pod \"7765df9b-4c2e-4ac9-b862-4c300d69acb5\" (UID: \"7765df9b-4c2e-4ac9-b862-4c300d69acb5\") " Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.069955 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7765df9b-4c2e-4ac9-b862-4c300d69acb5-kube-api-access-j2z4d" (OuterVolumeSpecName: "kube-api-access-j2z4d") pod "7765df9b-4c2e-4ac9-b862-4c300d69acb5" (UID: "7765df9b-4c2e-4ac9-b862-4c300d69acb5"). InnerVolumeSpecName "kube-api-access-j2z4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.166014 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2z4d\" (UniqueName: \"kubernetes.io/projected/7765df9b-4c2e-4ac9-b862-4c300d69acb5-kube-api-access-j2z4d\") on node \"crc\" DevicePath \"\"" Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.424257 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wwl7g" Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.424300 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wwl7g" Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.646321 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536878-zp9lh" event={"ID":"7765df9b-4c2e-4ac9-b862-4c300d69acb5","Type":"ContainerDied","Data":"8d30a6f98833e22430d334595f7d2269f7eb9ee712b992f79df5def0b6273216"} Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.646398 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d30a6f98833e22430d334595f7d2269f7eb9ee712b992f79df5def0b6273216" Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.646505 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536878-zp9lh" Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.649782 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qmhw5" event={"ID":"b1a1f4b1-e13f-4b11-bdeb-5d815af7b98b","Type":"ContainerStarted","Data":"f7d31f494af1fe06b05188eebe26020f53f70cc2d9332ead3329ba80b3b57baa"} Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.676649 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qmhw5" podStartSLOduration=3.069203902 podStartE2EDuration="13.676625265s" podCreationTimestamp="2026-02-27 17:17:57 +0000 UTC" firstStartedPulling="2026-02-27 17:17:59.496694615 +0000 UTC m=+1039.482007362" lastFinishedPulling="2026-02-27 17:18:10.104115978 +0000 UTC m=+1050.089428725" observedRunningTime="2026-02-27 17:18:10.668909531 +0000 UTC m=+1050.654222288" watchObservedRunningTime="2026-02-27 17:18:10.676625265 +0000 UTC m=+1050.661938022" Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.827221 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6db6876945-nmv94"] Feb 27 17:18:10 crc kubenswrapper[4700]: E0227 17:18:10.827658 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7765df9b-4c2e-4ac9-b862-4c300d69acb5" containerName="oc" Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.827668 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="7765df9b-4c2e-4ac9-b862-4c300d69acb5" containerName="oc" Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.827790 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="7765df9b-4c2e-4ac9-b862-4c300d69acb5" containerName="oc" Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.828179 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6db6876945-nmv94" Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.830186 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-cmf9x" Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.841602 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6db6876945-nmv94"] Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.851730 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-55d77d7b5c-dc72n"] Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.852731 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-dc72n" Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.856017 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-5d87c9d997-9mxq4"] Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.856885 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-5d87c9d997-9mxq4" Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.858228 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-b8lsj" Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.860235 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-5q59n" Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.874371 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bq9vk\" (UniqueName: \"kubernetes.io/projected/37c36ebd-e968-4024-a8c5-77c6507b95cf-kube-api-access-bq9vk\") pod \"barbican-operator-controller-manager-6db6876945-nmv94\" (UID: \"37c36ebd-e968-4024-a8c5-77c6507b95cf\") " pod="openstack-operators/barbican-operator-controller-manager-6db6876945-nmv94" Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.883421 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-55d77d7b5c-dc72n"] Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.888144 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-64db6967f8-l2542"] Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.888902 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-64db6967f8-l2542" Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.890553 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-g25kh" Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.934168 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-64db6967f8-l2542"] Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.953749 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-5d87c9d997-9mxq4"] Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.958258 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-cf99c678f-n55xj"] Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.959093 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-cf99c678f-n55xj" Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.962902 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-7jdfm" Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.963893 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-dm8wj"] Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.964697 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-dm8wj" Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.976991 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-k524x" Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.978103 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bq9vk\" (UniqueName: \"kubernetes.io/projected/37c36ebd-e968-4024-a8c5-77c6507b95cf-kube-api-access-bq9vk\") pod \"barbican-operator-controller-manager-6db6876945-nmv94\" (UID: \"37c36ebd-e968-4024-a8c5-77c6507b95cf\") " pod="openstack-operators/barbican-operator-controller-manager-6db6876945-nmv94" Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.978190 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9jqx\" (UniqueName: \"kubernetes.io/projected/181557a9-fcb7-4554-82ae-d3446b460867-kube-api-access-x9jqx\") pod \"glance-operator-controller-manager-64db6967f8-l2542\" (UID: \"181557a9-fcb7-4554-82ae-d3446b460867\") " pod="openstack-operators/glance-operator-controller-manager-64db6967f8-l2542" Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.978228 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vx6sh\" (UniqueName: \"kubernetes.io/projected/9b575398-8936-48f1-a412-ec26c21303cf-kube-api-access-vx6sh\") pod \"designate-operator-controller-manager-5d87c9d997-9mxq4\" (UID: \"9b575398-8936-48f1-a412-ec26c21303cf\") " pod="openstack-operators/designate-operator-controller-manager-5d87c9d997-9mxq4" Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.978269 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfsbf\" (UniqueName: \"kubernetes.io/projected/dcdee90e-b946-4cf5-807c-6babcd83071f-kube-api-access-nfsbf\") pod \"cinder-operator-controller-manager-55d77d7b5c-dc72n\" (UID: \"dcdee90e-b946-4cf5-807c-6babcd83071f\") " pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-dc72n" Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.978602 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-dm8wj"] Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.991356 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-cf99c678f-n55xj"] Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.993707 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-f7fcc58b9-84zw4"] Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.994666 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-84zw4" Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.997531 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-zkvlr" Feb 27 17:18:10 crc kubenswrapper[4700]: I0227 17:18:10.997577 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.008701 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-545456dc4-stc8v"] Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.009447 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-545456dc4-stc8v" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.010852 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bq9vk\" (UniqueName: \"kubernetes.io/projected/37c36ebd-e968-4024-a8c5-77c6507b95cf-kube-api-access-bq9vk\") pod \"barbican-operator-controller-manager-6db6876945-nmv94\" (UID: \"37c36ebd-e968-4024-a8c5-77c6507b95cf\") " pod="openstack-operators/barbican-operator-controller-manager-6db6876945-nmv94" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.015180 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-f7fcc58b9-84zw4"] Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.021843 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-545456dc4-stc8v"] Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.026762 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-mc9nd" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.034526 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55ffd4876b-q66w8"] Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.035566 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-55ffd4876b-q66w8" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.038867 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-gq5wn" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.063950 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-67d996989d-f625b"] Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.064762 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-67d996989d-f625b" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.068745 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-r64rw" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.073917 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55ffd4876b-q66w8"] Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.080211 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-67d996989d-f625b"] Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.080843 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7ngq\" (UniqueName: \"kubernetes.io/projected/7cd71afd-89a7-42b9-9913-6827ea7a22d3-kube-api-access-n7ngq\") pod \"keystone-operator-controller-manager-55ffd4876b-q66w8\" (UID: \"7cd71afd-89a7-42b9-9913-6827ea7a22d3\") " pod="openstack-operators/keystone-operator-controller-manager-55ffd4876b-q66w8" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.080899 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/56bc1032-efb2-4480-9091-11e529096a3c-cert\") pod \"infra-operator-controller-manager-f7fcc58b9-84zw4\" (UID: \"56bc1032-efb2-4480-9091-11e529096a3c\") " pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-84zw4" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.080928 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfxrz\" (UniqueName: \"kubernetes.io/projected/56bc1032-efb2-4480-9091-11e529096a3c-kube-api-access-lfxrz\") pod \"infra-operator-controller-manager-f7fcc58b9-84zw4\" (UID: \"56bc1032-efb2-4480-9091-11e529096a3c\") " pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-84zw4" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.080954 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tmj8\" (UniqueName: \"kubernetes.io/projected/c3ac9c3b-2142-409d-8c49-1bf2837252d7-kube-api-access-5tmj8\") pod \"horizon-operator-controller-manager-78bc7f9bd9-dm8wj\" (UID: \"c3ac9c3b-2142-409d-8c49-1bf2837252d7\") " pod="openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-dm8wj" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.080978 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9jqx\" (UniqueName: \"kubernetes.io/projected/181557a9-fcb7-4554-82ae-d3446b460867-kube-api-access-x9jqx\") pod \"glance-operator-controller-manager-64db6967f8-l2542\" (UID: \"181557a9-fcb7-4554-82ae-d3446b460867\") " pod="openstack-operators/glance-operator-controller-manager-64db6967f8-l2542" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.080998 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vx6sh\" (UniqueName: \"kubernetes.io/projected/9b575398-8936-48f1-a412-ec26c21303cf-kube-api-access-vx6sh\") pod \"designate-operator-controller-manager-5d87c9d997-9mxq4\" (UID: \"9b575398-8936-48f1-a412-ec26c21303cf\") " pod="openstack-operators/designate-operator-controller-manager-5d87c9d997-9mxq4" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.081016 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvksb\" (UniqueName: \"kubernetes.io/projected/ab00dea0-59fa-49ea-92f9-a230da6ea536-kube-api-access-fvksb\") pod \"heat-operator-controller-manager-cf99c678f-n55xj\" (UID: \"ab00dea0-59fa-49ea-92f9-a230da6ea536\") " pod="openstack-operators/heat-operator-controller-manager-cf99c678f-n55xj" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.081042 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbmtm\" (UniqueName: \"kubernetes.io/projected/9c1db815-93de-4bcc-9698-22bf6588ac88-kube-api-access-zbmtm\") pod \"ironic-operator-controller-manager-545456dc4-stc8v\" (UID: \"9c1db815-93de-4bcc-9698-22bf6588ac88\") " pod="openstack-operators/ironic-operator-controller-manager-545456dc4-stc8v" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.081074 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfsbf\" (UniqueName: \"kubernetes.io/projected/dcdee90e-b946-4cf5-807c-6babcd83071f-kube-api-access-nfsbf\") pod \"cinder-operator-controller-manager-55d77d7b5c-dc72n\" (UID: \"dcdee90e-b946-4cf5-807c-6babcd83071f\") " pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-dc72n" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.089383 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-556b8b874-4bmbt"] Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.090233 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-556b8b874-4bmbt" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.097270 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-dxqwm" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.101148 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-556b8b874-4bmbt"] Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.130986 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vx6sh\" (UniqueName: \"kubernetes.io/projected/9b575398-8936-48f1-a412-ec26c21303cf-kube-api-access-vx6sh\") pod \"designate-operator-controller-manager-5d87c9d997-9mxq4\" (UID: \"9b575398-8936-48f1-a412-ec26c21303cf\") " pod="openstack-operators/designate-operator-controller-manager-5d87c9d997-9mxq4" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.141812 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfsbf\" (UniqueName: \"kubernetes.io/projected/dcdee90e-b946-4cf5-807c-6babcd83071f-kube-api-access-nfsbf\") pod \"cinder-operator-controller-manager-55d77d7b5c-dc72n\" (UID: \"dcdee90e-b946-4cf5-807c-6babcd83071f\") " pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-dc72n" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.152377 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9jqx\" (UniqueName: \"kubernetes.io/projected/181557a9-fcb7-4554-82ae-d3446b460867-kube-api-access-x9jqx\") pod \"glance-operator-controller-manager-64db6967f8-l2542\" (UID: \"181557a9-fcb7-4554-82ae-d3446b460867\") " pod="openstack-operators/glance-operator-controller-manager-64db6967f8-l2542" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.161903 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-54688575f-8vhl9"] Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.162432 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6db6876945-nmv94" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.163352 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-54688575f-8vhl9" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.174019 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-7lxx7" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.183273 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jc9f\" (UniqueName: \"kubernetes.io/projected/e6d0216e-25c4-4957-8e6b-473b49dd7543-kube-api-access-5jc9f\") pod \"manila-operator-controller-manager-67d996989d-f625b\" (UID: \"e6d0216e-25c4-4957-8e6b-473b49dd7543\") " pod="openstack-operators/manila-operator-controller-manager-67d996989d-f625b" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.183345 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7ngq\" (UniqueName: \"kubernetes.io/projected/7cd71afd-89a7-42b9-9913-6827ea7a22d3-kube-api-access-n7ngq\") pod \"keystone-operator-controller-manager-55ffd4876b-q66w8\" (UID: \"7cd71afd-89a7-42b9-9913-6827ea7a22d3\") " pod="openstack-operators/keystone-operator-controller-manager-55ffd4876b-q66w8" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.183382 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/56bc1032-efb2-4480-9091-11e529096a3c-cert\") pod \"infra-operator-controller-manager-f7fcc58b9-84zw4\" (UID: \"56bc1032-efb2-4480-9091-11e529096a3c\") " pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-84zw4" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.183408 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfxrz\" (UniqueName: \"kubernetes.io/projected/56bc1032-efb2-4480-9091-11e529096a3c-kube-api-access-lfxrz\") pod \"infra-operator-controller-manager-f7fcc58b9-84zw4\" (UID: \"56bc1032-efb2-4480-9091-11e529096a3c\") " pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-84zw4" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.183431 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dm8dd\" (UniqueName: \"kubernetes.io/projected/c1fdff54-9c43-4fe4-925b-f87590c75e72-kube-api-access-dm8dd\") pod \"mariadb-operator-controller-manager-556b8b874-4bmbt\" (UID: \"c1fdff54-9c43-4fe4-925b-f87590c75e72\") " pod="openstack-operators/mariadb-operator-controller-manager-556b8b874-4bmbt" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.183449 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tmj8\" (UniqueName: \"kubernetes.io/projected/c3ac9c3b-2142-409d-8c49-1bf2837252d7-kube-api-access-5tmj8\") pod \"horizon-operator-controller-manager-78bc7f9bd9-dm8wj\" (UID: \"c3ac9c3b-2142-409d-8c49-1bf2837252d7\") " pod="openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-dm8wj" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.183528 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvksb\" (UniqueName: \"kubernetes.io/projected/ab00dea0-59fa-49ea-92f9-a230da6ea536-kube-api-access-fvksb\") pod \"heat-operator-controller-manager-cf99c678f-n55xj\" (UID: \"ab00dea0-59fa-49ea-92f9-a230da6ea536\") " pod="openstack-operators/heat-operator-controller-manager-cf99c678f-n55xj" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.183558 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbmtm\" (UniqueName: \"kubernetes.io/projected/9c1db815-93de-4bcc-9698-22bf6588ac88-kube-api-access-zbmtm\") pod \"ironic-operator-controller-manager-545456dc4-stc8v\" (UID: \"9c1db815-93de-4bcc-9698-22bf6588ac88\") " pod="openstack-operators/ironic-operator-controller-manager-545456dc4-stc8v" Feb 27 17:18:11 crc kubenswrapper[4700]: E0227 17:18:11.183738 4700 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 27 17:18:11 crc kubenswrapper[4700]: E0227 17:18:11.183794 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/56bc1032-efb2-4480-9091-11e529096a3c-cert podName:56bc1032-efb2-4480-9091-11e529096a3c nodeName:}" failed. No retries permitted until 2026-02-27 17:18:11.683778609 +0000 UTC m=+1051.669091356 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/56bc1032-efb2-4480-9091-11e529096a3c-cert") pod "infra-operator-controller-manager-f7fcc58b9-84zw4" (UID: "56bc1032-efb2-4480-9091-11e529096a3c") : secret "infra-operator-webhook-server-cert" not found Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.189932 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-dc72n" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.200761 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-5d87c9d997-9mxq4" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.213855 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-74b6b5dc96-72l74"] Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.214696 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-74b6b5dc96-72l74" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.217560 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-64db6967f8-l2542" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.217885 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-2ht29" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.220120 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfxrz\" (UniqueName: \"kubernetes.io/projected/56bc1032-efb2-4480-9091-11e529096a3c-kube-api-access-lfxrz\") pod \"infra-operator-controller-manager-f7fcc58b9-84zw4\" (UID: \"56bc1032-efb2-4480-9091-11e529096a3c\") " pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-84zw4" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.220285 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbmtm\" (UniqueName: \"kubernetes.io/projected/9c1db815-93de-4bcc-9698-22bf6588ac88-kube-api-access-zbmtm\") pod \"ironic-operator-controller-manager-545456dc4-stc8v\" (UID: \"9c1db815-93de-4bcc-9698-22bf6588ac88\") " pod="openstack-operators/ironic-operator-controller-manager-545456dc4-stc8v" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.228382 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvksb\" (UniqueName: \"kubernetes.io/projected/ab00dea0-59fa-49ea-92f9-a230da6ea536-kube-api-access-fvksb\") pod \"heat-operator-controller-manager-cf99c678f-n55xj\" (UID: \"ab00dea0-59fa-49ea-92f9-a230da6ea536\") " pod="openstack-operators/heat-operator-controller-manager-cf99c678f-n55xj" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.236620 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7ngq\" (UniqueName: \"kubernetes.io/projected/7cd71afd-89a7-42b9-9913-6827ea7a22d3-kube-api-access-n7ngq\") pod \"keystone-operator-controller-manager-55ffd4876b-q66w8\" (UID: \"7cd71afd-89a7-42b9-9913-6827ea7a22d3\") " pod="openstack-operators/keystone-operator-controller-manager-55ffd4876b-q66w8" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.249061 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tmj8\" (UniqueName: \"kubernetes.io/projected/c3ac9c3b-2142-409d-8c49-1bf2837252d7-kube-api-access-5tmj8\") pod \"horizon-operator-controller-manager-78bc7f9bd9-dm8wj\" (UID: \"c3ac9c3b-2142-409d-8c49-1bf2837252d7\") " pod="openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-dm8wj" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.250728 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-54688575f-8vhl9"] Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.269119 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536872-tkkz5"] Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.274672 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536872-tkkz5"] Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.279345 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-cf99c678f-n55xj" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.279752 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-vt248"] Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.280556 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-vt248" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.283803 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-rw55d" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.284417 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqbn7\" (UniqueName: \"kubernetes.io/projected/e00942b5-6f82-48d4-8413-2f568daf5d7f-kube-api-access-hqbn7\") pod \"neutron-operator-controller-manager-54688575f-8vhl9\" (UID: \"e00942b5-6f82-48d4-8413-2f568daf5d7f\") " pod="openstack-operators/neutron-operator-controller-manager-54688575f-8vhl9" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.284470 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dm8dd\" (UniqueName: \"kubernetes.io/projected/c1fdff54-9c43-4fe4-925b-f87590c75e72-kube-api-access-dm8dd\") pod \"mariadb-operator-controller-manager-556b8b874-4bmbt\" (UID: \"c1fdff54-9c43-4fe4-925b-f87590c75e72\") " pod="openstack-operators/mariadb-operator-controller-manager-556b8b874-4bmbt" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.284875 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jc9f\" (UniqueName: \"kubernetes.io/projected/e6d0216e-25c4-4957-8e6b-473b49dd7543-kube-api-access-5jc9f\") pod \"manila-operator-controller-manager-67d996989d-f625b\" (UID: \"e6d0216e-25c4-4957-8e6b-473b49dd7543\") " pod="openstack-operators/manila-operator-controller-manager-67d996989d-f625b" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.284905 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5kmq\" (UniqueName: \"kubernetes.io/projected/5bbcc1bb-6070-4196-8212-cf8e04eaa923-kube-api-access-p5kmq\") pod \"nova-operator-controller-manager-74b6b5dc96-72l74\" (UID: \"5bbcc1bb-6070-4196-8212-cf8e04eaa923\") " pod="openstack-operators/nova-operator-controller-manager-74b6b5dc96-72l74" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.290104 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-dm8wj" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.293827 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-74b6b5dc96-72l74"] Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.302582 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jc9f\" (UniqueName: \"kubernetes.io/projected/e6d0216e-25c4-4957-8e6b-473b49dd7543-kube-api-access-5jc9f\") pod \"manila-operator-controller-manager-67d996989d-f625b\" (UID: \"e6d0216e-25c4-4957-8e6b-473b49dd7543\") " pod="openstack-operators/manila-operator-controller-manager-67d996989d-f625b" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.304213 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dm8dd\" (UniqueName: \"kubernetes.io/projected/c1fdff54-9c43-4fe4-925b-f87590c75e72-kube-api-access-dm8dd\") pod \"mariadb-operator-controller-manager-556b8b874-4bmbt\" (UID: \"c1fdff54-9c43-4fe4-925b-f87590c75e72\") " pod="openstack-operators/mariadb-operator-controller-manager-556b8b874-4bmbt" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.308005 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-vt248"] Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.313074 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg"] Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.314022 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.317831 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.318035 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-7h4gq" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.321019 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-75684d597f-6cfj6"] Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.332709 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-75684d597f-6cfj6" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.336409 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-f6dv8" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.347881 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-75684d597f-6cfj6"] Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.366900 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-545456dc4-stc8v" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.392694 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-55ffd4876b-q66w8" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.396150 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzqzc\" (UniqueName: \"kubernetes.io/projected/00a21bc7-24a6-4a6d-ba48-18db1d9233c6-kube-api-access-hzqzc\") pod \"ovn-operator-controller-manager-75684d597f-6cfj6\" (UID: \"00a21bc7-24a6-4a6d-ba48-18db1d9233c6\") " pod="openstack-operators/ovn-operator-controller-manager-75684d597f-6cfj6" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.396209 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqbn7\" (UniqueName: \"kubernetes.io/projected/e00942b5-6f82-48d4-8413-2f568daf5d7f-kube-api-access-hqbn7\") pod \"neutron-operator-controller-manager-54688575f-8vhl9\" (UID: \"e00942b5-6f82-48d4-8413-2f568daf5d7f\") " pod="openstack-operators/neutron-operator-controller-manager-54688575f-8vhl9" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.396240 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98qxm\" (UniqueName: \"kubernetes.io/projected/f6a3d37c-2b18-40bc-a990-56dc67093abf-kube-api-access-98qxm\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg\" (UID: \"f6a3d37c-2b18-40bc-a990-56dc67093abf\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.396271 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f6a3d37c-2b18-40bc-a990-56dc67093abf-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg\" (UID: \"f6a3d37c-2b18-40bc-a990-56dc67093abf\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.396301 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km2h8\" (UniqueName: \"kubernetes.io/projected/bd70bcca-beb1-464b-a1a3-b154ff45c1f5-kube-api-access-km2h8\") pod \"octavia-operator-controller-manager-5d86c7ddb7-vt248\" (UID: \"bd70bcca-beb1-464b-a1a3-b154ff45c1f5\") " pod="openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-vt248" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.396332 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5kmq\" (UniqueName: \"kubernetes.io/projected/5bbcc1bb-6070-4196-8212-cf8e04eaa923-kube-api-access-p5kmq\") pod \"nova-operator-controller-manager-74b6b5dc96-72l74\" (UID: \"5bbcc1bb-6070-4196-8212-cf8e04eaa923\") " pod="openstack-operators/nova-operator-controller-manager-74b6b5dc96-72l74" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.397793 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-67d996989d-f625b" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.408980 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg"] Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.417366 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5kmq\" (UniqueName: \"kubernetes.io/projected/5bbcc1bb-6070-4196-8212-cf8e04eaa923-kube-api-access-p5kmq\") pod \"nova-operator-controller-manager-74b6b5dc96-72l74\" (UID: \"5bbcc1bb-6070-4196-8212-cf8e04eaa923\") " pod="openstack-operators/nova-operator-controller-manager-74b6b5dc96-72l74" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.418821 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-648564c9fc-dmd7w"] Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.419731 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-648564c9fc-dmd7w" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.419739 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqbn7\" (UniqueName: \"kubernetes.io/projected/e00942b5-6f82-48d4-8413-2f568daf5d7f-kube-api-access-hqbn7\") pod \"neutron-operator-controller-manager-54688575f-8vhl9\" (UID: \"e00942b5-6f82-48d4-8413-2f568daf5d7f\") " pod="openstack-operators/neutron-operator-controller-manager-54688575f-8vhl9" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.422047 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-h8ztn" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.467853 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-9b9ff9f4d-cb7w5"] Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.468757 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-648564c9fc-dmd7w"] Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.468831 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9b9ff9f4d-cb7w5" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.472691 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-ndj9m" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.481760 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-wwl7g" podUID="e548373a-d9bf-46bd-8efa-b4adf2dedc4b" containerName="registry-server" probeResult="failure" output=< Feb 27 17:18:11 crc kubenswrapper[4700]: timeout: failed to connect service ":50051" within 1s Feb 27 17:18:11 crc kubenswrapper[4700]: > Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.481938 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9b9ff9f4d-cb7w5"] Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.498682 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzqzc\" (UniqueName: \"kubernetes.io/projected/00a21bc7-24a6-4a6d-ba48-18db1d9233c6-kube-api-access-hzqzc\") pod \"ovn-operator-controller-manager-75684d597f-6cfj6\" (UID: \"00a21bc7-24a6-4a6d-ba48-18db1d9233c6\") " pod="openstack-operators/ovn-operator-controller-manager-75684d597f-6cfj6" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.498760 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98qxm\" (UniqueName: \"kubernetes.io/projected/f6a3d37c-2b18-40bc-a990-56dc67093abf-kube-api-access-98qxm\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg\" (UID: \"f6a3d37c-2b18-40bc-a990-56dc67093abf\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.498794 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f6a3d37c-2b18-40bc-a990-56dc67093abf-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg\" (UID: \"f6a3d37c-2b18-40bc-a990-56dc67093abf\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.498828 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km2h8\" (UniqueName: \"kubernetes.io/projected/bd70bcca-beb1-464b-a1a3-b154ff45c1f5-kube-api-access-km2h8\") pod \"octavia-operator-controller-manager-5d86c7ddb7-vt248\" (UID: \"bd70bcca-beb1-464b-a1a3-b154ff45c1f5\") " pod="openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-vt248" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.498856 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7m8nj\" (UniqueName: \"kubernetes.io/projected/5d94c9a7-4f4e-4370-a3da-9d80309f327a-kube-api-access-7m8nj\") pod \"placement-operator-controller-manager-648564c9fc-dmd7w\" (UID: \"5d94c9a7-4f4e-4370-a3da-9d80309f327a\") " pod="openstack-operators/placement-operator-controller-manager-648564c9fc-dmd7w" Feb 27 17:18:11 crc kubenswrapper[4700]: E0227 17:18:11.499329 4700 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 27 17:18:11 crc kubenswrapper[4700]: E0227 17:18:11.499366 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f6a3d37c-2b18-40bc-a990-56dc67093abf-cert podName:f6a3d37c-2b18-40bc-a990-56dc67093abf nodeName:}" failed. No retries permitted until 2026-02-27 17:18:11.99935328 +0000 UTC m=+1051.984666027 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f6a3d37c-2b18-40bc-a990-56dc67093abf-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg" (UID: "f6a3d37c-2b18-40bc-a990-56dc67093abf") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.513795 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5fdb694969-44djz"] Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.516884 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5fdb694969-44djz"] Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.516974 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5fdb694969-44djz" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.519169 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzqzc\" (UniqueName: \"kubernetes.io/projected/00a21bc7-24a6-4a6d-ba48-18db1d9233c6-kube-api-access-hzqzc\") pod \"ovn-operator-controller-manager-75684d597f-6cfj6\" (UID: \"00a21bc7-24a6-4a6d-ba48-18db1d9233c6\") " pod="openstack-operators/ovn-operator-controller-manager-75684d597f-6cfj6" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.522794 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-nfjbj" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.525268 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-55b5ff4dbb-xgf2x"] Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.526402 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-xgf2x" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.530185 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km2h8\" (UniqueName: \"kubernetes.io/projected/bd70bcca-beb1-464b-a1a3-b154ff45c1f5-kube-api-access-km2h8\") pod \"octavia-operator-controller-manager-5d86c7ddb7-vt248\" (UID: \"bd70bcca-beb1-464b-a1a3-b154ff45c1f5\") " pod="openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-vt248" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.530918 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-qlcq4" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.534526 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-55b5ff4dbb-xgf2x"] Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.540676 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-556b8b874-4bmbt" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.545860 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98qxm\" (UniqueName: \"kubernetes.io/projected/f6a3d37c-2b18-40bc-a990-56dc67093abf-kube-api-access-98qxm\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg\" (UID: \"f6a3d37c-2b18-40bc-a990-56dc67093abf\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.552562 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-8f7484475-8pdn6"] Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.553537 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-8f7484475-8pdn6" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.554901 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-54688575f-8vhl9" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.557591 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-8f7484475-8pdn6"] Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.566180 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-74b6b5dc96-72l74" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.568543 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-8prt6" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.600967 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-vt248" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.602012 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7m8nj\" (UniqueName: \"kubernetes.io/projected/5d94c9a7-4f4e-4370-a3da-9d80309f327a-kube-api-access-7m8nj\") pod \"placement-operator-controller-manager-648564c9fc-dmd7w\" (UID: \"5d94c9a7-4f4e-4370-a3da-9d80309f327a\") " pod="openstack-operators/placement-operator-controller-manager-648564c9fc-dmd7w" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.602106 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6mg7\" (UniqueName: \"kubernetes.io/projected/629a3ab5-c8ce-49cb-bd57-355d643253c5-kube-api-access-q6mg7\") pod \"telemetry-operator-controller-manager-5fdb694969-44djz\" (UID: \"629a3ab5-c8ce-49cb-bd57-355d643253c5\") " pod="openstack-operators/telemetry-operator-controller-manager-5fdb694969-44djz" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.602173 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmv6w\" (UniqueName: \"kubernetes.io/projected/269a9c0b-7242-44dc-b860-386757217052-kube-api-access-hmv6w\") pod \"test-operator-controller-manager-55b5ff4dbb-xgf2x\" (UID: \"269a9c0b-7242-44dc-b860-386757217052\") " pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-xgf2x" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.602199 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2smz\" (UniqueName: \"kubernetes.io/projected/15764e3b-3a40-4742-91a4-3d339cb86a0b-kube-api-access-w2smz\") pod \"swift-operator-controller-manager-9b9ff9f4d-cb7w5\" (UID: \"15764e3b-3a40-4742-91a4-3d339cb86a0b\") " pod="openstack-operators/swift-operator-controller-manager-9b9ff9f4d-cb7w5" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.602399 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgcdc\" (UniqueName: \"kubernetes.io/projected/33fd8eaf-c78f-4949-be53-86d8981ca296-kube-api-access-mgcdc\") pod \"watcher-operator-controller-manager-8f7484475-8pdn6\" (UID: \"33fd8eaf-c78f-4949-be53-86d8981ca296\") " pod="openstack-operators/watcher-operator-controller-manager-8f7484475-8pdn6" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.603662 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7ff59cf98b-bntz4"] Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.606881 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7ff59cf98b-bntz4" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.611076 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-nb726" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.611300 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.611410 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.642065 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7m8nj\" (UniqueName: \"kubernetes.io/projected/5d94c9a7-4f4e-4370-a3da-9d80309f327a-kube-api-access-7m8nj\") pod \"placement-operator-controller-manager-648564c9fc-dmd7w\" (UID: \"5d94c9a7-4f4e-4370-a3da-9d80309f327a\") " pod="openstack-operators/placement-operator-controller-manager-648564c9fc-dmd7w" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.647424 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7ff59cf98b-bntz4"] Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.658519 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qf2g9"] Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.659341 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-75684d597f-6cfj6" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.659688 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qf2g9" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.662625 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-ftw8p" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.672024 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qf2g9"] Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.705078 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6mg7\" (UniqueName: \"kubernetes.io/projected/629a3ab5-c8ce-49cb-bd57-355d643253c5-kube-api-access-q6mg7\") pod \"telemetry-operator-controller-manager-5fdb694969-44djz\" (UID: \"629a3ab5-c8ce-49cb-bd57-355d643253c5\") " pod="openstack-operators/telemetry-operator-controller-manager-5fdb694969-44djz" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.705136 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-webhook-certs\") pod \"openstack-operator-controller-manager-7ff59cf98b-bntz4\" (UID: \"fe5a4feb-297b-4c17-bc6c-cab1bf1dee41\") " pod="openstack-operators/openstack-operator-controller-manager-7ff59cf98b-bntz4" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.705159 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82s8m\" (UniqueName: \"kubernetes.io/projected/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-kube-api-access-82s8m\") pod \"openstack-operator-controller-manager-7ff59cf98b-bntz4\" (UID: \"fe5a4feb-297b-4c17-bc6c-cab1bf1dee41\") " pod="openstack-operators/openstack-operator-controller-manager-7ff59cf98b-bntz4" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.705207 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmv6w\" (UniqueName: \"kubernetes.io/projected/269a9c0b-7242-44dc-b860-386757217052-kube-api-access-hmv6w\") pod \"test-operator-controller-manager-55b5ff4dbb-xgf2x\" (UID: \"269a9c0b-7242-44dc-b860-386757217052\") " pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-xgf2x" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.705228 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2smz\" (UniqueName: \"kubernetes.io/projected/15764e3b-3a40-4742-91a4-3d339cb86a0b-kube-api-access-w2smz\") pod \"swift-operator-controller-manager-9b9ff9f4d-cb7w5\" (UID: \"15764e3b-3a40-4742-91a4-3d339cb86a0b\") " pod="openstack-operators/swift-operator-controller-manager-9b9ff9f4d-cb7w5" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.705246 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-metrics-certs\") pod \"openstack-operator-controller-manager-7ff59cf98b-bntz4\" (UID: \"fe5a4feb-297b-4c17-bc6c-cab1bf1dee41\") " pod="openstack-operators/openstack-operator-controller-manager-7ff59cf98b-bntz4" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.705277 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/56bc1032-efb2-4480-9091-11e529096a3c-cert\") pod \"infra-operator-controller-manager-f7fcc58b9-84zw4\" (UID: \"56bc1032-efb2-4480-9091-11e529096a3c\") " pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-84zw4" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.705338 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fcvd\" (UniqueName: \"kubernetes.io/projected/1010adc0-3454-486d-a676-80d329745d4c-kube-api-access-5fcvd\") pod \"rabbitmq-cluster-operator-manager-668c99d594-qf2g9\" (UID: \"1010adc0-3454-486d-a676-80d329745d4c\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qf2g9" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.705356 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgcdc\" (UniqueName: \"kubernetes.io/projected/33fd8eaf-c78f-4949-be53-86d8981ca296-kube-api-access-mgcdc\") pod \"watcher-operator-controller-manager-8f7484475-8pdn6\" (UID: \"33fd8eaf-c78f-4949-be53-86d8981ca296\") " pod="openstack-operators/watcher-operator-controller-manager-8f7484475-8pdn6" Feb 27 17:18:11 crc kubenswrapper[4700]: E0227 17:18:11.706011 4700 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 27 17:18:11 crc kubenswrapper[4700]: E0227 17:18:11.706065 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/56bc1032-efb2-4480-9091-11e529096a3c-cert podName:56bc1032-efb2-4480-9091-11e529096a3c nodeName:}" failed. No retries permitted until 2026-02-27 17:18:12.706047463 +0000 UTC m=+1052.691360210 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/56bc1032-efb2-4480-9091-11e529096a3c-cert") pod "infra-operator-controller-manager-f7fcc58b9-84zw4" (UID: "56bc1032-efb2-4480-9091-11e529096a3c") : secret "infra-operator-webhook-server-cert" not found Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.726991 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgcdc\" (UniqueName: \"kubernetes.io/projected/33fd8eaf-c78f-4949-be53-86d8981ca296-kube-api-access-mgcdc\") pod \"watcher-operator-controller-manager-8f7484475-8pdn6\" (UID: \"33fd8eaf-c78f-4949-be53-86d8981ca296\") " pod="openstack-operators/watcher-operator-controller-manager-8f7484475-8pdn6" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.727573 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmv6w\" (UniqueName: \"kubernetes.io/projected/269a9c0b-7242-44dc-b860-386757217052-kube-api-access-hmv6w\") pod \"test-operator-controller-manager-55b5ff4dbb-xgf2x\" (UID: \"269a9c0b-7242-44dc-b860-386757217052\") " pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-xgf2x" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.728387 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2smz\" (UniqueName: \"kubernetes.io/projected/15764e3b-3a40-4742-91a4-3d339cb86a0b-kube-api-access-w2smz\") pod \"swift-operator-controller-manager-9b9ff9f4d-cb7w5\" (UID: \"15764e3b-3a40-4742-91a4-3d339cb86a0b\") " pod="openstack-operators/swift-operator-controller-manager-9b9ff9f4d-cb7w5" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.728710 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6mg7\" (UniqueName: \"kubernetes.io/projected/629a3ab5-c8ce-49cb-bd57-355d643253c5-kube-api-access-q6mg7\") pod \"telemetry-operator-controller-manager-5fdb694969-44djz\" (UID: \"629a3ab5-c8ce-49cb-bd57-355d643253c5\") " pod="openstack-operators/telemetry-operator-controller-manager-5fdb694969-44djz" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.731638 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5fdb694969-44djz" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.744262 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-648564c9fc-dmd7w" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.754955 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-xgf2x" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.778042 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-8f7484475-8pdn6" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.806090 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-webhook-certs\") pod \"openstack-operator-controller-manager-7ff59cf98b-bntz4\" (UID: \"fe5a4feb-297b-4c17-bc6c-cab1bf1dee41\") " pod="openstack-operators/openstack-operator-controller-manager-7ff59cf98b-bntz4" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.806131 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82s8m\" (UniqueName: \"kubernetes.io/projected/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-kube-api-access-82s8m\") pod \"openstack-operator-controller-manager-7ff59cf98b-bntz4\" (UID: \"fe5a4feb-297b-4c17-bc6c-cab1bf1dee41\") " pod="openstack-operators/openstack-operator-controller-manager-7ff59cf98b-bntz4" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.806213 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-metrics-certs\") pod \"openstack-operator-controller-manager-7ff59cf98b-bntz4\" (UID: \"fe5a4feb-297b-4c17-bc6c-cab1bf1dee41\") " pod="openstack-operators/openstack-operator-controller-manager-7ff59cf98b-bntz4" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.806289 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fcvd\" (UniqueName: \"kubernetes.io/projected/1010adc0-3454-486d-a676-80d329745d4c-kube-api-access-5fcvd\") pod \"rabbitmq-cluster-operator-manager-668c99d594-qf2g9\" (UID: \"1010adc0-3454-486d-a676-80d329745d4c\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qf2g9" Feb 27 17:18:11 crc kubenswrapper[4700]: E0227 17:18:11.806836 4700 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 27 17:18:11 crc kubenswrapper[4700]: E0227 17:18:11.806903 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-webhook-certs podName:fe5a4feb-297b-4c17-bc6c-cab1bf1dee41 nodeName:}" failed. No retries permitted until 2026-02-27 17:18:12.306884118 +0000 UTC m=+1052.292196855 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-webhook-certs") pod "openstack-operator-controller-manager-7ff59cf98b-bntz4" (UID: "fe5a4feb-297b-4c17-bc6c-cab1bf1dee41") : secret "webhook-server-cert" not found Feb 27 17:18:11 crc kubenswrapper[4700]: E0227 17:18:11.807143 4700 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 27 17:18:11 crc kubenswrapper[4700]: E0227 17:18:11.807168 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-metrics-certs podName:fe5a4feb-297b-4c17-bc6c-cab1bf1dee41 nodeName:}" failed. No retries permitted until 2026-02-27 17:18:12.307159835 +0000 UTC m=+1052.292472582 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-metrics-certs") pod "openstack-operator-controller-manager-7ff59cf98b-bntz4" (UID: "fe5a4feb-297b-4c17-bc6c-cab1bf1dee41") : secret "metrics-server-cert" not found Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.847340 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fcvd\" (UniqueName: \"kubernetes.io/projected/1010adc0-3454-486d-a676-80d329745d4c-kube-api-access-5fcvd\") pod \"rabbitmq-cluster-operator-manager-668c99d594-qf2g9\" (UID: \"1010adc0-3454-486d-a676-80d329745d4c\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qf2g9" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.848277 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82s8m\" (UniqueName: \"kubernetes.io/projected/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-kube-api-access-82s8m\") pod \"openstack-operator-controller-manager-7ff59cf98b-bntz4\" (UID: \"fe5a4feb-297b-4c17-bc6c-cab1bf1dee41\") " pod="openstack-operators/openstack-operator-controller-manager-7ff59cf98b-bntz4" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.885526 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qf2g9" Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.977270 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6db6876945-nmv94"] Feb 27 17:18:11 crc kubenswrapper[4700]: I0227 17:18:11.985866 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9b9ff9f4d-cb7w5" Feb 27 17:18:12 crc kubenswrapper[4700]: I0227 17:18:12.010032 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f6a3d37c-2b18-40bc-a990-56dc67093abf-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg\" (UID: \"f6a3d37c-2b18-40bc-a990-56dc67093abf\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg" Feb 27 17:18:12 crc kubenswrapper[4700]: E0227 17:18:12.010386 4700 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 27 17:18:12 crc kubenswrapper[4700]: E0227 17:18:12.010472 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f6a3d37c-2b18-40bc-a990-56dc67093abf-cert podName:f6a3d37c-2b18-40bc-a990-56dc67093abf nodeName:}" failed. No retries permitted until 2026-02-27 17:18:13.010438328 +0000 UTC m=+1052.995751075 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f6a3d37c-2b18-40bc-a990-56dc67093abf-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg" (UID: "f6a3d37c-2b18-40bc-a990-56dc67093abf") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 27 17:18:12 crc kubenswrapper[4700]: W0227 17:18:12.093652 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37c36ebd_e968_4024_a8c5_77c6507b95cf.slice/crio-ec7817f435cf082f74753a9a7a9be3fcaba36f3cda1b2084c36c5eaecf7349c2 WatchSource:0}: Error finding container ec7817f435cf082f74753a9a7a9be3fcaba36f3cda1b2084c36c5eaecf7349c2: Status 404 returned error can't find the container with id ec7817f435cf082f74753a9a7a9be3fcaba36f3cda1b2084c36c5eaecf7349c2 Feb 27 17:18:12 crc kubenswrapper[4700]: I0227 17:18:12.324435 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-metrics-certs\") pod \"openstack-operator-controller-manager-7ff59cf98b-bntz4\" (UID: \"fe5a4feb-297b-4c17-bc6c-cab1bf1dee41\") " pod="openstack-operators/openstack-operator-controller-manager-7ff59cf98b-bntz4" Feb 27 17:18:12 crc kubenswrapper[4700]: I0227 17:18:12.324773 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-webhook-certs\") pod \"openstack-operator-controller-manager-7ff59cf98b-bntz4\" (UID: \"fe5a4feb-297b-4c17-bc6c-cab1bf1dee41\") " pod="openstack-operators/openstack-operator-controller-manager-7ff59cf98b-bntz4" Feb 27 17:18:12 crc kubenswrapper[4700]: E0227 17:18:12.324883 4700 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 27 17:18:12 crc kubenswrapper[4700]: E0227 17:18:12.324887 4700 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 27 17:18:12 crc kubenswrapper[4700]: E0227 17:18:12.324929 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-webhook-certs podName:fe5a4feb-297b-4c17-bc6c-cab1bf1dee41 nodeName:}" failed. No retries permitted until 2026-02-27 17:18:13.32491651 +0000 UTC m=+1053.310229257 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-webhook-certs") pod "openstack-operator-controller-manager-7ff59cf98b-bntz4" (UID: "fe5a4feb-297b-4c17-bc6c-cab1bf1dee41") : secret "webhook-server-cert" not found Feb 27 17:18:12 crc kubenswrapper[4700]: E0227 17:18:12.324957 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-metrics-certs podName:fe5a4feb-297b-4c17-bc6c-cab1bf1dee41 nodeName:}" failed. No retries permitted until 2026-02-27 17:18:13.32493609 +0000 UTC m=+1053.310248837 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-metrics-certs") pod "openstack-operator-controller-manager-7ff59cf98b-bntz4" (UID: "fe5a4feb-297b-4c17-bc6c-cab1bf1dee41") : secret "metrics-server-cert" not found Feb 27 17:18:12 crc kubenswrapper[4700]: I0227 17:18:12.674755 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6db6876945-nmv94" event={"ID":"37c36ebd-e968-4024-a8c5-77c6507b95cf","Type":"ContainerStarted","Data":"ec7817f435cf082f74753a9a7a9be3fcaba36f3cda1b2084c36c5eaecf7349c2"} Feb 27 17:18:12 crc kubenswrapper[4700]: I0227 17:18:12.729662 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-67d996989d-f625b"] Feb 27 17:18:12 crc kubenswrapper[4700]: I0227 17:18:12.730997 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/56bc1032-efb2-4480-9091-11e529096a3c-cert\") pod \"infra-operator-controller-manager-f7fcc58b9-84zw4\" (UID: \"56bc1032-efb2-4480-9091-11e529096a3c\") " pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-84zw4" Feb 27 17:18:12 crc kubenswrapper[4700]: E0227 17:18:12.731159 4700 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 27 17:18:12 crc kubenswrapper[4700]: E0227 17:18:12.731209 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/56bc1032-efb2-4480-9091-11e529096a3c-cert podName:56bc1032-efb2-4480-9091-11e529096a3c nodeName:}" failed. No retries permitted until 2026-02-27 17:18:14.731194927 +0000 UTC m=+1054.716507674 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/56bc1032-efb2-4480-9091-11e529096a3c-cert") pod "infra-operator-controller-manager-f7fcc58b9-84zw4" (UID: "56bc1032-efb2-4480-9091-11e529096a3c") : secret "infra-operator-webhook-server-cert" not found Feb 27 17:18:12 crc kubenswrapper[4700]: I0227 17:18:12.747336 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-54688575f-8vhl9"] Feb 27 17:18:12 crc kubenswrapper[4700]: I0227 17:18:12.755501 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-64db6967f8-l2542"] Feb 27 17:18:12 crc kubenswrapper[4700]: W0227 17:18:12.763882 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode00942b5_6f82_48d4_8413_2f568daf5d7f.slice/crio-8371351802b9989c3e5d3a2bd46fc032b95852770634e08b719f0d8893652211 WatchSource:0}: Error finding container 8371351802b9989c3e5d3a2bd46fc032b95852770634e08b719f0d8893652211: Status 404 returned error can't find the container with id 8371351802b9989c3e5d3a2bd46fc032b95852770634e08b719f0d8893652211 Feb 27 17:18:12 crc kubenswrapper[4700]: I0227 17:18:12.777729 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-dm8wj"] Feb 27 17:18:12 crc kubenswrapper[4700]: I0227 17:18:12.792396 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-556b8b874-4bmbt"] Feb 27 17:18:12 crc kubenswrapper[4700]: I0227 17:18:12.806762 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-5d87c9d997-9mxq4"] Feb 27 17:18:12 crc kubenswrapper[4700]: I0227 17:18:12.819645 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-74b6b5dc96-72l74"] Feb 27 17:18:12 crc kubenswrapper[4700]: I0227 17:18:12.828507 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-vt248"] Feb 27 17:18:12 crc kubenswrapper[4700]: I0227 17:18:12.834381 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-cf99c678f-n55xj"] Feb 27 17:18:12 crc kubenswrapper[4700]: I0227 17:18:12.841111 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55ffd4876b-q66w8"] Feb 27 17:18:12 crc kubenswrapper[4700]: I0227 17:18:12.856016 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-545456dc4-stc8v"] Feb 27 17:18:12 crc kubenswrapper[4700]: I0227 17:18:12.863174 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-75684d597f-6cfj6"] Feb 27 17:18:12 crc kubenswrapper[4700]: W0227 17:18:12.863558 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod00a21bc7_24a6_4a6d_ba48_18db1d9233c6.slice/crio-5e977290f2d16100ef6d3d99a4116a289839d42edc49d6e504f36ab112185d1c WatchSource:0}: Error finding container 5e977290f2d16100ef6d3d99a4116a289839d42edc49d6e504f36ab112185d1c: Status 404 returned error can't find the container with id 5e977290f2d16100ef6d3d99a4116a289839d42edc49d6e504f36ab112185d1c Feb 27 17:18:12 crc kubenswrapper[4700]: W0227 17:18:12.865654 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9c1db815_93de_4bcc_9698_22bf6588ac88.slice/crio-f8fbac1ff2e7dcdb1f79630795a7f4c6cbbda254c550f70079ded791d91b1006 WatchSource:0}: Error finding container f8fbac1ff2e7dcdb1f79630795a7f4c6cbbda254c550f70079ded791d91b1006: Status 404 returned error can't find the container with id f8fbac1ff2e7dcdb1f79630795a7f4c6cbbda254c550f70079ded791d91b1006 Feb 27 17:18:12 crc kubenswrapper[4700]: E0227 17:18:12.873609 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:ee642fcf655f9897d480460008cba2e98b497d3ffdf7ab1d48ea460eb20c2053,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fvksb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-cf99c678f-n55xj_openstack-operators(ab00dea0-59fa-49ea-92f9-a230da6ea536): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 27 17:18:12 crc kubenswrapper[4700]: E0227 17:18:12.874790 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/heat-operator-controller-manager-cf99c678f-n55xj" podUID="ab00dea0-59fa-49ea-92f9-a230da6ea536" Feb 27 17:18:12 crc kubenswrapper[4700]: E0227 17:18:12.876065 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:e41dfadd2c3bbcae29f8c43cd2feea6724a48cdef127d65d1d37816bb9945a01,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zbmtm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-545456dc4-stc8v_openstack-operators(9c1db815-93de-4bcc-9698-22bf6588ac88): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 27 17:18:12 crc kubenswrapper[4700]: E0227 17:18:12.877203 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ironic-operator-controller-manager-545456dc4-stc8v" podUID="9c1db815-93de-4bcc-9698-22bf6588ac88" Feb 27 17:18:12 crc kubenswrapper[4700]: I0227 17:18:12.941537 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-55d77d7b5c-dc72n"] Feb 27 17:18:12 crc kubenswrapper[4700]: W0227 17:18:12.942719 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddcdee90e_b946_4cf5_807c_6babcd83071f.slice/crio-97d3cfda54be05ed204ac403e9114fb10774eec72563b0bfff9afd4af4ebed7a WatchSource:0}: Error finding container 97d3cfda54be05ed204ac403e9114fb10774eec72563b0bfff9afd4af4ebed7a: Status 404 returned error can't find the container with id 97d3cfda54be05ed204ac403e9114fb10774eec72563b0bfff9afd4af4ebed7a Feb 27 17:18:12 crc kubenswrapper[4700]: E0227 17:18:12.947687 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:7961c67cfc87de69055f8330771af625f73d857426c4bb17ebb888ead843fff3,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nfsbf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-55d77d7b5c-dc72n_openstack-operators(dcdee90e-b946-4cf5-807c-6babcd83071f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 27 17:18:12 crc kubenswrapper[4700]: E0227 17:18:12.948888 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-dc72n" podUID="dcdee90e-b946-4cf5-807c-6babcd83071f" Feb 27 17:18:12 crc kubenswrapper[4700]: I0227 17:18:12.989939 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e5bcb83-503a-4684-9e4c-3d3c9badbae5" path="/var/lib/kubelet/pods/6e5bcb83-503a-4684-9e4c-3d3c9badbae5/volumes" Feb 27 17:18:13 crc kubenswrapper[4700]: I0227 17:18:13.034982 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f6a3d37c-2b18-40bc-a990-56dc67093abf-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg\" (UID: \"f6a3d37c-2b18-40bc-a990-56dc67093abf\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg" Feb 27 17:18:13 crc kubenswrapper[4700]: E0227 17:18:13.035146 4700 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 27 17:18:13 crc kubenswrapper[4700]: E0227 17:18:13.035229 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f6a3d37c-2b18-40bc-a990-56dc67093abf-cert podName:f6a3d37c-2b18-40bc-a990-56dc67093abf nodeName:}" failed. No retries permitted until 2026-02-27 17:18:15.035205182 +0000 UTC m=+1055.020517929 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f6a3d37c-2b18-40bc-a990-56dc67093abf-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg" (UID: "f6a3d37c-2b18-40bc-a990-56dc67093abf") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 27 17:18:13 crc kubenswrapper[4700]: I0227 17:18:13.097196 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5fdb694969-44djz"] Feb 27 17:18:13 crc kubenswrapper[4700]: W0227 17:18:13.124530 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod629a3ab5_c8ce_49cb_bd57_355d643253c5.slice/crio-a122ba748fe1d4f048547493f50f291c252006fa4fac11c50881979d659bc220 WatchSource:0}: Error finding container a122ba748fe1d4f048547493f50f291c252006fa4fac11c50881979d659bc220: Status 404 returned error can't find the container with id a122ba748fe1d4f048547493f50f291c252006fa4fac11c50881979d659bc220 Feb 27 17:18:13 crc kubenswrapper[4700]: I0227 17:18:13.173430 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-8f7484475-8pdn6"] Feb 27 17:18:13 crc kubenswrapper[4700]: W0227 17:18:13.177290 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d94c9a7_4f4e_4370_a3da_9d80309f327a.slice/crio-713e371a675ef04a9107e5863e98fac6ca391e0cdd4cca19ca73ca4b461a1db5 WatchSource:0}: Error finding container 713e371a675ef04a9107e5863e98fac6ca391e0cdd4cca19ca73ca4b461a1db5: Status 404 returned error can't find the container with id 713e371a675ef04a9107e5863e98fac6ca391e0cdd4cca19ca73ca4b461a1db5 Feb 27 17:18:13 crc kubenswrapper[4700]: E0227 17:18:13.178688 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.145:5001/openstack-k8s-operators/watcher-operator:eaf82eeed7c641cca4b0e467ff9bfd7468ff8986,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mgcdc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-8f7484475-8pdn6_openstack-operators(33fd8eaf-c78f-4949-be53-86d8981ca296): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 27 17:18:13 crc kubenswrapper[4700]: E0227 17:18:13.179951 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-8f7484475-8pdn6" podUID="33fd8eaf-c78f-4949-be53-86d8981ca296" Feb 27 17:18:13 crc kubenswrapper[4700]: E0227 17:18:13.181170 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:9d03f03aa9a460f1fcac8875064808c03e4ecd0388873bbfb9c7dc58331f3968,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hmv6w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-55b5ff4dbb-xgf2x_openstack-operators(269a9c0b-7242-44dc-b860-386757217052): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 27 17:18:13 crc kubenswrapper[4700]: E0227 17:18:13.182618 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-xgf2x" podUID="269a9c0b-7242-44dc-b860-386757217052" Feb 27 17:18:13 crc kubenswrapper[4700]: E0227 17:18:13.194162 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:bb939885bd04593ad03af901adb77ee2a2d18529b328c23288c7cc7a2ba5282e,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7m8nj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-648564c9fc-dmd7w_openstack-operators(5d94c9a7-4f4e-4370-a3da-9d80309f327a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 27 17:18:13 crc kubenswrapper[4700]: E0227 17:18:13.195943 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-648564c9fc-dmd7w" podUID="5d94c9a7-4f4e-4370-a3da-9d80309f327a" Feb 27 17:18:13 crc kubenswrapper[4700]: E0227 17:18:13.197767 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:f309cdea8084a4b1e8cbcd732d6e250fd93c55cfd1b48ba9026907c8591faab7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-w2smz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-9b9ff9f4d-cb7w5_openstack-operators(15764e3b-3a40-4742-91a4-3d339cb86a0b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 27 17:18:13 crc kubenswrapper[4700]: E0227 17:18:13.199882 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-9b9ff9f4d-cb7w5" podUID="15764e3b-3a40-4742-91a4-3d339cb86a0b" Feb 27 17:18:13 crc kubenswrapper[4700]: I0227 17:18:13.210605 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qf2g9"] Feb 27 17:18:13 crc kubenswrapper[4700]: I0227 17:18:13.222664 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-55b5ff4dbb-xgf2x"] Feb 27 17:18:13 crc kubenswrapper[4700]: I0227 17:18:13.229663 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-648564c9fc-dmd7w"] Feb 27 17:18:13 crc kubenswrapper[4700]: I0227 17:18:13.233184 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9b9ff9f4d-cb7w5"] Feb 27 17:18:13 crc kubenswrapper[4700]: I0227 17:18:13.360098 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-metrics-certs\") pod \"openstack-operator-controller-manager-7ff59cf98b-bntz4\" (UID: \"fe5a4feb-297b-4c17-bc6c-cab1bf1dee41\") " pod="openstack-operators/openstack-operator-controller-manager-7ff59cf98b-bntz4" Feb 27 17:18:13 crc kubenswrapper[4700]: I0227 17:18:13.360204 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-webhook-certs\") pod \"openstack-operator-controller-manager-7ff59cf98b-bntz4\" (UID: \"fe5a4feb-297b-4c17-bc6c-cab1bf1dee41\") " pod="openstack-operators/openstack-operator-controller-manager-7ff59cf98b-bntz4" Feb 27 17:18:13 crc kubenswrapper[4700]: E0227 17:18:13.360295 4700 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 27 17:18:13 crc kubenswrapper[4700]: E0227 17:18:13.360332 4700 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 27 17:18:13 crc kubenswrapper[4700]: E0227 17:18:13.360369 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-metrics-certs podName:fe5a4feb-297b-4c17-bc6c-cab1bf1dee41 nodeName:}" failed. No retries permitted until 2026-02-27 17:18:15.360348956 +0000 UTC m=+1055.345661703 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-metrics-certs") pod "openstack-operator-controller-manager-7ff59cf98b-bntz4" (UID: "fe5a4feb-297b-4c17-bc6c-cab1bf1dee41") : secret "metrics-server-cert" not found Feb 27 17:18:13 crc kubenswrapper[4700]: E0227 17:18:13.360386 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-webhook-certs podName:fe5a4feb-297b-4c17-bc6c-cab1bf1dee41 nodeName:}" failed. No retries permitted until 2026-02-27 17:18:15.360380917 +0000 UTC m=+1055.345693664 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-webhook-certs") pod "openstack-operator-controller-manager-7ff59cf98b-bntz4" (UID: "fe5a4feb-297b-4c17-bc6c-cab1bf1dee41") : secret "webhook-server-cert" not found Feb 27 17:18:13 crc kubenswrapper[4700]: I0227 17:18:13.687921 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-vt248" event={"ID":"bd70bcca-beb1-464b-a1a3-b154ff45c1f5","Type":"ContainerStarted","Data":"fec092a9b0ea9655ec23cd13eef3d63fb6ff0e8f1504ca4b84cb4de024f163ea"} Feb 27 17:18:13 crc kubenswrapper[4700]: I0227 17:18:13.689481 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-556b8b874-4bmbt" event={"ID":"c1fdff54-9c43-4fe4-925b-f87590c75e72","Type":"ContainerStarted","Data":"e6b5cc5bcea78d23456f64294d4783dbb3455bfce49852e898289867c62c7e96"} Feb 27 17:18:13 crc kubenswrapper[4700]: I0227 17:18:13.690483 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-dm8wj" event={"ID":"c3ac9c3b-2142-409d-8c49-1bf2837252d7","Type":"ContainerStarted","Data":"9a13046a50c17be4e56beb85dffe03c995d6467fde8658026491f205b31067a4"} Feb 27 17:18:13 crc kubenswrapper[4700]: I0227 17:18:13.691746 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-cf99c678f-n55xj" event={"ID":"ab00dea0-59fa-49ea-92f9-a230da6ea536","Type":"ContainerStarted","Data":"b745b58d037a3d8697db0ef3459f1cda8ee66f3721e7da515ab2862a20378ba9"} Feb 27 17:18:13 crc kubenswrapper[4700]: E0227 17:18:13.693303 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:ee642fcf655f9897d480460008cba2e98b497d3ffdf7ab1d48ea460eb20c2053\\\"\"" pod="openstack-operators/heat-operator-controller-manager-cf99c678f-n55xj" podUID="ab00dea0-59fa-49ea-92f9-a230da6ea536" Feb 27 17:18:13 crc kubenswrapper[4700]: I0227 17:18:13.694484 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-8f7484475-8pdn6" event={"ID":"33fd8eaf-c78f-4949-be53-86d8981ca296","Type":"ContainerStarted","Data":"e81f95fdf289d3687835159df431208fd6451655f295fcb0cbfba0959388b8d9"} Feb 27 17:18:13 crc kubenswrapper[4700]: E0227 17:18:13.696384 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.145:5001/openstack-k8s-operators/watcher-operator:eaf82eeed7c641cca4b0e467ff9bfd7468ff8986\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-8f7484475-8pdn6" podUID="33fd8eaf-c78f-4949-be53-86d8981ca296" Feb 27 17:18:13 crc kubenswrapper[4700]: I0227 17:18:13.696466 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5fdb694969-44djz" event={"ID":"629a3ab5-c8ce-49cb-bd57-355d643253c5","Type":"ContainerStarted","Data":"a122ba748fe1d4f048547493f50f291c252006fa4fac11c50881979d659bc220"} Feb 27 17:18:13 crc kubenswrapper[4700]: I0227 17:18:13.699727 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-dc72n" event={"ID":"dcdee90e-b946-4cf5-807c-6babcd83071f","Type":"ContainerStarted","Data":"97d3cfda54be05ed204ac403e9114fb10774eec72563b0bfff9afd4af4ebed7a"} Feb 27 17:18:13 crc kubenswrapper[4700]: E0227 17:18:13.700866 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/cinder-operator@sha256:7961c67cfc87de69055f8330771af625f73d857426c4bb17ebb888ead843fff3\\\"\"" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-dc72n" podUID="dcdee90e-b946-4cf5-807c-6babcd83071f" Feb 27 17:18:13 crc kubenswrapper[4700]: I0227 17:18:13.700911 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-648564c9fc-dmd7w" event={"ID":"5d94c9a7-4f4e-4370-a3da-9d80309f327a","Type":"ContainerStarted","Data":"713e371a675ef04a9107e5863e98fac6ca391e0cdd4cca19ca73ca4b461a1db5"} Feb 27 17:18:13 crc kubenswrapper[4700]: E0227 17:18:13.701975 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:bb939885bd04593ad03af901adb77ee2a2d18529b328c23288c7cc7a2ba5282e\\\"\"" pod="openstack-operators/placement-operator-controller-manager-648564c9fc-dmd7w" podUID="5d94c9a7-4f4e-4370-a3da-9d80309f327a" Feb 27 17:18:13 crc kubenswrapper[4700]: I0227 17:18:13.702389 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9b9ff9f4d-cb7w5" event={"ID":"15764e3b-3a40-4742-91a4-3d339cb86a0b","Type":"ContainerStarted","Data":"6af1c0050079b7445112ab467a203a632520fad51d0c26cf075ac2c8baf80c86"} Feb 27 17:18:13 crc kubenswrapper[4700]: E0227 17:18:13.703346 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:f309cdea8084a4b1e8cbcd732d6e250fd93c55cfd1b48ba9026907c8591faab7\\\"\"" pod="openstack-operators/swift-operator-controller-manager-9b9ff9f4d-cb7w5" podUID="15764e3b-3a40-4742-91a4-3d339cb86a0b" Feb 27 17:18:13 crc kubenswrapper[4700]: I0227 17:18:13.704227 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-55ffd4876b-q66w8" event={"ID":"7cd71afd-89a7-42b9-9913-6827ea7a22d3","Type":"ContainerStarted","Data":"797e7079f8da7e5e6a4f377958e30da3e03a793d93ed80d2a18542b4fee92908"} Feb 27 17:18:13 crc kubenswrapper[4700]: I0227 17:18:13.705977 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qf2g9" event={"ID":"1010adc0-3454-486d-a676-80d329745d4c","Type":"ContainerStarted","Data":"ce9008e373d8adc461cad769d8c4acf355039a553aeccd1bfe36bcf908fd9ae5"} Feb 27 17:18:13 crc kubenswrapper[4700]: I0227 17:18:13.713427 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-67d996989d-f625b" event={"ID":"e6d0216e-25c4-4957-8e6b-473b49dd7543","Type":"ContainerStarted","Data":"7a63031a073fda02b9d0fb90ddb7e92c2183202629f53a63275d4cf283fac9fe"} Feb 27 17:18:13 crc kubenswrapper[4700]: I0227 17:18:13.715114 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-74b6b5dc96-72l74" event={"ID":"5bbcc1bb-6070-4196-8212-cf8e04eaa923","Type":"ContainerStarted","Data":"34c3fa1c714182c98ebdc24d77c1106aa7a39ded2a2659fa8e60e7819d6695a9"} Feb 27 17:18:13 crc kubenswrapper[4700]: I0227 17:18:13.717377 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-75684d597f-6cfj6" event={"ID":"00a21bc7-24a6-4a6d-ba48-18db1d9233c6","Type":"ContainerStarted","Data":"5e977290f2d16100ef6d3d99a4116a289839d42edc49d6e504f36ab112185d1c"} Feb 27 17:18:13 crc kubenswrapper[4700]: I0227 17:18:13.720799 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-xgf2x" event={"ID":"269a9c0b-7242-44dc-b860-386757217052","Type":"ContainerStarted","Data":"f27e651806ffed59b8631e349ae007a609d24d3ecf494f58086d1c725d840b61"} Feb 27 17:18:13 crc kubenswrapper[4700]: E0227 17:18:13.723620 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:9d03f03aa9a460f1fcac8875064808c03e4ecd0388873bbfb9c7dc58331f3968\\\"\"" pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-xgf2x" podUID="269a9c0b-7242-44dc-b860-386757217052" Feb 27 17:18:13 crc kubenswrapper[4700]: I0227 17:18:13.724805 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-64db6967f8-l2542" event={"ID":"181557a9-fcb7-4554-82ae-d3446b460867","Type":"ContainerStarted","Data":"4452f726e942600e7d93b00bb97f3480180c06329cb01eedd054bb3c743875d6"} Feb 27 17:18:13 crc kubenswrapper[4700]: I0227 17:18:13.726500 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-54688575f-8vhl9" event={"ID":"e00942b5-6f82-48d4-8413-2f568daf5d7f","Type":"ContainerStarted","Data":"8371351802b9989c3e5d3a2bd46fc032b95852770634e08b719f0d8893652211"} Feb 27 17:18:13 crc kubenswrapper[4700]: I0227 17:18:13.729706 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-545456dc4-stc8v" event={"ID":"9c1db815-93de-4bcc-9698-22bf6588ac88","Type":"ContainerStarted","Data":"f8fbac1ff2e7dcdb1f79630795a7f4c6cbbda254c550f70079ded791d91b1006"} Feb 27 17:18:13 crc kubenswrapper[4700]: E0227 17:18:13.731590 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:e41dfadd2c3bbcae29f8c43cd2feea6724a48cdef127d65d1d37816bb9945a01\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-545456dc4-stc8v" podUID="9c1db815-93de-4bcc-9698-22bf6588ac88" Feb 27 17:18:13 crc kubenswrapper[4700]: I0227 17:18:13.743938 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-5d87c9d997-9mxq4" event={"ID":"9b575398-8936-48f1-a412-ec26c21303cf","Type":"ContainerStarted","Data":"a6f6eed34ad29057bf509ce2a53ae01de7f4eafd37e1dd7490a7c33707efb2db"} Feb 27 17:18:14 crc kubenswrapper[4700]: E0227 17:18:14.751115 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/cinder-operator@sha256:7961c67cfc87de69055f8330771af625f73d857426c4bb17ebb888ead843fff3\\\"\"" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-dc72n" podUID="dcdee90e-b946-4cf5-807c-6babcd83071f" Feb 27 17:18:14 crc kubenswrapper[4700]: E0227 17:18:14.751837 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.145:5001/openstack-k8s-operators/watcher-operator:eaf82eeed7c641cca4b0e467ff9bfd7468ff8986\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-8f7484475-8pdn6" podUID="33fd8eaf-c78f-4949-be53-86d8981ca296" Feb 27 17:18:14 crc kubenswrapper[4700]: E0227 17:18:14.751841 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:f309cdea8084a4b1e8cbcd732d6e250fd93c55cfd1b48ba9026907c8591faab7\\\"\"" pod="openstack-operators/swift-operator-controller-manager-9b9ff9f4d-cb7w5" podUID="15764e3b-3a40-4742-91a4-3d339cb86a0b" Feb 27 17:18:14 crc kubenswrapper[4700]: E0227 17:18:14.752269 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:9d03f03aa9a460f1fcac8875064808c03e4ecd0388873bbfb9c7dc58331f3968\\\"\"" pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-xgf2x" podUID="269a9c0b-7242-44dc-b860-386757217052" Feb 27 17:18:14 crc kubenswrapper[4700]: E0227 17:18:14.752695 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:bb939885bd04593ad03af901adb77ee2a2d18529b328c23288c7cc7a2ba5282e\\\"\"" pod="openstack-operators/placement-operator-controller-manager-648564c9fc-dmd7w" podUID="5d94c9a7-4f4e-4370-a3da-9d80309f327a" Feb 27 17:18:14 crc kubenswrapper[4700]: E0227 17:18:14.752793 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:ee642fcf655f9897d480460008cba2e98b497d3ffdf7ab1d48ea460eb20c2053\\\"\"" pod="openstack-operators/heat-operator-controller-manager-cf99c678f-n55xj" podUID="ab00dea0-59fa-49ea-92f9-a230da6ea536" Feb 27 17:18:14 crc kubenswrapper[4700]: E0227 17:18:14.752953 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:e41dfadd2c3bbcae29f8c43cd2feea6724a48cdef127d65d1d37816bb9945a01\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-545456dc4-stc8v" podUID="9c1db815-93de-4bcc-9698-22bf6588ac88" Feb 27 17:18:14 crc kubenswrapper[4700]: I0227 17:18:14.788572 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/56bc1032-efb2-4480-9091-11e529096a3c-cert\") pod \"infra-operator-controller-manager-f7fcc58b9-84zw4\" (UID: \"56bc1032-efb2-4480-9091-11e529096a3c\") " pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-84zw4" Feb 27 17:18:14 crc kubenswrapper[4700]: E0227 17:18:14.788736 4700 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 27 17:18:14 crc kubenswrapper[4700]: E0227 17:18:14.788790 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/56bc1032-efb2-4480-9091-11e529096a3c-cert podName:56bc1032-efb2-4480-9091-11e529096a3c nodeName:}" failed. No retries permitted until 2026-02-27 17:18:18.788775538 +0000 UTC m=+1058.774088285 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/56bc1032-efb2-4480-9091-11e529096a3c-cert") pod "infra-operator-controller-manager-f7fcc58b9-84zw4" (UID: "56bc1032-efb2-4480-9091-11e529096a3c") : secret "infra-operator-webhook-server-cert" not found Feb 27 17:18:15 crc kubenswrapper[4700]: I0227 17:18:15.092504 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f6a3d37c-2b18-40bc-a990-56dc67093abf-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg\" (UID: \"f6a3d37c-2b18-40bc-a990-56dc67093abf\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg" Feb 27 17:18:15 crc kubenswrapper[4700]: E0227 17:18:15.094707 4700 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 27 17:18:15 crc kubenswrapper[4700]: E0227 17:18:15.094793 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f6a3d37c-2b18-40bc-a990-56dc67093abf-cert podName:f6a3d37c-2b18-40bc-a990-56dc67093abf nodeName:}" failed. No retries permitted until 2026-02-27 17:18:19.094773666 +0000 UTC m=+1059.080086413 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f6a3d37c-2b18-40bc-a990-56dc67093abf-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg" (UID: "f6a3d37c-2b18-40bc-a990-56dc67093abf") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 27 17:18:15 crc kubenswrapper[4700]: I0227 17:18:15.398952 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-metrics-certs\") pod \"openstack-operator-controller-manager-7ff59cf98b-bntz4\" (UID: \"fe5a4feb-297b-4c17-bc6c-cab1bf1dee41\") " pod="openstack-operators/openstack-operator-controller-manager-7ff59cf98b-bntz4" Feb 27 17:18:15 crc kubenswrapper[4700]: E0227 17:18:15.399247 4700 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 27 17:18:15 crc kubenswrapper[4700]: E0227 17:18:15.399319 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-metrics-certs podName:fe5a4feb-297b-4c17-bc6c-cab1bf1dee41 nodeName:}" failed. No retries permitted until 2026-02-27 17:18:19.399301034 +0000 UTC m=+1059.384613781 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-metrics-certs") pod "openstack-operator-controller-manager-7ff59cf98b-bntz4" (UID: "fe5a4feb-297b-4c17-bc6c-cab1bf1dee41") : secret "metrics-server-cert" not found Feb 27 17:18:15 crc kubenswrapper[4700]: I0227 17:18:15.399347 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-webhook-certs\") pod \"openstack-operator-controller-manager-7ff59cf98b-bntz4\" (UID: \"fe5a4feb-297b-4c17-bc6c-cab1bf1dee41\") " pod="openstack-operators/openstack-operator-controller-manager-7ff59cf98b-bntz4" Feb 27 17:18:15 crc kubenswrapper[4700]: E0227 17:18:15.399553 4700 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 27 17:18:15 crc kubenswrapper[4700]: E0227 17:18:15.399576 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-webhook-certs podName:fe5a4feb-297b-4c17-bc6c-cab1bf1dee41 nodeName:}" failed. No retries permitted until 2026-02-27 17:18:19.399569401 +0000 UTC m=+1059.384882148 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-webhook-certs") pod "openstack-operator-controller-manager-7ff59cf98b-bntz4" (UID: "fe5a4feb-297b-4c17-bc6c-cab1bf1dee41") : secret "webhook-server-cert" not found Feb 27 17:18:18 crc kubenswrapper[4700]: I0227 17:18:18.084652 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qmhw5" Feb 27 17:18:18 crc kubenswrapper[4700]: I0227 17:18:18.085581 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qmhw5" Feb 27 17:18:18 crc kubenswrapper[4700]: I0227 17:18:18.140993 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qmhw5" Feb 27 17:18:18 crc kubenswrapper[4700]: I0227 17:18:18.818313 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qmhw5" Feb 27 17:18:18 crc kubenswrapper[4700]: I0227 17:18:18.853878 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/56bc1032-efb2-4480-9091-11e529096a3c-cert\") pod \"infra-operator-controller-manager-f7fcc58b9-84zw4\" (UID: \"56bc1032-efb2-4480-9091-11e529096a3c\") " pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-84zw4" Feb 27 17:18:18 crc kubenswrapper[4700]: E0227 17:18:18.854118 4700 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 27 17:18:18 crc kubenswrapper[4700]: E0227 17:18:18.854162 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/56bc1032-efb2-4480-9091-11e529096a3c-cert podName:56bc1032-efb2-4480-9091-11e529096a3c nodeName:}" failed. No retries permitted until 2026-02-27 17:18:26.854149561 +0000 UTC m=+1066.839462308 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/56bc1032-efb2-4480-9091-11e529096a3c-cert") pod "infra-operator-controller-manager-f7fcc58b9-84zw4" (UID: "56bc1032-efb2-4480-9091-11e529096a3c") : secret "infra-operator-webhook-server-cert" not found Feb 27 17:18:18 crc kubenswrapper[4700]: I0227 17:18:18.882526 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qmhw5"] Feb 27 17:18:18 crc kubenswrapper[4700]: I0227 17:18:18.924733 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qjw68"] Feb 27 17:18:18 crc kubenswrapper[4700]: I0227 17:18:18.925038 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qjw68" podUID="be7f5a17-2d75-446f-90ed-b5802f57b07a" containerName="registry-server" containerID="cri-o://ec7b94744c9a0936ac2d59e837f01b7cca71f21f1aa6341c6d941c7bed387f31" gracePeriod=2 Feb 27 17:18:19 crc kubenswrapper[4700]: I0227 17:18:19.161565 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f6a3d37c-2b18-40bc-a990-56dc67093abf-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg\" (UID: \"f6a3d37c-2b18-40bc-a990-56dc67093abf\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg" Feb 27 17:18:19 crc kubenswrapper[4700]: E0227 17:18:19.161755 4700 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 27 17:18:19 crc kubenswrapper[4700]: E0227 17:18:19.161843 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f6a3d37c-2b18-40bc-a990-56dc67093abf-cert podName:f6a3d37c-2b18-40bc-a990-56dc67093abf nodeName:}" failed. No retries permitted until 2026-02-27 17:18:27.161813582 +0000 UTC m=+1067.147126329 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f6a3d37c-2b18-40bc-a990-56dc67093abf-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg" (UID: "f6a3d37c-2b18-40bc-a990-56dc67093abf") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 27 17:18:19 crc kubenswrapper[4700]: I0227 17:18:19.466742 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-webhook-certs\") pod \"openstack-operator-controller-manager-7ff59cf98b-bntz4\" (UID: \"fe5a4feb-297b-4c17-bc6c-cab1bf1dee41\") " pod="openstack-operators/openstack-operator-controller-manager-7ff59cf98b-bntz4" Feb 27 17:18:19 crc kubenswrapper[4700]: I0227 17:18:19.466821 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-metrics-certs\") pod \"openstack-operator-controller-manager-7ff59cf98b-bntz4\" (UID: \"fe5a4feb-297b-4c17-bc6c-cab1bf1dee41\") " pod="openstack-operators/openstack-operator-controller-manager-7ff59cf98b-bntz4" Feb 27 17:18:19 crc kubenswrapper[4700]: E0227 17:18:19.466959 4700 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 27 17:18:19 crc kubenswrapper[4700]: E0227 17:18:19.467007 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-metrics-certs podName:fe5a4feb-297b-4c17-bc6c-cab1bf1dee41 nodeName:}" failed. No retries permitted until 2026-02-27 17:18:27.466992288 +0000 UTC m=+1067.452305025 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-metrics-certs") pod "openstack-operator-controller-manager-7ff59cf98b-bntz4" (UID: "fe5a4feb-297b-4c17-bc6c-cab1bf1dee41") : secret "metrics-server-cert" not found Feb 27 17:18:19 crc kubenswrapper[4700]: E0227 17:18:19.467609 4700 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 27 17:18:19 crc kubenswrapper[4700]: E0227 17:18:19.467640 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-webhook-certs podName:fe5a4feb-297b-4c17-bc6c-cab1bf1dee41 nodeName:}" failed. No retries permitted until 2026-02-27 17:18:27.467632075 +0000 UTC m=+1067.452944822 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-webhook-certs") pod "openstack-operator-controller-manager-7ff59cf98b-bntz4" (UID: "fe5a4feb-297b-4c17-bc6c-cab1bf1dee41") : secret "webhook-server-cert" not found Feb 27 17:18:19 crc kubenswrapper[4700]: I0227 17:18:19.787337 4700 generic.go:334] "Generic (PLEG): container finished" podID="be7f5a17-2d75-446f-90ed-b5802f57b07a" containerID="ec7b94744c9a0936ac2d59e837f01b7cca71f21f1aa6341c6d941c7bed387f31" exitCode=0 Feb 27 17:18:19 crc kubenswrapper[4700]: I0227 17:18:19.787417 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qjw68" event={"ID":"be7f5a17-2d75-446f-90ed-b5802f57b07a","Type":"ContainerDied","Data":"ec7b94744c9a0936ac2d59e837f01b7cca71f21f1aa6341c6d941c7bed387f31"} Feb 27 17:18:20 crc kubenswrapper[4700]: I0227 17:18:20.467221 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wwl7g" Feb 27 17:18:20 crc kubenswrapper[4700]: I0227 17:18:20.505611 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wwl7g" Feb 27 17:18:22 crc kubenswrapper[4700]: I0227 17:18:22.784921 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wwl7g"] Feb 27 17:18:22 crc kubenswrapper[4700]: I0227 17:18:22.785155 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wwl7g" podUID="e548373a-d9bf-46bd-8efa-b4adf2dedc4b" containerName="registry-server" containerID="cri-o://1136a57c54af307682805b5db5c45047ff68468150674f891a45786ee0d36bff" gracePeriod=2 Feb 27 17:18:23 crc kubenswrapper[4700]: I0227 17:18:23.821562 4700 generic.go:334] "Generic (PLEG): container finished" podID="e548373a-d9bf-46bd-8efa-b4adf2dedc4b" containerID="1136a57c54af307682805b5db5c45047ff68468150674f891a45786ee0d36bff" exitCode=0 Feb 27 17:18:23 crc kubenswrapper[4700]: I0227 17:18:23.821648 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wwl7g" event={"ID":"e548373a-d9bf-46bd-8efa-b4adf2dedc4b","Type":"ContainerDied","Data":"1136a57c54af307682805b5db5c45047ff68468150674f891a45786ee0d36bff"} Feb 27 17:18:26 crc kubenswrapper[4700]: I0227 17:18:26.900970 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/56bc1032-efb2-4480-9091-11e529096a3c-cert\") pod \"infra-operator-controller-manager-f7fcc58b9-84zw4\" (UID: \"56bc1032-efb2-4480-9091-11e529096a3c\") " pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-84zw4" Feb 27 17:18:26 crc kubenswrapper[4700]: I0227 17:18:26.908596 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/56bc1032-efb2-4480-9091-11e529096a3c-cert\") pod \"infra-operator-controller-manager-f7fcc58b9-84zw4\" (UID: \"56bc1032-efb2-4480-9091-11e529096a3c\") " pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-84zw4" Feb 27 17:18:26 crc kubenswrapper[4700]: I0227 17:18:26.911923 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-84zw4" Feb 27 17:18:27 crc kubenswrapper[4700]: I0227 17:18:27.206607 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f6a3d37c-2b18-40bc-a990-56dc67093abf-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg\" (UID: \"f6a3d37c-2b18-40bc-a990-56dc67093abf\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg" Feb 27 17:18:27 crc kubenswrapper[4700]: I0227 17:18:27.212392 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f6a3d37c-2b18-40bc-a990-56dc67093abf-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg\" (UID: \"f6a3d37c-2b18-40bc-a990-56dc67093abf\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg" Feb 27 17:18:27 crc kubenswrapper[4700]: I0227 17:18:27.231292 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg" Feb 27 17:18:27 crc kubenswrapper[4700]: I0227 17:18:27.512147 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-webhook-certs\") pod \"openstack-operator-controller-manager-7ff59cf98b-bntz4\" (UID: \"fe5a4feb-297b-4c17-bc6c-cab1bf1dee41\") " pod="openstack-operators/openstack-operator-controller-manager-7ff59cf98b-bntz4" Feb 27 17:18:27 crc kubenswrapper[4700]: E0227 17:18:27.512399 4700 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 27 17:18:27 crc kubenswrapper[4700]: I0227 17:18:27.512628 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-metrics-certs\") pod \"openstack-operator-controller-manager-7ff59cf98b-bntz4\" (UID: \"fe5a4feb-297b-4c17-bc6c-cab1bf1dee41\") " pod="openstack-operators/openstack-operator-controller-manager-7ff59cf98b-bntz4" Feb 27 17:18:27 crc kubenswrapper[4700]: E0227 17:18:27.512647 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-webhook-certs podName:fe5a4feb-297b-4c17-bc6c-cab1bf1dee41 nodeName:}" failed. No retries permitted until 2026-02-27 17:18:43.512620522 +0000 UTC m=+1083.497933299 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-webhook-certs") pod "openstack-operator-controller-manager-7ff59cf98b-bntz4" (UID: "fe5a4feb-297b-4c17-bc6c-cab1bf1dee41") : secret "webhook-server-cert" not found Feb 27 17:18:27 crc kubenswrapper[4700]: E0227 17:18:27.512847 4700 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 27 17:18:27 crc kubenswrapper[4700]: E0227 17:18:27.512935 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-metrics-certs podName:fe5a4feb-297b-4c17-bc6c-cab1bf1dee41 nodeName:}" failed. No retries permitted until 2026-02-27 17:18:43.512911449 +0000 UTC m=+1083.498224226 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-metrics-certs") pod "openstack-operator-controller-manager-7ff59cf98b-bntz4" (UID: "fe5a4feb-297b-4c17-bc6c-cab1bf1dee41") : secret "metrics-server-cert" not found Feb 27 17:18:28 crc kubenswrapper[4700]: E0227 17:18:28.229671 4700 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ec7b94744c9a0936ac2d59e837f01b7cca71f21f1aa6341c6d941c7bed387f31 is running failed: container process not found" containerID="ec7b94744c9a0936ac2d59e837f01b7cca71f21f1aa6341c6d941c7bed387f31" cmd=["grpc_health_probe","-addr=:50051"] Feb 27 17:18:28 crc kubenswrapper[4700]: E0227 17:18:28.230107 4700 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ec7b94744c9a0936ac2d59e837f01b7cca71f21f1aa6341c6d941c7bed387f31 is running failed: container process not found" containerID="ec7b94744c9a0936ac2d59e837f01b7cca71f21f1aa6341c6d941c7bed387f31" cmd=["grpc_health_probe","-addr=:50051"] Feb 27 17:18:28 crc kubenswrapper[4700]: E0227 17:18:28.230394 4700 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ec7b94744c9a0936ac2d59e837f01b7cca71f21f1aa6341c6d941c7bed387f31 is running failed: container process not found" containerID="ec7b94744c9a0936ac2d59e837f01b7cca71f21f1aa6341c6d941c7bed387f31" cmd=["grpc_health_probe","-addr=:50051"] Feb 27 17:18:28 crc kubenswrapper[4700]: E0227 17:18:28.230439 4700 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ec7b94744c9a0936ac2d59e837f01b7cca71f21f1aa6341c6d941c7bed387f31 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-qjw68" podUID="be7f5a17-2d75-446f-90ed-b5802f57b07a" containerName="registry-server" Feb 27 17:18:30 crc kubenswrapper[4700]: E0227 17:18:30.425246 4700 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1136a57c54af307682805b5db5c45047ff68468150674f891a45786ee0d36bff is running failed: container process not found" containerID="1136a57c54af307682805b5db5c45047ff68468150674f891a45786ee0d36bff" cmd=["grpc_health_probe","-addr=:50051"] Feb 27 17:18:30 crc kubenswrapper[4700]: E0227 17:18:30.426384 4700 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1136a57c54af307682805b5db5c45047ff68468150674f891a45786ee0d36bff is running failed: container process not found" containerID="1136a57c54af307682805b5db5c45047ff68468150674f891a45786ee0d36bff" cmd=["grpc_health_probe","-addr=:50051"] Feb 27 17:18:30 crc kubenswrapper[4700]: E0227 17:18:30.426983 4700 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1136a57c54af307682805b5db5c45047ff68468150674f891a45786ee0d36bff is running failed: container process not found" containerID="1136a57c54af307682805b5db5c45047ff68468150674f891a45786ee0d36bff" cmd=["grpc_health_probe","-addr=:50051"] Feb 27 17:18:30 crc kubenswrapper[4700]: E0227 17:18:30.427158 4700 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1136a57c54af307682805b5db5c45047ff68468150674f891a45786ee0d36bff is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-wwl7g" podUID="e548373a-d9bf-46bd-8efa-b4adf2dedc4b" containerName="registry-server" Feb 27 17:18:31 crc kubenswrapper[4700]: E0227 17:18:31.671006 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:71f2ab3bb41d1743287a3270dd49e32192b347d8ba7353d2250cbd7e8528219b" Feb 27 17:18:31 crc kubenswrapper[4700]: E0227 17:18:31.671700 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:71f2ab3bb41d1743287a3270dd49e32192b347d8ba7353d2250cbd7e8528219b,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dm8dd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-556b8b874-4bmbt_openstack-operators(c1fdff54-9c43-4fe4-925b-f87590c75e72): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 17:18:31 crc kubenswrapper[4700]: E0227 17:18:31.673354 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-556b8b874-4bmbt" podUID="c1fdff54-9c43-4fe4-925b-f87590c75e72" Feb 27 17:18:31 crc kubenswrapper[4700]: E0227 17:18:31.887443 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:71f2ab3bb41d1743287a3270dd49e32192b347d8ba7353d2250cbd7e8528219b\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-556b8b874-4bmbt" podUID="c1fdff54-9c43-4fe4-925b-f87590c75e72" Feb 27 17:18:32 crc kubenswrapper[4700]: E0227 17:18:32.334183 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:f1158ec4d879c4646eee4323bc501eba4d377beb2ad6fbe08ed30070c441ac26" Feb 27 17:18:32 crc kubenswrapper[4700]: E0227 17:18:32.334747 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:f1158ec4d879c4646eee4323bc501eba4d377beb2ad6fbe08ed30070c441ac26,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5jc9f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-67d996989d-f625b_openstack-operators(e6d0216e-25c4-4957-8e6b-473b49dd7543): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 17:18:32 crc kubenswrapper[4700]: E0227 17:18:32.335956 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-67d996989d-f625b" podUID="e6d0216e-25c4-4957-8e6b-473b49dd7543" Feb 27 17:18:32 crc kubenswrapper[4700]: E0227 17:18:32.895081 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:f1158ec4d879c4646eee4323bc501eba4d377beb2ad6fbe08ed30070c441ac26\\\"\"" pod="openstack-operators/manila-operator-controller-manager-67d996989d-f625b" podUID="e6d0216e-25c4-4957-8e6b-473b49dd7543" Feb 27 17:18:33 crc kubenswrapper[4700]: E0227 17:18:33.080920 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/telemetry-operator@sha256:1b9074a4ce16396d8bd2d30a475fc8c2f004f75a023e3eef8950661e89c0bcc6" Feb 27 17:18:33 crc kubenswrapper[4700]: E0227 17:18:33.081134 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:1b9074a4ce16396d8bd2d30a475fc8c2f004f75a023e3eef8950661e89c0bcc6,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-q6mg7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-5fdb694969-44djz_openstack-operators(629a3ab5-c8ce-49cb-bd57-355d643253c5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 17:18:33 crc kubenswrapper[4700]: E0227 17:18:33.082365 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-5fdb694969-44djz" podUID="629a3ab5-c8ce-49cb-bd57-355d643253c5" Feb 27 17:18:33 crc kubenswrapper[4700]: E0227 17:18:33.676212 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:172f24bd4603ac3498536a8a2c8fffb07cf9113dd52bc132778ea0aa275c6b84" Feb 27 17:18:33 crc kubenswrapper[4700]: E0227 17:18:33.676377 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:172f24bd4603ac3498536a8a2c8fffb07cf9113dd52bc132778ea0aa275c6b84,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-p5kmq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-74b6b5dc96-72l74_openstack-operators(5bbcc1bb-6070-4196-8212-cf8e04eaa923): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 17:18:33 crc kubenswrapper[4700]: E0227 17:18:33.678073 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-74b6b5dc96-72l74" podUID="5bbcc1bb-6070-4196-8212-cf8e04eaa923" Feb 27 17:18:33 crc kubenswrapper[4700]: E0227 17:18:33.901443 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:172f24bd4603ac3498536a8a2c8fffb07cf9113dd52bc132778ea0aa275c6b84\\\"\"" pod="openstack-operators/nova-operator-controller-manager-74b6b5dc96-72l74" podUID="5bbcc1bb-6070-4196-8212-cf8e04eaa923" Feb 27 17:18:33 crc kubenswrapper[4700]: E0227 17:18:33.901513 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:1b9074a4ce16396d8bd2d30a475fc8c2f004f75a023e3eef8950661e89c0bcc6\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-5fdb694969-44djz" podUID="629a3ab5-c8ce-49cb-bd57-355d643253c5" Feb 27 17:18:34 crc kubenswrapper[4700]: I0227 17:18:34.772596 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qjw68" Feb 27 17:18:34 crc kubenswrapper[4700]: I0227 17:18:34.826148 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srwkf\" (UniqueName: \"kubernetes.io/projected/be7f5a17-2d75-446f-90ed-b5802f57b07a-kube-api-access-srwkf\") pod \"be7f5a17-2d75-446f-90ed-b5802f57b07a\" (UID: \"be7f5a17-2d75-446f-90ed-b5802f57b07a\") " Feb 27 17:18:34 crc kubenswrapper[4700]: I0227 17:18:34.826994 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be7f5a17-2d75-446f-90ed-b5802f57b07a-catalog-content\") pod \"be7f5a17-2d75-446f-90ed-b5802f57b07a\" (UID: \"be7f5a17-2d75-446f-90ed-b5802f57b07a\") " Feb 27 17:18:34 crc kubenswrapper[4700]: I0227 17:18:34.827141 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be7f5a17-2d75-446f-90ed-b5802f57b07a-utilities\") pod \"be7f5a17-2d75-446f-90ed-b5802f57b07a\" (UID: \"be7f5a17-2d75-446f-90ed-b5802f57b07a\") " Feb 27 17:18:34 crc kubenswrapper[4700]: I0227 17:18:34.828496 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be7f5a17-2d75-446f-90ed-b5802f57b07a-utilities" (OuterVolumeSpecName: "utilities") pod "be7f5a17-2d75-446f-90ed-b5802f57b07a" (UID: "be7f5a17-2d75-446f-90ed-b5802f57b07a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:18:34 crc kubenswrapper[4700]: I0227 17:18:34.834247 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be7f5a17-2d75-446f-90ed-b5802f57b07a-kube-api-access-srwkf" (OuterVolumeSpecName: "kube-api-access-srwkf") pod "be7f5a17-2d75-446f-90ed-b5802f57b07a" (UID: "be7f5a17-2d75-446f-90ed-b5802f57b07a"). InnerVolumeSpecName "kube-api-access-srwkf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:18:34 crc kubenswrapper[4700]: I0227 17:18:34.902655 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be7f5a17-2d75-446f-90ed-b5802f57b07a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "be7f5a17-2d75-446f-90ed-b5802f57b07a" (UID: "be7f5a17-2d75-446f-90ed-b5802f57b07a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:18:34 crc kubenswrapper[4700]: I0227 17:18:34.910104 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qjw68" event={"ID":"be7f5a17-2d75-446f-90ed-b5802f57b07a","Type":"ContainerDied","Data":"3609c6636ec2088b66d518e9299d36dce19fd54fdd030360c99d633c7bcbe5a8"} Feb 27 17:18:34 crc kubenswrapper[4700]: I0227 17:18:34.910162 4700 scope.go:117] "RemoveContainer" containerID="ec7b94744c9a0936ac2d59e837f01b7cca71f21f1aa6341c6d941c7bed387f31" Feb 27 17:18:34 crc kubenswrapper[4700]: I0227 17:18:34.910178 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qjw68" Feb 27 17:18:34 crc kubenswrapper[4700]: I0227 17:18:34.928334 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be7f5a17-2d75-446f-90ed-b5802f57b07a-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:18:34 crc kubenswrapper[4700]: I0227 17:18:34.928363 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be7f5a17-2d75-446f-90ed-b5802f57b07a-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:18:34 crc kubenswrapper[4700]: I0227 17:18:34.928377 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srwkf\" (UniqueName: \"kubernetes.io/projected/be7f5a17-2d75-446f-90ed-b5802f57b07a-kube-api-access-srwkf\") on node \"crc\" DevicePath \"\"" Feb 27 17:18:34 crc kubenswrapper[4700]: I0227 17:18:34.948190 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qjw68"] Feb 27 17:18:34 crc kubenswrapper[4700]: I0227 17:18:34.954713 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qjw68"] Feb 27 17:18:34 crc kubenswrapper[4700]: I0227 17:18:34.993056 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be7f5a17-2d75-446f-90ed-b5802f57b07a" path="/var/lib/kubelet/pods/be7f5a17-2d75-446f-90ed-b5802f57b07a/volumes" Feb 27 17:18:35 crc kubenswrapper[4700]: E0227 17:18:35.423623 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:b242403a27609ac87a0ed3a7dd788aceaf8f3da3620981cf5e000d56862d77a4" Feb 27 17:18:35 crc kubenswrapper[4700]: E0227 17:18:35.423771 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:b242403a27609ac87a0ed3a7dd788aceaf8f3da3620981cf5e000d56862d77a4,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hqbn7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-54688575f-8vhl9_openstack-operators(e00942b5-6f82-48d4-8413-2f568daf5d7f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 17:18:35 crc kubenswrapper[4700]: E0227 17:18:35.425307 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-54688575f-8vhl9" podUID="e00942b5-6f82-48d4-8413-2f568daf5d7f" Feb 27 17:18:35 crc kubenswrapper[4700]: E0227 17:18:35.922835 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:b242403a27609ac87a0ed3a7dd788aceaf8f3da3620981cf5e000d56862d77a4\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-54688575f-8vhl9" podUID="e00942b5-6f82-48d4-8413-2f568daf5d7f" Feb 27 17:18:35 crc kubenswrapper[4700]: E0227 17:18:35.983404 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:12fa31d2a2dfe1a832c6a2c0eb58876a3a62595a1a1f49b13c2a1f9b6d378735" Feb 27 17:18:35 crc kubenswrapper[4700]: E0227 17:18:35.983827 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:12fa31d2a2dfe1a832c6a2c0eb58876a3a62595a1a1f49b13c2a1f9b6d378735,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n7ngq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-55ffd4876b-q66w8_openstack-operators(7cd71afd-89a7-42b9-9913-6827ea7a22d3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 17:18:35 crc kubenswrapper[4700]: E0227 17:18:35.985414 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-55ffd4876b-q66w8" podUID="7cd71afd-89a7-42b9-9913-6827ea7a22d3" Feb 27 17:18:36 crc kubenswrapper[4700]: E0227 17:18:36.928828 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:12fa31d2a2dfe1a832c6a2c0eb58876a3a62595a1a1f49b13c2a1f9b6d378735\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-55ffd4876b-q66w8" podUID="7cd71afd-89a7-42b9-9913-6827ea7a22d3" Feb 27 17:18:39 crc kubenswrapper[4700]: E0227 17:18:39.729553 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Feb 27 17:18:39 crc kubenswrapper[4700]: E0227 17:18:39.730174 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5fcvd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-qf2g9_openstack-operators(1010adc0-3454-486d-a676-80d329745d4c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 17:18:39 crc kubenswrapper[4700]: E0227 17:18:39.731517 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qf2g9" podUID="1010adc0-3454-486d-a676-80d329745d4c" Feb 27 17:18:39 crc kubenswrapper[4700]: E0227 17:18:39.950014 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qf2g9" podUID="1010adc0-3454-486d-a676-80d329745d4c" Feb 27 17:18:40 crc kubenswrapper[4700]: I0227 17:18:40.285815 4700 scope.go:117] "RemoveContainer" containerID="79f033dd71797894f9ab36ca15d1f70844f19243ce596b305dc0ac7ad61fbcc9" Feb 27 17:18:40 crc kubenswrapper[4700]: I0227 17:18:40.297685 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wwl7g" Feb 27 17:18:40 crc kubenswrapper[4700]: I0227 17:18:40.410705 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e548373a-d9bf-46bd-8efa-b4adf2dedc4b-utilities\") pod \"e548373a-d9bf-46bd-8efa-b4adf2dedc4b\" (UID: \"e548373a-d9bf-46bd-8efa-b4adf2dedc4b\") " Feb 27 17:18:40 crc kubenswrapper[4700]: I0227 17:18:40.410769 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddpll\" (UniqueName: \"kubernetes.io/projected/e548373a-d9bf-46bd-8efa-b4adf2dedc4b-kube-api-access-ddpll\") pod \"e548373a-d9bf-46bd-8efa-b4adf2dedc4b\" (UID: \"e548373a-d9bf-46bd-8efa-b4adf2dedc4b\") " Feb 27 17:18:40 crc kubenswrapper[4700]: I0227 17:18:40.410793 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e548373a-d9bf-46bd-8efa-b4adf2dedc4b-catalog-content\") pod \"e548373a-d9bf-46bd-8efa-b4adf2dedc4b\" (UID: \"e548373a-d9bf-46bd-8efa-b4adf2dedc4b\") " Feb 27 17:18:40 crc kubenswrapper[4700]: I0227 17:18:40.411699 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e548373a-d9bf-46bd-8efa-b4adf2dedc4b-utilities" (OuterVolumeSpecName: "utilities") pod "e548373a-d9bf-46bd-8efa-b4adf2dedc4b" (UID: "e548373a-d9bf-46bd-8efa-b4adf2dedc4b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:18:40 crc kubenswrapper[4700]: I0227 17:18:40.417686 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e548373a-d9bf-46bd-8efa-b4adf2dedc4b-kube-api-access-ddpll" (OuterVolumeSpecName: "kube-api-access-ddpll") pod "e548373a-d9bf-46bd-8efa-b4adf2dedc4b" (UID: "e548373a-d9bf-46bd-8efa-b4adf2dedc4b"). InnerVolumeSpecName "kube-api-access-ddpll". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:18:40 crc kubenswrapper[4700]: I0227 17:18:40.438537 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e548373a-d9bf-46bd-8efa-b4adf2dedc4b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e548373a-d9bf-46bd-8efa-b4adf2dedc4b" (UID: "e548373a-d9bf-46bd-8efa-b4adf2dedc4b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:18:40 crc kubenswrapper[4700]: I0227 17:18:40.512133 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ddpll\" (UniqueName: \"kubernetes.io/projected/e548373a-d9bf-46bd-8efa-b4adf2dedc4b-kube-api-access-ddpll\") on node \"crc\" DevicePath \"\"" Feb 27 17:18:40 crc kubenswrapper[4700]: I0227 17:18:40.512163 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e548373a-d9bf-46bd-8efa-b4adf2dedc4b-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:18:40 crc kubenswrapper[4700]: I0227 17:18:40.512178 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e548373a-d9bf-46bd-8efa-b4adf2dedc4b-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:18:40 crc kubenswrapper[4700]: I0227 17:18:40.581290 4700 scope.go:117] "RemoveContainer" containerID="3fa4c7ab2a569c2126247a326cf410d76863b937c971a582179761dc4d86b276" Feb 27 17:18:40 crc kubenswrapper[4700]: I0227 17:18:40.842671 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg"] Feb 27 17:18:40 crc kubenswrapper[4700]: I0227 17:18:40.926406 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-f7fcc58b9-84zw4"] Feb 27 17:18:40 crc kubenswrapper[4700]: W0227 17:18:40.929735 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod56bc1032_efb2_4480_9091_11e529096a3c.slice/crio-94c9494470871ee8503faa9fddd2978d891ee790a6423e2fd7c9b197de472836 WatchSource:0}: Error finding container 94c9494470871ee8503faa9fddd2978d891ee790a6423e2fd7c9b197de472836: Status 404 returned error can't find the container with id 94c9494470871ee8503faa9fddd2978d891ee790a6423e2fd7c9b197de472836 Feb 27 17:18:40 crc kubenswrapper[4700]: I0227 17:18:40.958472 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9b9ff9f4d-cb7w5" event={"ID":"15764e3b-3a40-4742-91a4-3d339cb86a0b","Type":"ContainerStarted","Data":"09c6cb7729ba45ccf8d247d73edc2cc2efb40d4c25e7a4ff6d0e32b9426b4298"} Feb 27 17:18:40 crc kubenswrapper[4700]: I0227 17:18:40.958662 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-9b9ff9f4d-cb7w5" Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.009162 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-9b9ff9f4d-cb7w5" podStartSLOduration=2.760187463 podStartE2EDuration="30.009145705s" podCreationTimestamp="2026-02-27 17:18:11 +0000 UTC" firstStartedPulling="2026-02-27 17:18:13.197577878 +0000 UTC m=+1053.182890625" lastFinishedPulling="2026-02-27 17:18:40.44653612 +0000 UTC m=+1080.431848867" observedRunningTime="2026-02-27 17:18:40.986589136 +0000 UTC m=+1080.971901883" watchObservedRunningTime="2026-02-27 17:18:41.009145705 +0000 UTC m=+1080.994458452" Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.014590 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-75684d597f-6cfj6" Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.014620 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-75684d597f-6cfj6" event={"ID":"00a21bc7-24a6-4a6d-ba48-18db1d9233c6","Type":"ContainerStarted","Data":"5910fd2a75940d588d7efce67c326ac6bb5acc14456bbbf5a5b0cd5cf441df15"} Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.018595 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6db6876945-nmv94" event={"ID":"37c36ebd-e968-4024-a8c5-77c6507b95cf","Type":"ContainerStarted","Data":"ae28c75bcd49beaad8786c9e7ea8c57cf2923daa4da93db61019d041b04a039b"} Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.019049 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-6db6876945-nmv94" Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.030769 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-545456dc4-stc8v" event={"ID":"9c1db815-93de-4bcc-9698-22bf6588ac88","Type":"ContainerStarted","Data":"4762e7fa2995139519f13446bac228e27846df633ed053edabd012a9604ef662"} Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.031446 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-545456dc4-stc8v" Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.034593 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg" event={"ID":"f6a3d37c-2b18-40bc-a990-56dc67093abf","Type":"ContainerStarted","Data":"2829148a86ccc507274e4f4ed67118b662e321d24c5902c6f1f76b5218ab612a"} Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.037226 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-8f7484475-8pdn6" event={"ID":"33fd8eaf-c78f-4949-be53-86d8981ca296","Type":"ContainerStarted","Data":"4cda6906abcf2b9d1153f6e7dc98aefe4af958486479fb23ebcf3f397f194de3"} Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.037508 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-8f7484475-8pdn6" Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.041543 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-dc72n" event={"ID":"dcdee90e-b946-4cf5-807c-6babcd83071f","Type":"ContainerStarted","Data":"393f5c2f4209bfdc7831d339b8926f60f92b0800efbda4013311384b51813c0f"} Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.042431 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-dc72n" Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.044081 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-648564c9fc-dmd7w" event={"ID":"5d94c9a7-4f4e-4370-a3da-9d80309f327a","Type":"ContainerStarted","Data":"fb8e2103889a53f90b9db7172d47592a65eb251942e4467ad6b9c29701157e3f"} Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.044707 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-648564c9fc-dmd7w" Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.050715 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-84zw4" event={"ID":"56bc1032-efb2-4480-9091-11e529096a3c","Type":"ContainerStarted","Data":"94c9494470871ee8503faa9fddd2978d891ee790a6423e2fd7c9b197de472836"} Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.064741 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-vt248" event={"ID":"bd70bcca-beb1-464b-a1a3-b154ff45c1f5","Type":"ContainerStarted","Data":"4bcb3d4f5af1b7fd0b6949c341ab10de90340503d697ebf122758a1985fe3cd9"} Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.065302 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-vt248" Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.085011 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-xgf2x" event={"ID":"269a9c0b-7242-44dc-b860-386757217052","Type":"ContainerStarted","Data":"3160bda50adeb5e779a8a3dd3d559bd716f376bdb530dd70b547d38a9f57686b"} Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.085767 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-xgf2x" Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.113734 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-dm8wj" event={"ID":"c3ac9c3b-2142-409d-8c49-1bf2837252d7","Type":"ContainerStarted","Data":"4474c07d0579bed9214c4c4f8cfb8b198eba999e7869842d2b3583a1c400cbc0"} Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.114490 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-dm8wj" Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.119371 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-6db6876945-nmv94" podStartSLOduration=8.527444755 podStartE2EDuration="31.119354528s" podCreationTimestamp="2026-02-27 17:18:10 +0000 UTC" firstStartedPulling="2026-02-27 17:18:12.108658513 +0000 UTC m=+1052.093971260" lastFinishedPulling="2026-02-27 17:18:34.700568276 +0000 UTC m=+1074.685881033" observedRunningTime="2026-02-27 17:18:41.114729865 +0000 UTC m=+1081.100042612" watchObservedRunningTime="2026-02-27 17:18:41.119354528 +0000 UTC m=+1081.104667275" Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.146645 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-cf99c678f-n55xj" event={"ID":"ab00dea0-59fa-49ea-92f9-a230da6ea536","Type":"ContainerStarted","Data":"eb6f2677f1fd18306cd9cdf965ba95f95f220ec369e2306a45b87d1a554a24f2"} Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.147356 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-cf99c678f-n55xj" Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.171644 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-5d87c9d997-9mxq4" event={"ID":"9b575398-8936-48f1-a412-ec26c21303cf","Type":"ContainerStarted","Data":"09cd6f30c48dee25556329b5d515f1dc2aa4ef2eb53f600c3c7de6a76c9c336b"} Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.172365 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-5d87c9d997-9mxq4" Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.178184 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-dc72n" podStartSLOduration=3.6368324210000003 podStartE2EDuration="31.178173128s" podCreationTimestamp="2026-02-27 17:18:10 +0000 UTC" firstStartedPulling="2026-02-27 17:18:12.947569997 +0000 UTC m=+1052.932882744" lastFinishedPulling="2026-02-27 17:18:40.488910704 +0000 UTC m=+1080.474223451" observedRunningTime="2026-02-27 17:18:41.139245996 +0000 UTC m=+1081.124558753" watchObservedRunningTime="2026-02-27 17:18:41.178173128 +0000 UTC m=+1081.163485875" Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.178269 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-648564c9fc-dmd7w" podStartSLOduration=2.954440425 podStartE2EDuration="30.178266341s" podCreationTimestamp="2026-02-27 17:18:11 +0000 UTC" firstStartedPulling="2026-02-27 17:18:13.193777887 +0000 UTC m=+1053.179090634" lastFinishedPulling="2026-02-27 17:18:40.417603793 +0000 UTC m=+1080.402916550" observedRunningTime="2026-02-27 17:18:41.176936516 +0000 UTC m=+1081.162249263" watchObservedRunningTime="2026-02-27 17:18:41.178266341 +0000 UTC m=+1081.163579088" Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.195792 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wwl7g" Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.196349 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wwl7g" event={"ID":"e548373a-d9bf-46bd-8efa-b4adf2dedc4b","Type":"ContainerDied","Data":"e240c8fabd13b1b1dad1ae559c41c98d2ab7f927d48e1f1204cbcc97f5308a60"} Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.196383 4700 scope.go:117] "RemoveContainer" containerID="1136a57c54af307682805b5db5c45047ff68468150674f891a45786ee0d36bff" Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.200723 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-64db6967f8-l2542" event={"ID":"181557a9-fcb7-4554-82ae-d3446b460867","Type":"ContainerStarted","Data":"32441c5a000a18c8b5a35d66b760b4db56deaf7daca32428795d11402081515c"} Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.201332 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-64db6967f8-l2542" Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.217238 4700 scope.go:117] "RemoveContainer" containerID="d216a732af5a93b29dbe0c142cc4ac56b9a23e406c28e7b3db1c40ba7c4ca2d2" Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.229400 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-545456dc4-stc8v" podStartSLOduration=3.693498734 podStartE2EDuration="31.229384387s" podCreationTimestamp="2026-02-27 17:18:10 +0000 UTC" firstStartedPulling="2026-02-27 17:18:12.875950897 +0000 UTC m=+1052.861263644" lastFinishedPulling="2026-02-27 17:18:40.41183655 +0000 UTC m=+1080.397149297" observedRunningTime="2026-02-27 17:18:41.223880431 +0000 UTC m=+1081.209193178" watchObservedRunningTime="2026-02-27 17:18:41.229384387 +0000 UTC m=+1081.214697134" Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.252152 4700 scope.go:117] "RemoveContainer" containerID="84ba22f971daed6b03ac4db6cac8d8ed639abe496d8deb4ca197bde236d699b6" Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.277386 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-vt248" podStartSLOduration=4.805924332 podStartE2EDuration="30.27737173s" podCreationTimestamp="2026-02-27 17:18:11 +0000 UTC" firstStartedPulling="2026-02-27 17:18:12.847317458 +0000 UTC m=+1052.832630205" lastFinishedPulling="2026-02-27 17:18:38.318764856 +0000 UTC m=+1078.304077603" observedRunningTime="2026-02-27 17:18:41.249408878 +0000 UTC m=+1081.234721625" watchObservedRunningTime="2026-02-27 17:18:41.27737173 +0000 UTC m=+1081.262684477" Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.305718 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-75684d597f-6cfj6" podStartSLOduration=8.477707323 podStartE2EDuration="30.305703041s" podCreationTimestamp="2026-02-27 17:18:11 +0000 UTC" firstStartedPulling="2026-02-27 17:18:12.872768493 +0000 UTC m=+1052.858081240" lastFinishedPulling="2026-02-27 17:18:34.700764201 +0000 UTC m=+1074.686076958" observedRunningTime="2026-02-27 17:18:41.304605012 +0000 UTC m=+1081.289917759" watchObservedRunningTime="2026-02-27 17:18:41.305703041 +0000 UTC m=+1081.291015788" Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.306517 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-8f7484475-8pdn6" podStartSLOduration=2.833909179 podStartE2EDuration="30.306513103s" podCreationTimestamp="2026-02-27 17:18:11 +0000 UTC" firstStartedPulling="2026-02-27 17:18:13.178549194 +0000 UTC m=+1053.163861941" lastFinishedPulling="2026-02-27 17:18:40.651153118 +0000 UTC m=+1080.636465865" observedRunningTime="2026-02-27 17:18:41.281404517 +0000 UTC m=+1081.266717264" watchObservedRunningTime="2026-02-27 17:18:41.306513103 +0000 UTC m=+1081.291825850" Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.343486 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wwl7g"] Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.347805 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wwl7g"] Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.360854 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-5d87c9d997-9mxq4" podStartSLOduration=8.20909862 podStartE2EDuration="31.360838244s" podCreationTimestamp="2026-02-27 17:18:10 +0000 UTC" firstStartedPulling="2026-02-27 17:18:12.813298955 +0000 UTC m=+1052.798611702" lastFinishedPulling="2026-02-27 17:18:35.965038569 +0000 UTC m=+1075.950351326" observedRunningTime="2026-02-27 17:18:41.357707631 +0000 UTC m=+1081.343020378" watchObservedRunningTime="2026-02-27 17:18:41.360838244 +0000 UTC m=+1081.346150991" Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.383136 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-64db6967f8-l2542" podStartSLOduration=8.73158345 podStartE2EDuration="31.383117965s" podCreationTimestamp="2026-02-27 17:18:10 +0000 UTC" firstStartedPulling="2026-02-27 17:18:12.768025024 +0000 UTC m=+1052.753337771" lastFinishedPulling="2026-02-27 17:18:35.419559529 +0000 UTC m=+1075.404872286" observedRunningTime="2026-02-27 17:18:41.38254315 +0000 UTC m=+1081.367855887" watchObservedRunningTime="2026-02-27 17:18:41.383117965 +0000 UTC m=+1081.368430712" Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.411000 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-dm8wj" podStartSLOduration=8.24037077 podStartE2EDuration="31.410978894s" podCreationTimestamp="2026-02-27 17:18:10 +0000 UTC" firstStartedPulling="2026-02-27 17:18:12.805807397 +0000 UTC m=+1052.791120144" lastFinishedPulling="2026-02-27 17:18:35.976415491 +0000 UTC m=+1075.961728268" observedRunningTime="2026-02-27 17:18:41.408046656 +0000 UTC m=+1081.393359403" watchObservedRunningTime="2026-02-27 17:18:41.410978894 +0000 UTC m=+1081.396291641" Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.424831 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-xgf2x" podStartSLOduration=3.188556257 podStartE2EDuration="30.424809431s" podCreationTimestamp="2026-02-27 17:18:11 +0000 UTC" firstStartedPulling="2026-02-27 17:18:13.181054251 +0000 UTC m=+1053.166366998" lastFinishedPulling="2026-02-27 17:18:40.417307415 +0000 UTC m=+1080.402620172" observedRunningTime="2026-02-27 17:18:41.421417111 +0000 UTC m=+1081.406729858" watchObservedRunningTime="2026-02-27 17:18:41.424809431 +0000 UTC m=+1081.410122178" Feb 27 17:18:41 crc kubenswrapper[4700]: I0227 17:18:41.441446 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-cf99c678f-n55xj" podStartSLOduration=3.8972615900000003 podStartE2EDuration="31.441432812s" podCreationTimestamp="2026-02-27 17:18:10 +0000 UTC" firstStartedPulling="2026-02-27 17:18:12.873452361 +0000 UTC m=+1052.858765108" lastFinishedPulling="2026-02-27 17:18:40.417623543 +0000 UTC m=+1080.402936330" observedRunningTime="2026-02-27 17:18:41.437850787 +0000 UTC m=+1081.423163534" watchObservedRunningTime="2026-02-27 17:18:41.441432812 +0000 UTC m=+1081.426745559" Feb 27 17:18:42 crc kubenswrapper[4700]: I0227 17:18:42.998191 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e548373a-d9bf-46bd-8efa-b4adf2dedc4b" path="/var/lib/kubelet/pods/e548373a-d9bf-46bd-8efa-b4adf2dedc4b/volumes" Feb 27 17:18:43 crc kubenswrapper[4700]: I0227 17:18:43.552044 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-webhook-certs\") pod \"openstack-operator-controller-manager-7ff59cf98b-bntz4\" (UID: \"fe5a4feb-297b-4c17-bc6c-cab1bf1dee41\") " pod="openstack-operators/openstack-operator-controller-manager-7ff59cf98b-bntz4" Feb 27 17:18:43 crc kubenswrapper[4700]: I0227 17:18:43.552116 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-metrics-certs\") pod \"openstack-operator-controller-manager-7ff59cf98b-bntz4\" (UID: \"fe5a4feb-297b-4c17-bc6c-cab1bf1dee41\") " pod="openstack-operators/openstack-operator-controller-manager-7ff59cf98b-bntz4" Feb 27 17:18:43 crc kubenswrapper[4700]: I0227 17:18:43.558913 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-metrics-certs\") pod \"openstack-operator-controller-manager-7ff59cf98b-bntz4\" (UID: \"fe5a4feb-297b-4c17-bc6c-cab1bf1dee41\") " pod="openstack-operators/openstack-operator-controller-manager-7ff59cf98b-bntz4" Feb 27 17:18:43 crc kubenswrapper[4700]: I0227 17:18:43.559660 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/fe5a4feb-297b-4c17-bc6c-cab1bf1dee41-webhook-certs\") pod \"openstack-operator-controller-manager-7ff59cf98b-bntz4\" (UID: \"fe5a4feb-297b-4c17-bc6c-cab1bf1dee41\") " pod="openstack-operators/openstack-operator-controller-manager-7ff59cf98b-bntz4" Feb 27 17:18:43 crc kubenswrapper[4700]: I0227 17:18:43.675673 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-nb726" Feb 27 17:18:43 crc kubenswrapper[4700]: I0227 17:18:43.684318 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7ff59cf98b-bntz4" Feb 27 17:18:44 crc kubenswrapper[4700]: I0227 17:18:44.219755 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg" event={"ID":"f6a3d37c-2b18-40bc-a990-56dc67093abf","Type":"ContainerStarted","Data":"9b3bda1dc90f49c9ff580eb83c7a8226a2c585ebb5d1e3050e741877bb0048bd"} Feb 27 17:18:44 crc kubenswrapper[4700]: I0227 17:18:44.220637 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg" Feb 27 17:18:44 crc kubenswrapper[4700]: I0227 17:18:44.222336 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-84zw4" event={"ID":"56bc1032-efb2-4480-9091-11e529096a3c","Type":"ContainerStarted","Data":"90861a25c371a925b27e959f64767a21d52507451c6732a0a77698963f057e6c"} Feb 27 17:18:44 crc kubenswrapper[4700]: I0227 17:18:44.222520 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-84zw4" Feb 27 17:18:44 crc kubenswrapper[4700]: I0227 17:18:44.247661 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg" podStartSLOduration=30.144284969 podStartE2EDuration="33.247642772s" podCreationTimestamp="2026-02-27 17:18:11 +0000 UTC" firstStartedPulling="2026-02-27 17:18:40.860288226 +0000 UTC m=+1080.845600973" lastFinishedPulling="2026-02-27 17:18:43.963646029 +0000 UTC m=+1083.948958776" observedRunningTime="2026-02-27 17:18:44.24491157 +0000 UTC m=+1084.230224327" watchObservedRunningTime="2026-02-27 17:18:44.247642772 +0000 UTC m=+1084.232955529" Feb 27 17:18:44 crc kubenswrapper[4700]: I0227 17:18:44.266863 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-84zw4" podStartSLOduration=31.274539065 podStartE2EDuration="34.266841781s" podCreationTimestamp="2026-02-27 17:18:10 +0000 UTC" firstStartedPulling="2026-02-27 17:18:40.948040544 +0000 UTC m=+1080.933353291" lastFinishedPulling="2026-02-27 17:18:43.94034326 +0000 UTC m=+1083.925656007" observedRunningTime="2026-02-27 17:18:44.25960676 +0000 UTC m=+1084.244919527" watchObservedRunningTime="2026-02-27 17:18:44.266841781 +0000 UTC m=+1084.252154538" Feb 27 17:18:44 crc kubenswrapper[4700]: I0227 17:18:44.453888 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7ff59cf98b-bntz4"] Feb 27 17:18:44 crc kubenswrapper[4700]: W0227 17:18:44.458133 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfe5a4feb_297b_4c17_bc6c_cab1bf1dee41.slice/crio-436f338c30306f27e8e3f3ca338401281df84bff835c1a359b80c7fc1c55ee98 WatchSource:0}: Error finding container 436f338c30306f27e8e3f3ca338401281df84bff835c1a359b80c7fc1c55ee98: Status 404 returned error can't find the container with id 436f338c30306f27e8e3f3ca338401281df84bff835c1a359b80c7fc1c55ee98 Feb 27 17:18:45 crc kubenswrapper[4700]: I0227 17:18:45.233297 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7ff59cf98b-bntz4" event={"ID":"fe5a4feb-297b-4c17-bc6c-cab1bf1dee41","Type":"ContainerStarted","Data":"a8801801b6cd33fec40dca5056e15823c586812f7015471034f0bfcb31f05681"} Feb 27 17:18:45 crc kubenswrapper[4700]: I0227 17:18:45.233359 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7ff59cf98b-bntz4" event={"ID":"fe5a4feb-297b-4c17-bc6c-cab1bf1dee41","Type":"ContainerStarted","Data":"436f338c30306f27e8e3f3ca338401281df84bff835c1a359b80c7fc1c55ee98"} Feb 27 17:18:45 crc kubenswrapper[4700]: I0227 17:18:45.233416 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-7ff59cf98b-bntz4" Feb 27 17:18:45 crc kubenswrapper[4700]: I0227 17:18:45.235838 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-556b8b874-4bmbt" event={"ID":"c1fdff54-9c43-4fe4-925b-f87590c75e72","Type":"ContainerStarted","Data":"addc0e2864607524e975a6f9978b3beb1b9fb3eea2493fdb9f54af42143ddec5"} Feb 27 17:18:45 crc kubenswrapper[4700]: I0227 17:18:45.283759 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-7ff59cf98b-bntz4" podStartSLOduration=34.283731127 podStartE2EDuration="34.283731127s" podCreationTimestamp="2026-02-27 17:18:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:18:45.281155429 +0000 UTC m=+1085.266468206" watchObservedRunningTime="2026-02-27 17:18:45.283731127 +0000 UTC m=+1085.269043914" Feb 27 17:18:45 crc kubenswrapper[4700]: I0227 17:18:45.315942 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-556b8b874-4bmbt" podStartSLOduration=2.70082809 podStartE2EDuration="34.315916491s" podCreationTimestamp="2026-02-27 17:18:11 +0000 UTC" firstStartedPulling="2026-02-27 17:18:12.819907941 +0000 UTC m=+1052.805220688" lastFinishedPulling="2026-02-27 17:18:44.434996312 +0000 UTC m=+1084.420309089" observedRunningTime="2026-02-27 17:18:45.303015048 +0000 UTC m=+1085.288327825" watchObservedRunningTime="2026-02-27 17:18:45.315916491 +0000 UTC m=+1085.301229278" Feb 27 17:18:45 crc kubenswrapper[4700]: I0227 17:18:45.516344 4700 scope.go:117] "RemoveContainer" containerID="f21ac7cba5e13686e60de613f90c6209fb6b1b815f8b55deec107290160ab448" Feb 27 17:18:45 crc kubenswrapper[4700]: E0227 17:18:45.832638 4700 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode548373a_d9bf_46bd_8efa_b4adf2dedc4b.slice/crio-e240c8fabd13b1b1dad1ae559c41c98d2ab7f927d48e1f1204cbcc97f5308a60\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode548373a_d9bf_46bd_8efa_b4adf2dedc4b.slice\": RecentStats: unable to find data in memory cache]" Feb 27 17:18:47 crc kubenswrapper[4700]: I0227 17:18:47.254184 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-67d996989d-f625b" event={"ID":"e6d0216e-25c4-4957-8e6b-473b49dd7543","Type":"ContainerStarted","Data":"f7dea5c0d0ef0d925907a183acf77938fc7204f968cdce15d1c942ba670ea707"} Feb 27 17:18:47 crc kubenswrapper[4700]: I0227 17:18:47.255169 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-67d996989d-f625b" Feb 27 17:18:47 crc kubenswrapper[4700]: I0227 17:18:47.285052 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-67d996989d-f625b" podStartSLOduration=3.5969253329999997 podStartE2EDuration="37.285026395s" podCreationTimestamp="2026-02-27 17:18:10 +0000 UTC" firstStartedPulling="2026-02-27 17:18:12.755602855 +0000 UTC m=+1052.740915602" lastFinishedPulling="2026-02-27 17:18:46.443703907 +0000 UTC m=+1086.429016664" observedRunningTime="2026-02-27 17:18:47.277720331 +0000 UTC m=+1087.263033118" watchObservedRunningTime="2026-02-27 17:18:47.285026395 +0000 UTC m=+1087.270339182" Feb 27 17:18:49 crc kubenswrapper[4700]: I0227 17:18:49.275863 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-54688575f-8vhl9" event={"ID":"e00942b5-6f82-48d4-8413-2f568daf5d7f","Type":"ContainerStarted","Data":"248645c71439bdeefedba807c37d79e4f14d17653d863f4eadc1add08df3142d"} Feb 27 17:18:49 crc kubenswrapper[4700]: I0227 17:18:49.277230 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-54688575f-8vhl9" Feb 27 17:18:49 crc kubenswrapper[4700]: I0227 17:18:49.277745 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5fdb694969-44djz" event={"ID":"629a3ab5-c8ce-49cb-bd57-355d643253c5","Type":"ContainerStarted","Data":"37c4148190b4f332c10491a3e50f2cfe524b3a9733093b9a88457a095ee1c0c7"} Feb 27 17:18:49 crc kubenswrapper[4700]: I0227 17:18:49.278019 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-5fdb694969-44djz" Feb 27 17:18:49 crc kubenswrapper[4700]: I0227 17:18:49.304641 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-54688575f-8vhl9" podStartSLOduration=2.5722749179999997 podStartE2EDuration="38.304622228s" podCreationTimestamp="2026-02-27 17:18:11 +0000 UTC" firstStartedPulling="2026-02-27 17:18:12.765886248 +0000 UTC m=+1052.751198995" lastFinishedPulling="2026-02-27 17:18:48.498233518 +0000 UTC m=+1088.483546305" observedRunningTime="2026-02-27 17:18:49.300756606 +0000 UTC m=+1089.286069393" watchObservedRunningTime="2026-02-27 17:18:49.304622228 +0000 UTC m=+1089.289934985" Feb 27 17:18:49 crc kubenswrapper[4700]: I0227 17:18:49.321505 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-5fdb694969-44djz" podStartSLOduration=2.9675970449999998 podStartE2EDuration="38.321457405s" podCreationTimestamp="2026-02-27 17:18:11 +0000 UTC" firstStartedPulling="2026-02-27 17:18:13.144649565 +0000 UTC m=+1053.129962312" lastFinishedPulling="2026-02-27 17:18:48.498509925 +0000 UTC m=+1088.483822672" observedRunningTime="2026-02-27 17:18:49.31711696 +0000 UTC m=+1089.302429717" watchObservedRunningTime="2026-02-27 17:18:49.321457405 +0000 UTC m=+1089.306770192" Feb 27 17:18:50 crc kubenswrapper[4700]: I0227 17:18:50.287593 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-74b6b5dc96-72l74" event={"ID":"5bbcc1bb-6070-4196-8212-cf8e04eaa923","Type":"ContainerStarted","Data":"e59b9deb835c03829f66271b7e873e879a1b672d7300443dfc342c28372589dd"} Feb 27 17:18:50 crc kubenswrapper[4700]: I0227 17:18:50.313258 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-74b6b5dc96-72l74" podStartSLOduration=2.7008495200000002 podStartE2EDuration="39.313232124s" podCreationTimestamp="2026-02-27 17:18:11 +0000 UTC" firstStartedPulling="2026-02-27 17:18:12.847214925 +0000 UTC m=+1052.832527672" lastFinishedPulling="2026-02-27 17:18:49.459597519 +0000 UTC m=+1089.444910276" observedRunningTime="2026-02-27 17:18:50.305039987 +0000 UTC m=+1090.290352774" watchObservedRunningTime="2026-02-27 17:18:50.313232124 +0000 UTC m=+1090.298544911" Feb 27 17:18:51 crc kubenswrapper[4700]: I0227 17:18:51.166346 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-6db6876945-nmv94" Feb 27 17:18:51 crc kubenswrapper[4700]: I0227 17:18:51.195371 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-dc72n" Feb 27 17:18:51 crc kubenswrapper[4700]: I0227 17:18:51.204791 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-5d87c9d997-9mxq4" Feb 27 17:18:51 crc kubenswrapper[4700]: I0227 17:18:51.222579 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-64db6967f8-l2542" Feb 27 17:18:51 crc kubenswrapper[4700]: I0227 17:18:51.282699 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-cf99c678f-n55xj" Feb 27 17:18:51 crc kubenswrapper[4700]: I0227 17:18:51.295478 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-78bc7f9bd9-dm8wj" Feb 27 17:18:51 crc kubenswrapper[4700]: I0227 17:18:51.369999 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-545456dc4-stc8v" Feb 27 17:18:51 crc kubenswrapper[4700]: I0227 17:18:51.401776 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-67d996989d-f625b" Feb 27 17:18:51 crc kubenswrapper[4700]: I0227 17:18:51.541679 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-556b8b874-4bmbt" Feb 27 17:18:51 crc kubenswrapper[4700]: I0227 17:18:51.544754 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-556b8b874-4bmbt" Feb 27 17:18:51 crc kubenswrapper[4700]: I0227 17:18:51.567219 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-74b6b5dc96-72l74" Feb 27 17:18:51 crc kubenswrapper[4700]: I0227 17:18:51.610267 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-vt248" Feb 27 17:18:51 crc kubenswrapper[4700]: I0227 17:18:51.661992 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-75684d597f-6cfj6" Feb 27 17:18:51 crc kubenswrapper[4700]: I0227 17:18:51.762653 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-648564c9fc-dmd7w" Feb 27 17:18:51 crc kubenswrapper[4700]: I0227 17:18:51.786096 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-xgf2x" Feb 27 17:18:51 crc kubenswrapper[4700]: I0227 17:18:51.820356 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-8f7484475-8pdn6" Feb 27 17:18:51 crc kubenswrapper[4700]: I0227 17:18:51.993183 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-9b9ff9f4d-cb7w5" Feb 27 17:18:52 crc kubenswrapper[4700]: I0227 17:18:52.310395 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qf2g9" event={"ID":"1010adc0-3454-486d-a676-80d329745d4c","Type":"ContainerStarted","Data":"45c5e0f588d3e5263a3ee0a91db9fdfc07449cb56e81ab6901accba18f4c9f25"} Feb 27 17:18:52 crc kubenswrapper[4700]: I0227 17:18:52.311886 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-55ffd4876b-q66w8" event={"ID":"7cd71afd-89a7-42b9-9913-6827ea7a22d3","Type":"ContainerStarted","Data":"359882db591c20c5174e7790b0f2e738708c860b044352b3066232b6c04a7ec4"} Feb 27 17:18:52 crc kubenswrapper[4700]: I0227 17:18:52.312203 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-55ffd4876b-q66w8" Feb 27 17:18:52 crc kubenswrapper[4700]: I0227 17:18:52.342529 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qf2g9" podStartSLOduration=2.755780738 podStartE2EDuration="41.342508215s" podCreationTimestamp="2026-02-27 17:18:11 +0000 UTC" firstStartedPulling="2026-02-27 17:18:13.174066226 +0000 UTC m=+1053.159378973" lastFinishedPulling="2026-02-27 17:18:51.760793703 +0000 UTC m=+1091.746106450" observedRunningTime="2026-02-27 17:18:52.327110956 +0000 UTC m=+1092.312423713" watchObservedRunningTime="2026-02-27 17:18:52.342508215 +0000 UTC m=+1092.327820972" Feb 27 17:18:52 crc kubenswrapper[4700]: I0227 17:18:52.347916 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-55ffd4876b-q66w8" podStartSLOduration=3.742668759 podStartE2EDuration="42.347902768s" podCreationTimestamp="2026-02-27 17:18:10 +0000 UTC" firstStartedPulling="2026-02-27 17:18:12.854603981 +0000 UTC m=+1052.839916728" lastFinishedPulling="2026-02-27 17:18:51.45983798 +0000 UTC m=+1091.445150737" observedRunningTime="2026-02-27 17:18:52.341547379 +0000 UTC m=+1092.326860166" watchObservedRunningTime="2026-02-27 17:18:52.347902768 +0000 UTC m=+1092.333215525" Feb 27 17:18:53 crc kubenswrapper[4700]: I0227 17:18:53.694629 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-7ff59cf98b-bntz4" Feb 27 17:18:56 crc kubenswrapper[4700]: E0227 17:18:56.032014 4700 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode548373a_d9bf_46bd_8efa_b4adf2dedc4b.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode548373a_d9bf_46bd_8efa_b4adf2dedc4b.slice/crio-e240c8fabd13b1b1dad1ae559c41c98d2ab7f927d48e1f1204cbcc97f5308a60\": RecentStats: unable to find data in memory cache]" Feb 27 17:18:56 crc kubenswrapper[4700]: I0227 17:18:56.920845 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-f7fcc58b9-84zw4" Feb 27 17:18:57 crc kubenswrapper[4700]: I0227 17:18:57.242889 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg" Feb 27 17:19:01 crc kubenswrapper[4700]: I0227 17:19:01.396430 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-55ffd4876b-q66w8" Feb 27 17:19:01 crc kubenswrapper[4700]: I0227 17:19:01.561083 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-54688575f-8vhl9" Feb 27 17:19:01 crc kubenswrapper[4700]: I0227 17:19:01.568557 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-74b6b5dc96-72l74" Feb 27 17:19:01 crc kubenswrapper[4700]: I0227 17:19:01.734490 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-5fdb694969-44djz" Feb 27 17:19:06 crc kubenswrapper[4700]: E0227 17:19:06.245237 4700 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode548373a_d9bf_46bd_8efa_b4adf2dedc4b.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode548373a_d9bf_46bd_8efa_b4adf2dedc4b.slice/crio-e240c8fabd13b1b1dad1ae559c41c98d2ab7f927d48e1f1204cbcc97f5308a60\": RecentStats: unable to find data in memory cache]" Feb 27 17:19:16 crc kubenswrapper[4700]: E0227 17:19:16.451511 4700 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode548373a_d9bf_46bd_8efa_b4adf2dedc4b.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode548373a_d9bf_46bd_8efa_b4adf2dedc4b.slice/crio-e240c8fabd13b1b1dad1ae559c41c98d2ab7f927d48e1f1204cbcc97f5308a60\": RecentStats: unable to find data in memory cache]" Feb 27 17:19:21 crc kubenswrapper[4700]: I0227 17:19:21.770453 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bd877c5c-t5pxf"] Feb 27 17:19:21 crc kubenswrapper[4700]: E0227 17:19:21.771430 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be7f5a17-2d75-446f-90ed-b5802f57b07a" containerName="extract-content" Feb 27 17:19:21 crc kubenswrapper[4700]: I0227 17:19:21.771447 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="be7f5a17-2d75-446f-90ed-b5802f57b07a" containerName="extract-content" Feb 27 17:19:21 crc kubenswrapper[4700]: E0227 17:19:21.771519 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e548373a-d9bf-46bd-8efa-b4adf2dedc4b" containerName="extract-content" Feb 27 17:19:21 crc kubenswrapper[4700]: I0227 17:19:21.771530 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="e548373a-d9bf-46bd-8efa-b4adf2dedc4b" containerName="extract-content" Feb 27 17:19:21 crc kubenswrapper[4700]: E0227 17:19:21.771544 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e548373a-d9bf-46bd-8efa-b4adf2dedc4b" containerName="registry-server" Feb 27 17:19:21 crc kubenswrapper[4700]: I0227 17:19:21.771552 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="e548373a-d9bf-46bd-8efa-b4adf2dedc4b" containerName="registry-server" Feb 27 17:19:21 crc kubenswrapper[4700]: E0227 17:19:21.771569 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be7f5a17-2d75-446f-90ed-b5802f57b07a" containerName="registry-server" Feb 27 17:19:21 crc kubenswrapper[4700]: I0227 17:19:21.771577 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="be7f5a17-2d75-446f-90ed-b5802f57b07a" containerName="registry-server" Feb 27 17:19:21 crc kubenswrapper[4700]: E0227 17:19:21.771603 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e548373a-d9bf-46bd-8efa-b4adf2dedc4b" containerName="extract-utilities" Feb 27 17:19:21 crc kubenswrapper[4700]: I0227 17:19:21.771612 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="e548373a-d9bf-46bd-8efa-b4adf2dedc4b" containerName="extract-utilities" Feb 27 17:19:21 crc kubenswrapper[4700]: E0227 17:19:21.771630 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be7f5a17-2d75-446f-90ed-b5802f57b07a" containerName="extract-utilities" Feb 27 17:19:21 crc kubenswrapper[4700]: I0227 17:19:21.771639 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="be7f5a17-2d75-446f-90ed-b5802f57b07a" containerName="extract-utilities" Feb 27 17:19:21 crc kubenswrapper[4700]: I0227 17:19:21.771797 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="be7f5a17-2d75-446f-90ed-b5802f57b07a" containerName="registry-server" Feb 27 17:19:21 crc kubenswrapper[4700]: I0227 17:19:21.771817 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="e548373a-d9bf-46bd-8efa-b4adf2dedc4b" containerName="registry-server" Feb 27 17:19:21 crc kubenswrapper[4700]: I0227 17:19:21.773908 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bd877c5c-t5pxf" Feb 27 17:19:21 crc kubenswrapper[4700]: I0227 17:19:21.776772 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Feb 27 17:19:21 crc kubenswrapper[4700]: I0227 17:19:21.788724 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-l9zmg" Feb 27 17:19:21 crc kubenswrapper[4700]: I0227 17:19:21.789651 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Feb 27 17:19:21 crc kubenswrapper[4700]: I0227 17:19:21.790725 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Feb 27 17:19:21 crc kubenswrapper[4700]: I0227 17:19:21.804006 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc996031-bc38-4f18-82b1-c7adf611b54b-config\") pod \"dnsmasq-dns-bd877c5c-t5pxf\" (UID: \"cc996031-bc38-4f18-82b1-c7adf611b54b\") " pod="openstack/dnsmasq-dns-bd877c5c-t5pxf" Feb 27 17:19:21 crc kubenswrapper[4700]: I0227 17:19:21.804062 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7t4l4\" (UniqueName: \"kubernetes.io/projected/cc996031-bc38-4f18-82b1-c7adf611b54b-kube-api-access-7t4l4\") pod \"dnsmasq-dns-bd877c5c-t5pxf\" (UID: \"cc996031-bc38-4f18-82b1-c7adf611b54b\") " pod="openstack/dnsmasq-dns-bd877c5c-t5pxf" Feb 27 17:19:21 crc kubenswrapper[4700]: I0227 17:19:21.807518 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bd877c5c-t5pxf"] Feb 27 17:19:21 crc kubenswrapper[4700]: I0227 17:19:21.837059 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f4c94c77-hw5gz"] Feb 27 17:19:21 crc kubenswrapper[4700]: I0227 17:19:21.838373 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f4c94c77-hw5gz" Feb 27 17:19:21 crc kubenswrapper[4700]: I0227 17:19:21.842013 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Feb 27 17:19:21 crc kubenswrapper[4700]: I0227 17:19:21.852243 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f4c94c77-hw5gz"] Feb 27 17:19:21 crc kubenswrapper[4700]: I0227 17:19:21.905322 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc996031-bc38-4f18-82b1-c7adf611b54b-config\") pod \"dnsmasq-dns-bd877c5c-t5pxf\" (UID: \"cc996031-bc38-4f18-82b1-c7adf611b54b\") " pod="openstack/dnsmasq-dns-bd877c5c-t5pxf" Feb 27 17:19:21 crc kubenswrapper[4700]: I0227 17:19:21.905695 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7t4l4\" (UniqueName: \"kubernetes.io/projected/cc996031-bc38-4f18-82b1-c7adf611b54b-kube-api-access-7t4l4\") pod \"dnsmasq-dns-bd877c5c-t5pxf\" (UID: \"cc996031-bc38-4f18-82b1-c7adf611b54b\") " pod="openstack/dnsmasq-dns-bd877c5c-t5pxf" Feb 27 17:19:21 crc kubenswrapper[4700]: I0227 17:19:21.906114 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc996031-bc38-4f18-82b1-c7adf611b54b-config\") pod \"dnsmasq-dns-bd877c5c-t5pxf\" (UID: \"cc996031-bc38-4f18-82b1-c7adf611b54b\") " pod="openstack/dnsmasq-dns-bd877c5c-t5pxf" Feb 27 17:19:21 crc kubenswrapper[4700]: I0227 17:19:21.934263 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7t4l4\" (UniqueName: \"kubernetes.io/projected/cc996031-bc38-4f18-82b1-c7adf611b54b-kube-api-access-7t4l4\") pod \"dnsmasq-dns-bd877c5c-t5pxf\" (UID: \"cc996031-bc38-4f18-82b1-c7adf611b54b\") " pod="openstack/dnsmasq-dns-bd877c5c-t5pxf" Feb 27 17:19:22 crc kubenswrapper[4700]: I0227 17:19:22.007249 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26b9a2b0-0140-415e-911b-14fa2eb6c3ff-dns-svc\") pod \"dnsmasq-dns-f4c94c77-hw5gz\" (UID: \"26b9a2b0-0140-415e-911b-14fa2eb6c3ff\") " pod="openstack/dnsmasq-dns-f4c94c77-hw5gz" Feb 27 17:19:22 crc kubenswrapper[4700]: I0227 17:19:22.007311 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgcmf\" (UniqueName: \"kubernetes.io/projected/26b9a2b0-0140-415e-911b-14fa2eb6c3ff-kube-api-access-cgcmf\") pod \"dnsmasq-dns-f4c94c77-hw5gz\" (UID: \"26b9a2b0-0140-415e-911b-14fa2eb6c3ff\") " pod="openstack/dnsmasq-dns-f4c94c77-hw5gz" Feb 27 17:19:22 crc kubenswrapper[4700]: I0227 17:19:22.007338 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26b9a2b0-0140-415e-911b-14fa2eb6c3ff-config\") pod \"dnsmasq-dns-f4c94c77-hw5gz\" (UID: \"26b9a2b0-0140-415e-911b-14fa2eb6c3ff\") " pod="openstack/dnsmasq-dns-f4c94c77-hw5gz" Feb 27 17:19:22 crc kubenswrapper[4700]: I0227 17:19:22.108683 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26b9a2b0-0140-415e-911b-14fa2eb6c3ff-dns-svc\") pod \"dnsmasq-dns-f4c94c77-hw5gz\" (UID: \"26b9a2b0-0140-415e-911b-14fa2eb6c3ff\") " pod="openstack/dnsmasq-dns-f4c94c77-hw5gz" Feb 27 17:19:22 crc kubenswrapper[4700]: I0227 17:19:22.109710 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgcmf\" (UniqueName: \"kubernetes.io/projected/26b9a2b0-0140-415e-911b-14fa2eb6c3ff-kube-api-access-cgcmf\") pod \"dnsmasq-dns-f4c94c77-hw5gz\" (UID: \"26b9a2b0-0140-415e-911b-14fa2eb6c3ff\") " pod="openstack/dnsmasq-dns-f4c94c77-hw5gz" Feb 27 17:19:22 crc kubenswrapper[4700]: I0227 17:19:22.109934 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26b9a2b0-0140-415e-911b-14fa2eb6c3ff-config\") pod \"dnsmasq-dns-f4c94c77-hw5gz\" (UID: \"26b9a2b0-0140-415e-911b-14fa2eb6c3ff\") " pod="openstack/dnsmasq-dns-f4c94c77-hw5gz" Feb 27 17:19:22 crc kubenswrapper[4700]: I0227 17:19:22.109950 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26b9a2b0-0140-415e-911b-14fa2eb6c3ff-dns-svc\") pod \"dnsmasq-dns-f4c94c77-hw5gz\" (UID: \"26b9a2b0-0140-415e-911b-14fa2eb6c3ff\") " pod="openstack/dnsmasq-dns-f4c94c77-hw5gz" Feb 27 17:19:22 crc kubenswrapper[4700]: I0227 17:19:22.110705 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26b9a2b0-0140-415e-911b-14fa2eb6c3ff-config\") pod \"dnsmasq-dns-f4c94c77-hw5gz\" (UID: \"26b9a2b0-0140-415e-911b-14fa2eb6c3ff\") " pod="openstack/dnsmasq-dns-f4c94c77-hw5gz" Feb 27 17:19:22 crc kubenswrapper[4700]: I0227 17:19:22.126738 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bd877c5c-t5pxf" Feb 27 17:19:22 crc kubenswrapper[4700]: I0227 17:19:22.144322 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgcmf\" (UniqueName: \"kubernetes.io/projected/26b9a2b0-0140-415e-911b-14fa2eb6c3ff-kube-api-access-cgcmf\") pod \"dnsmasq-dns-f4c94c77-hw5gz\" (UID: \"26b9a2b0-0140-415e-911b-14fa2eb6c3ff\") " pod="openstack/dnsmasq-dns-f4c94c77-hw5gz" Feb 27 17:19:22 crc kubenswrapper[4700]: I0227 17:19:22.163561 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f4c94c77-hw5gz" Feb 27 17:19:22 crc kubenswrapper[4700]: I0227 17:19:22.386862 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bd877c5c-t5pxf"] Feb 27 17:19:22 crc kubenswrapper[4700]: I0227 17:19:22.450136 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f4c94c77-hw5gz"] Feb 27 17:19:22 crc kubenswrapper[4700]: W0227 17:19:22.454612 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod26b9a2b0_0140_415e_911b_14fa2eb6c3ff.slice/crio-47c09d8aaf6ff9dfae75ce882fb913c1d192c6cec549f4112383b37064a71ae9 WatchSource:0}: Error finding container 47c09d8aaf6ff9dfae75ce882fb913c1d192c6cec549f4112383b37064a71ae9: Status 404 returned error can't find the container with id 47c09d8aaf6ff9dfae75ce882fb913c1d192c6cec549f4112383b37064a71ae9 Feb 27 17:19:22 crc kubenswrapper[4700]: I0227 17:19:22.615265 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bd877c5c-t5pxf" event={"ID":"cc996031-bc38-4f18-82b1-c7adf611b54b","Type":"ContainerStarted","Data":"c8d2af4f3291aef0078855d4beb941c6075ddea942bc81014f37930128b12f54"} Feb 27 17:19:22 crc kubenswrapper[4700]: I0227 17:19:22.616776 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f4c94c77-hw5gz" event={"ID":"26b9a2b0-0140-415e-911b-14fa2eb6c3ff","Type":"ContainerStarted","Data":"47c09d8aaf6ff9dfae75ce882fb913c1d192c6cec549f4112383b37064a71ae9"} Feb 27 17:19:25 crc kubenswrapper[4700]: I0227 17:19:25.609503 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bd877c5c-t5pxf"] Feb 27 17:19:25 crc kubenswrapper[4700]: I0227 17:19:25.640102 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7bbc67fc-tr49q"] Feb 27 17:19:25 crc kubenswrapper[4700]: I0227 17:19:25.641248 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bbc67fc-tr49q" Feb 27 17:19:25 crc kubenswrapper[4700]: I0227 17:19:25.649248 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bbc67fc-tr49q"] Feb 27 17:19:25 crc kubenswrapper[4700]: I0227 17:19:25.778098 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4448288d-21f1-418b-95bf-9287e914f52b-dns-svc\") pod \"dnsmasq-dns-7bbc67fc-tr49q\" (UID: \"4448288d-21f1-418b-95bf-9287e914f52b\") " pod="openstack/dnsmasq-dns-7bbc67fc-tr49q" Feb 27 17:19:25 crc kubenswrapper[4700]: I0227 17:19:25.778161 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4448288d-21f1-418b-95bf-9287e914f52b-config\") pod \"dnsmasq-dns-7bbc67fc-tr49q\" (UID: \"4448288d-21f1-418b-95bf-9287e914f52b\") " pod="openstack/dnsmasq-dns-7bbc67fc-tr49q" Feb 27 17:19:25 crc kubenswrapper[4700]: I0227 17:19:25.778411 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kr8ft\" (UniqueName: \"kubernetes.io/projected/4448288d-21f1-418b-95bf-9287e914f52b-kube-api-access-kr8ft\") pod \"dnsmasq-dns-7bbc67fc-tr49q\" (UID: \"4448288d-21f1-418b-95bf-9287e914f52b\") " pod="openstack/dnsmasq-dns-7bbc67fc-tr49q" Feb 27 17:19:25 crc kubenswrapper[4700]: I0227 17:19:25.877722 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f4c94c77-hw5gz"] Feb 27 17:19:25 crc kubenswrapper[4700]: I0227 17:19:25.879504 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kr8ft\" (UniqueName: \"kubernetes.io/projected/4448288d-21f1-418b-95bf-9287e914f52b-kube-api-access-kr8ft\") pod \"dnsmasq-dns-7bbc67fc-tr49q\" (UID: \"4448288d-21f1-418b-95bf-9287e914f52b\") " pod="openstack/dnsmasq-dns-7bbc67fc-tr49q" Feb 27 17:19:25 crc kubenswrapper[4700]: I0227 17:19:25.879569 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4448288d-21f1-418b-95bf-9287e914f52b-dns-svc\") pod \"dnsmasq-dns-7bbc67fc-tr49q\" (UID: \"4448288d-21f1-418b-95bf-9287e914f52b\") " pod="openstack/dnsmasq-dns-7bbc67fc-tr49q" Feb 27 17:19:25 crc kubenswrapper[4700]: I0227 17:19:25.879604 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4448288d-21f1-418b-95bf-9287e914f52b-config\") pod \"dnsmasq-dns-7bbc67fc-tr49q\" (UID: \"4448288d-21f1-418b-95bf-9287e914f52b\") " pod="openstack/dnsmasq-dns-7bbc67fc-tr49q" Feb 27 17:19:25 crc kubenswrapper[4700]: I0227 17:19:25.881229 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4448288d-21f1-418b-95bf-9287e914f52b-config\") pod \"dnsmasq-dns-7bbc67fc-tr49q\" (UID: \"4448288d-21f1-418b-95bf-9287e914f52b\") " pod="openstack/dnsmasq-dns-7bbc67fc-tr49q" Feb 27 17:19:25 crc kubenswrapper[4700]: I0227 17:19:25.882543 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4448288d-21f1-418b-95bf-9287e914f52b-dns-svc\") pod \"dnsmasq-dns-7bbc67fc-tr49q\" (UID: \"4448288d-21f1-418b-95bf-9287e914f52b\") " pod="openstack/dnsmasq-dns-7bbc67fc-tr49q" Feb 27 17:19:25 crc kubenswrapper[4700]: I0227 17:19:25.898117 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-598b499cc5-b79hp"] Feb 27 17:19:25 crc kubenswrapper[4700]: I0227 17:19:25.899338 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-598b499cc5-b79hp" Feb 27 17:19:25 crc kubenswrapper[4700]: I0227 17:19:25.913900 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kr8ft\" (UniqueName: \"kubernetes.io/projected/4448288d-21f1-418b-95bf-9287e914f52b-kube-api-access-kr8ft\") pod \"dnsmasq-dns-7bbc67fc-tr49q\" (UID: \"4448288d-21f1-418b-95bf-9287e914f52b\") " pod="openstack/dnsmasq-dns-7bbc67fc-tr49q" Feb 27 17:19:25 crc kubenswrapper[4700]: I0227 17:19:25.916672 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-598b499cc5-b79hp"] Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.000024 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bbc67fc-tr49q" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.083927 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e0e1d8f9-0461-400d-841a-a2fe02b6fba9-dns-svc\") pod \"dnsmasq-dns-598b499cc5-b79hp\" (UID: \"e0e1d8f9-0461-400d-841a-a2fe02b6fba9\") " pod="openstack/dnsmasq-dns-598b499cc5-b79hp" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.083970 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0e1d8f9-0461-400d-841a-a2fe02b6fba9-config\") pod \"dnsmasq-dns-598b499cc5-b79hp\" (UID: \"e0e1d8f9-0461-400d-841a-a2fe02b6fba9\") " pod="openstack/dnsmasq-dns-598b499cc5-b79hp" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.084017 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dqt7\" (UniqueName: \"kubernetes.io/projected/e0e1d8f9-0461-400d-841a-a2fe02b6fba9-kube-api-access-9dqt7\") pod \"dnsmasq-dns-598b499cc5-b79hp\" (UID: \"e0e1d8f9-0461-400d-841a-a2fe02b6fba9\") " pod="openstack/dnsmasq-dns-598b499cc5-b79hp" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.147288 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-598b499cc5-b79hp"] Feb 27 17:19:26 crc kubenswrapper[4700]: E0227 17:19:26.147744 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-9dqt7], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-598b499cc5-b79hp" podUID="e0e1d8f9-0461-400d-841a-a2fe02b6fba9" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.176890 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bd79c76b5-52bjb"] Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.178327 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bd79c76b5-52bjb" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.185244 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bd79c76b5-52bjb"] Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.186176 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e0e1d8f9-0461-400d-841a-a2fe02b6fba9-dns-svc\") pod \"dnsmasq-dns-598b499cc5-b79hp\" (UID: \"e0e1d8f9-0461-400d-841a-a2fe02b6fba9\") " pod="openstack/dnsmasq-dns-598b499cc5-b79hp" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.186217 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-td4pt\" (UniqueName: \"kubernetes.io/projected/cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90-kube-api-access-td4pt\") pod \"dnsmasq-dns-6bd79c76b5-52bjb\" (UID: \"cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90\") " pod="openstack/dnsmasq-dns-6bd79c76b5-52bjb" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.186237 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0e1d8f9-0461-400d-841a-a2fe02b6fba9-config\") pod \"dnsmasq-dns-598b499cc5-b79hp\" (UID: \"e0e1d8f9-0461-400d-841a-a2fe02b6fba9\") " pod="openstack/dnsmasq-dns-598b499cc5-b79hp" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.186279 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dqt7\" (UniqueName: \"kubernetes.io/projected/e0e1d8f9-0461-400d-841a-a2fe02b6fba9-kube-api-access-9dqt7\") pod \"dnsmasq-dns-598b499cc5-b79hp\" (UID: \"e0e1d8f9-0461-400d-841a-a2fe02b6fba9\") " pod="openstack/dnsmasq-dns-598b499cc5-b79hp" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.186309 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90-dns-svc\") pod \"dnsmasq-dns-6bd79c76b5-52bjb\" (UID: \"cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90\") " pod="openstack/dnsmasq-dns-6bd79c76b5-52bjb" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.186355 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90-config\") pod \"dnsmasq-dns-6bd79c76b5-52bjb\" (UID: \"cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90\") " pod="openstack/dnsmasq-dns-6bd79c76b5-52bjb" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.187075 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e0e1d8f9-0461-400d-841a-a2fe02b6fba9-dns-svc\") pod \"dnsmasq-dns-598b499cc5-b79hp\" (UID: \"e0e1d8f9-0461-400d-841a-a2fe02b6fba9\") " pod="openstack/dnsmasq-dns-598b499cc5-b79hp" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.187653 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0e1d8f9-0461-400d-841a-a2fe02b6fba9-config\") pod \"dnsmasq-dns-598b499cc5-b79hp\" (UID: \"e0e1d8f9-0461-400d-841a-a2fe02b6fba9\") " pod="openstack/dnsmasq-dns-598b499cc5-b79hp" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.207326 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dqt7\" (UniqueName: \"kubernetes.io/projected/e0e1d8f9-0461-400d-841a-a2fe02b6fba9-kube-api-access-9dqt7\") pod \"dnsmasq-dns-598b499cc5-b79hp\" (UID: \"e0e1d8f9-0461-400d-841a-a2fe02b6fba9\") " pod="openstack/dnsmasq-dns-598b499cc5-b79hp" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.290045 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90-dns-svc\") pod \"dnsmasq-dns-6bd79c76b5-52bjb\" (UID: \"cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90\") " pod="openstack/dnsmasq-dns-6bd79c76b5-52bjb" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.290111 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90-config\") pod \"dnsmasq-dns-6bd79c76b5-52bjb\" (UID: \"cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90\") " pod="openstack/dnsmasq-dns-6bd79c76b5-52bjb" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.290150 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-td4pt\" (UniqueName: \"kubernetes.io/projected/cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90-kube-api-access-td4pt\") pod \"dnsmasq-dns-6bd79c76b5-52bjb\" (UID: \"cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90\") " pod="openstack/dnsmasq-dns-6bd79c76b5-52bjb" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.291273 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90-config\") pod \"dnsmasq-dns-6bd79c76b5-52bjb\" (UID: \"cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90\") " pod="openstack/dnsmasq-dns-6bd79c76b5-52bjb" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.291286 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90-dns-svc\") pod \"dnsmasq-dns-6bd79c76b5-52bjb\" (UID: \"cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90\") " pod="openstack/dnsmasq-dns-6bd79c76b5-52bjb" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.308858 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-td4pt\" (UniqueName: \"kubernetes.io/projected/cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90-kube-api-access-td4pt\") pod \"dnsmasq-dns-6bd79c76b5-52bjb\" (UID: \"cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90\") " pod="openstack/dnsmasq-dns-6bd79c76b5-52bjb" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.500256 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bd79c76b5-52bjb" Feb 27 17:19:26 crc kubenswrapper[4700]: E0227 17:19:26.637532 4700 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode548373a_d9bf_46bd_8efa_b4adf2dedc4b.slice/crio-e240c8fabd13b1b1dad1ae559c41c98d2ab7f927d48e1f1204cbcc97f5308a60\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode548373a_d9bf_46bd_8efa_b4adf2dedc4b.slice\": RecentStats: unable to find data in memory cache]" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.643370 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-598b499cc5-b79hp" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.653095 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-598b499cc5-b79hp" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.764290 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.765638 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.767770 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.770627 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-7fq6k" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.770805 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.770913 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.771049 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.771156 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.772728 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.788441 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.794858 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dqt7\" (UniqueName: \"kubernetes.io/projected/e0e1d8f9-0461-400d-841a-a2fe02b6fba9-kube-api-access-9dqt7\") pod \"e0e1d8f9-0461-400d-841a-a2fe02b6fba9\" (UID: \"e0e1d8f9-0461-400d-841a-a2fe02b6fba9\") " Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.799118 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0e1d8f9-0461-400d-841a-a2fe02b6fba9-config\") pod \"e0e1d8f9-0461-400d-841a-a2fe02b6fba9\" (UID: \"e0e1d8f9-0461-400d-841a-a2fe02b6fba9\") " Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.799164 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e0e1d8f9-0461-400d-841a-a2fe02b6fba9-dns-svc\") pod \"e0e1d8f9-0461-400d-841a-a2fe02b6fba9\" (UID: \"e0e1d8f9-0461-400d-841a-a2fe02b6fba9\") " Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.798846 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0e1d8f9-0461-400d-841a-a2fe02b6fba9-kube-api-access-9dqt7" (OuterVolumeSpecName: "kube-api-access-9dqt7") pod "e0e1d8f9-0461-400d-841a-a2fe02b6fba9" (UID: "e0e1d8f9-0461-400d-841a-a2fe02b6fba9"). InnerVolumeSpecName "kube-api-access-9dqt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.800002 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0e1d8f9-0461-400d-841a-a2fe02b6fba9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e0e1d8f9-0461-400d-841a-a2fe02b6fba9" (UID: "e0e1d8f9-0461-400d-841a-a2fe02b6fba9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.800330 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0e1d8f9-0461-400d-841a-a2fe02b6fba9-config" (OuterVolumeSpecName: "config") pod "e0e1d8f9-0461-400d-841a-a2fe02b6fba9" (UID: "e0e1d8f9-0461-400d-841a-a2fe02b6fba9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.901561 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " pod="openstack/rabbitmq-server-0" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.901628 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/efe7d637-da3a-4995-a40f-fae00257ac1d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " pod="openstack/rabbitmq-server-0" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.901653 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/efe7d637-da3a-4995-a40f-fae00257ac1d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " pod="openstack/rabbitmq-server-0" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.901679 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/efe7d637-da3a-4995-a40f-fae00257ac1d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " pod="openstack/rabbitmq-server-0" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.901720 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/efe7d637-da3a-4995-a40f-fae00257ac1d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " pod="openstack/rabbitmq-server-0" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.901850 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/efe7d637-da3a-4995-a40f-fae00257ac1d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " pod="openstack/rabbitmq-server-0" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.901887 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/efe7d637-da3a-4995-a40f-fae00257ac1d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " pod="openstack/rabbitmq-server-0" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.901916 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/efe7d637-da3a-4995-a40f-fae00257ac1d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " pod="openstack/rabbitmq-server-0" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.902017 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/efe7d637-da3a-4995-a40f-fae00257ac1d-config-data\") pod \"rabbitmq-server-0\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " pod="openstack/rabbitmq-server-0" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.902070 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rw6kl\" (UniqueName: \"kubernetes.io/projected/efe7d637-da3a-4995-a40f-fae00257ac1d-kube-api-access-rw6kl\") pod \"rabbitmq-server-0\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " pod="openstack/rabbitmq-server-0" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.902094 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/efe7d637-da3a-4995-a40f-fae00257ac1d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " pod="openstack/rabbitmq-server-0" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.902187 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dqt7\" (UniqueName: \"kubernetes.io/projected/e0e1d8f9-0461-400d-841a-a2fe02b6fba9-kube-api-access-9dqt7\") on node \"crc\" DevicePath \"\"" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.902201 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0e1d8f9-0461-400d-841a-a2fe02b6fba9-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:19:26 crc kubenswrapper[4700]: I0227 17:19:26.902210 4700 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e0e1d8f9-0461-400d-841a-a2fe02b6fba9-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.003398 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/efe7d637-da3a-4995-a40f-fae00257ac1d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " pod="openstack/rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.003859 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/efe7d637-da3a-4995-a40f-fae00257ac1d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " pod="openstack/rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.003887 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/efe7d637-da3a-4995-a40f-fae00257ac1d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " pod="openstack/rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.003925 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/efe7d637-da3a-4995-a40f-fae00257ac1d-config-data\") pod \"rabbitmq-server-0\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " pod="openstack/rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.003957 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rw6kl\" (UniqueName: \"kubernetes.io/projected/efe7d637-da3a-4995-a40f-fae00257ac1d-kube-api-access-rw6kl\") pod \"rabbitmq-server-0\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " pod="openstack/rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.003978 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/efe7d637-da3a-4995-a40f-fae00257ac1d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " pod="openstack/rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.004007 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " pod="openstack/rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.004028 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/efe7d637-da3a-4995-a40f-fae00257ac1d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " pod="openstack/rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.004045 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/efe7d637-da3a-4995-a40f-fae00257ac1d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " pod="openstack/rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.004071 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/efe7d637-da3a-4995-a40f-fae00257ac1d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " pod="openstack/rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.004106 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/efe7d637-da3a-4995-a40f-fae00257ac1d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " pod="openstack/rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.004613 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/efe7d637-da3a-4995-a40f-fae00257ac1d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " pod="openstack/rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.005079 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/efe7d637-da3a-4995-a40f-fae00257ac1d-config-data\") pod \"rabbitmq-server-0\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " pod="openstack/rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.005143 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/efe7d637-da3a-4995-a40f-fae00257ac1d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " pod="openstack/rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.005561 4700 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.006416 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/efe7d637-da3a-4995-a40f-fae00257ac1d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " pod="openstack/rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.006430 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/efe7d637-da3a-4995-a40f-fae00257ac1d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " pod="openstack/rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.007000 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/efe7d637-da3a-4995-a40f-fae00257ac1d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " pod="openstack/rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.007481 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/efe7d637-da3a-4995-a40f-fae00257ac1d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " pod="openstack/rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.015769 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/efe7d637-da3a-4995-a40f-fae00257ac1d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " pod="openstack/rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.018140 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/efe7d637-da3a-4995-a40f-fae00257ac1d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " pod="openstack/rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.021329 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/notifications-rabbitmq-server-0"] Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.022542 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.025177 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"notifications-rabbitmq-default-user" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.025284 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-notifications-rabbitmq-svc" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.025515 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"notifications-rabbitmq-erlang-cookie" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.025623 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"notifications-rabbitmq-server-dockercfg-rh5q2" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.025630 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"notifications-rabbitmq-server-conf" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.025671 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"notifications-rabbitmq-plugins-conf" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.025995 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"notifications-rabbitmq-config-data" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.027212 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rw6kl\" (UniqueName: \"kubernetes.io/projected/efe7d637-da3a-4995-a40f-fae00257ac1d-kube-api-access-rw6kl\") pod \"rabbitmq-server-0\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " pod="openstack/rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.040872 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " pod="openstack/rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.042664 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/notifications-rabbitmq-server-0"] Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.089247 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.207389 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da-rabbitmq-erlang-cookie\") pod \"notifications-rabbitmq-server-0\" (UID: \"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da\") " pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.207472 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da-rabbitmq-plugins\") pod \"notifications-rabbitmq-server-0\" (UID: \"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da\") " pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.207504 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da-plugins-conf\") pod \"notifications-rabbitmq-server-0\" (UID: \"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da\") " pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.207538 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da-server-conf\") pod \"notifications-rabbitmq-server-0\" (UID: \"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da\") " pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.207556 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clzkh\" (UniqueName: \"kubernetes.io/projected/fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da-kube-api-access-clzkh\") pod \"notifications-rabbitmq-server-0\" (UID: \"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da\") " pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.207575 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da-config-data\") pod \"notifications-rabbitmq-server-0\" (UID: \"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da\") " pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.207594 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da-erlang-cookie-secret\") pod \"notifications-rabbitmq-server-0\" (UID: \"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da\") " pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.207620 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da-pod-info\") pod \"notifications-rabbitmq-server-0\" (UID: \"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da\") " pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.207646 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"notifications-rabbitmq-server-0\" (UID: \"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da\") " pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.207662 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da-rabbitmq-confd\") pod \"notifications-rabbitmq-server-0\" (UID: \"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da\") " pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.207683 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da-rabbitmq-tls\") pod \"notifications-rabbitmq-server-0\" (UID: \"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da\") " pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.310570 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.310628 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da-rabbitmq-plugins\") pod \"notifications-rabbitmq-server-0\" (UID: \"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da\") " pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.311147 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da-rabbitmq-plugins\") pod \"notifications-rabbitmq-server-0\" (UID: \"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da\") " pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.311208 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da-plugins-conf\") pod \"notifications-rabbitmq-server-0\" (UID: \"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da\") " pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.312001 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.313076 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da-plugins-conf\") pod \"notifications-rabbitmq-server-0\" (UID: \"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da\") " pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.313125 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da-server-conf\") pod \"notifications-rabbitmq-server-0\" (UID: \"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da\") " pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.313169 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clzkh\" (UniqueName: \"kubernetes.io/projected/fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da-kube-api-access-clzkh\") pod \"notifications-rabbitmq-server-0\" (UID: \"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da\") " pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.313196 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da-config-data\") pod \"notifications-rabbitmq-server-0\" (UID: \"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da\") " pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.313758 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da-server-conf\") pod \"notifications-rabbitmq-server-0\" (UID: \"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da\") " pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.314081 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da-config-data\") pod \"notifications-rabbitmq-server-0\" (UID: \"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da\") " pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.314153 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da-erlang-cookie-secret\") pod \"notifications-rabbitmq-server-0\" (UID: \"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da\") " pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.314299 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da-pod-info\") pod \"notifications-rabbitmq-server-0\" (UID: \"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da\") " pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.314727 4700 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"notifications-rabbitmq-server-0\" (UID: \"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.314831 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"notifications-rabbitmq-server-0\" (UID: \"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da\") " pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.314859 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da-rabbitmq-confd\") pod \"notifications-rabbitmq-server-0\" (UID: \"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da\") " pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.314888 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da-rabbitmq-tls\") pod \"notifications-rabbitmq-server-0\" (UID: \"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da\") " pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.314927 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da-rabbitmq-erlang-cookie\") pod \"notifications-rabbitmq-server-0\" (UID: \"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da\") " pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.315179 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da-rabbitmq-erlang-cookie\") pod \"notifications-rabbitmq-server-0\" (UID: \"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da\") " pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.315613 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.315938 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.316201 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-xxmpg" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.316326 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.316429 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.316552 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.316656 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.320412 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da-rabbitmq-confd\") pod \"notifications-rabbitmq-server-0\" (UID: \"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da\") " pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.320553 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da-pod-info\") pod \"notifications-rabbitmq-server-0\" (UID: \"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da\") " pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.321174 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da-rabbitmq-tls\") pod \"notifications-rabbitmq-server-0\" (UID: \"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da\") " pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.327325 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.332895 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da-erlang-cookie-secret\") pod \"notifications-rabbitmq-server-0\" (UID: \"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da\") " pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.359042 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"notifications-rabbitmq-server-0\" (UID: \"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da\") " pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.373661 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clzkh\" (UniqueName: \"kubernetes.io/projected/fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da-kube-api-access-clzkh\") pod \"notifications-rabbitmq-server-0\" (UID: \"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da\") " pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.382602 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.418391 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/458e6422-b856-4bcf-8308-4b6cb9ec8fef-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.418434 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/458e6422-b856-4bcf-8308-4b6cb9ec8fef-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.418488 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/458e6422-b856-4bcf-8308-4b6cb9ec8fef-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.418513 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/458e6422-b856-4bcf-8308-4b6cb9ec8fef-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.418529 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/458e6422-b856-4bcf-8308-4b6cb9ec8fef-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.418544 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/458e6422-b856-4bcf-8308-4b6cb9ec8fef-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.418591 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.418605 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/458e6422-b856-4bcf-8308-4b6cb9ec8fef-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.418621 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbjsk\" (UniqueName: \"kubernetes.io/projected/458e6422-b856-4bcf-8308-4b6cb9ec8fef-kube-api-access-dbjsk\") pod \"rabbitmq-cell1-server-0\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.418657 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/458e6422-b856-4bcf-8308-4b6cb9ec8fef-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.418681 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/458e6422-b856-4bcf-8308-4b6cb9ec8fef-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.520924 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/458e6422-b856-4bcf-8308-4b6cb9ec8fef-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.520977 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/458e6422-b856-4bcf-8308-4b6cb9ec8fef-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.521003 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/458e6422-b856-4bcf-8308-4b6cb9ec8fef-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.521052 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.521078 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/458e6422-b856-4bcf-8308-4b6cb9ec8fef-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.521101 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbjsk\" (UniqueName: \"kubernetes.io/projected/458e6422-b856-4bcf-8308-4b6cb9ec8fef-kube-api-access-dbjsk\") pod \"rabbitmq-cell1-server-0\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.521452 4700 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.522012 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/458e6422-b856-4bcf-8308-4b6cb9ec8fef-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.522089 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/458e6422-b856-4bcf-8308-4b6cb9ec8fef-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.522186 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/458e6422-b856-4bcf-8308-4b6cb9ec8fef-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.522222 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/458e6422-b856-4bcf-8308-4b6cb9ec8fef-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.522299 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/458e6422-b856-4bcf-8308-4b6cb9ec8fef-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.522983 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/458e6422-b856-4bcf-8308-4b6cb9ec8fef-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.525953 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/458e6422-b856-4bcf-8308-4b6cb9ec8fef-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.533126 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/458e6422-b856-4bcf-8308-4b6cb9ec8fef-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.533437 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/458e6422-b856-4bcf-8308-4b6cb9ec8fef-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.533668 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/458e6422-b856-4bcf-8308-4b6cb9ec8fef-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.534009 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/458e6422-b856-4bcf-8308-4b6cb9ec8fef-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.534172 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/458e6422-b856-4bcf-8308-4b6cb9ec8fef-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.534372 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/458e6422-b856-4bcf-8308-4b6cb9ec8fef-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.535787 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/458e6422-b856-4bcf-8308-4b6cb9ec8fef-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.538392 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbjsk\" (UniqueName: \"kubernetes.io/projected/458e6422-b856-4bcf-8308-4b6cb9ec8fef-kube-api-access-dbjsk\") pod \"rabbitmq-cell1-server-0\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.559650 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.650494 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-598b499cc5-b79hp" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.697282 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-598b499cc5-b79hp"] Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.702493 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:19:27 crc kubenswrapper[4700]: I0227 17:19:27.707272 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-598b499cc5-b79hp"] Feb 27 17:19:28 crc kubenswrapper[4700]: I0227 17:19:28.457333 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Feb 27 17:19:28 crc kubenswrapper[4700]: I0227 17:19:28.461383 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 27 17:19:28 crc kubenswrapper[4700]: I0227 17:19:28.464208 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-tmhcf" Feb 27 17:19:28 crc kubenswrapper[4700]: I0227 17:19:28.465053 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Feb 27 17:19:28 crc kubenswrapper[4700]: I0227 17:19:28.468014 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Feb 27 17:19:28 crc kubenswrapper[4700]: I0227 17:19:28.469139 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Feb 27 17:19:28 crc kubenswrapper[4700]: I0227 17:19:28.474072 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 27 17:19:28 crc kubenswrapper[4700]: I0227 17:19:28.475534 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Feb 27 17:19:28 crc kubenswrapper[4700]: I0227 17:19:28.640638 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f2325e29-9f4c-438f-9400-e643e2494808-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f2325e29-9f4c-438f-9400-e643e2494808\") " pod="openstack/openstack-galera-0" Feb 27 17:19:28 crc kubenswrapper[4700]: I0227 17:19:28.640745 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2325e29-9f4c-438f-9400-e643e2494808-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f2325e29-9f4c-438f-9400-e643e2494808\") " pod="openstack/openstack-galera-0" Feb 27 17:19:28 crc kubenswrapper[4700]: I0227 17:19:28.640798 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f2325e29-9f4c-438f-9400-e643e2494808-kolla-config\") pod \"openstack-galera-0\" (UID: \"f2325e29-9f4c-438f-9400-e643e2494808\") " pod="openstack/openstack-galera-0" Feb 27 17:19:28 crc kubenswrapper[4700]: I0227 17:19:28.640849 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2325e29-9f4c-438f-9400-e643e2494808-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"f2325e29-9f4c-438f-9400-e643e2494808\") " pod="openstack/openstack-galera-0" Feb 27 17:19:28 crc kubenswrapper[4700]: I0227 17:19:28.640889 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzwqt\" (UniqueName: \"kubernetes.io/projected/f2325e29-9f4c-438f-9400-e643e2494808-kube-api-access-qzwqt\") pod \"openstack-galera-0\" (UID: \"f2325e29-9f4c-438f-9400-e643e2494808\") " pod="openstack/openstack-galera-0" Feb 27 17:19:28 crc kubenswrapper[4700]: I0227 17:19:28.640976 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2325e29-9f4c-438f-9400-e643e2494808-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"f2325e29-9f4c-438f-9400-e643e2494808\") " pod="openstack/openstack-galera-0" Feb 27 17:19:28 crc kubenswrapper[4700]: I0227 17:19:28.641056 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f2325e29-9f4c-438f-9400-e643e2494808-config-data-default\") pod \"openstack-galera-0\" (UID: \"f2325e29-9f4c-438f-9400-e643e2494808\") " pod="openstack/openstack-galera-0" Feb 27 17:19:28 crc kubenswrapper[4700]: I0227 17:19:28.641145 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"f2325e29-9f4c-438f-9400-e643e2494808\") " pod="openstack/openstack-galera-0" Feb 27 17:19:28 crc kubenswrapper[4700]: I0227 17:19:28.742207 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"f2325e29-9f4c-438f-9400-e643e2494808\") " pod="openstack/openstack-galera-0" Feb 27 17:19:28 crc kubenswrapper[4700]: I0227 17:19:28.742303 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2325e29-9f4c-438f-9400-e643e2494808-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f2325e29-9f4c-438f-9400-e643e2494808\") " pod="openstack/openstack-galera-0" Feb 27 17:19:28 crc kubenswrapper[4700]: I0227 17:19:28.742332 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f2325e29-9f4c-438f-9400-e643e2494808-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f2325e29-9f4c-438f-9400-e643e2494808\") " pod="openstack/openstack-galera-0" Feb 27 17:19:28 crc kubenswrapper[4700]: I0227 17:19:28.742349 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f2325e29-9f4c-438f-9400-e643e2494808-kolla-config\") pod \"openstack-galera-0\" (UID: \"f2325e29-9f4c-438f-9400-e643e2494808\") " pod="openstack/openstack-galera-0" Feb 27 17:19:28 crc kubenswrapper[4700]: I0227 17:19:28.742381 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2325e29-9f4c-438f-9400-e643e2494808-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"f2325e29-9f4c-438f-9400-e643e2494808\") " pod="openstack/openstack-galera-0" Feb 27 17:19:28 crc kubenswrapper[4700]: I0227 17:19:28.742405 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzwqt\" (UniqueName: \"kubernetes.io/projected/f2325e29-9f4c-438f-9400-e643e2494808-kube-api-access-qzwqt\") pod \"openstack-galera-0\" (UID: \"f2325e29-9f4c-438f-9400-e643e2494808\") " pod="openstack/openstack-galera-0" Feb 27 17:19:28 crc kubenswrapper[4700]: I0227 17:19:28.742481 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2325e29-9f4c-438f-9400-e643e2494808-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"f2325e29-9f4c-438f-9400-e643e2494808\") " pod="openstack/openstack-galera-0" Feb 27 17:19:28 crc kubenswrapper[4700]: I0227 17:19:28.742518 4700 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"f2325e29-9f4c-438f-9400-e643e2494808\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/openstack-galera-0" Feb 27 17:19:28 crc kubenswrapper[4700]: I0227 17:19:28.742531 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f2325e29-9f4c-438f-9400-e643e2494808-config-data-default\") pod \"openstack-galera-0\" (UID: \"f2325e29-9f4c-438f-9400-e643e2494808\") " pod="openstack/openstack-galera-0" Feb 27 17:19:28 crc kubenswrapper[4700]: I0227 17:19:28.743639 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f2325e29-9f4c-438f-9400-e643e2494808-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f2325e29-9f4c-438f-9400-e643e2494808\") " pod="openstack/openstack-galera-0" Feb 27 17:19:28 crc kubenswrapper[4700]: I0227 17:19:28.743655 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f2325e29-9f4c-438f-9400-e643e2494808-kolla-config\") pod \"openstack-galera-0\" (UID: \"f2325e29-9f4c-438f-9400-e643e2494808\") " pod="openstack/openstack-galera-0" Feb 27 17:19:28 crc kubenswrapper[4700]: I0227 17:19:28.744596 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f2325e29-9f4c-438f-9400-e643e2494808-config-data-default\") pod \"openstack-galera-0\" (UID: \"f2325e29-9f4c-438f-9400-e643e2494808\") " pod="openstack/openstack-galera-0" Feb 27 17:19:28 crc kubenswrapper[4700]: I0227 17:19:28.747041 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2325e29-9f4c-438f-9400-e643e2494808-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f2325e29-9f4c-438f-9400-e643e2494808\") " pod="openstack/openstack-galera-0" Feb 27 17:19:28 crc kubenswrapper[4700]: I0227 17:19:28.749230 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2325e29-9f4c-438f-9400-e643e2494808-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"f2325e29-9f4c-438f-9400-e643e2494808\") " pod="openstack/openstack-galera-0" Feb 27 17:19:28 crc kubenswrapper[4700]: I0227 17:19:28.749904 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2325e29-9f4c-438f-9400-e643e2494808-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"f2325e29-9f4c-438f-9400-e643e2494808\") " pod="openstack/openstack-galera-0" Feb 27 17:19:28 crc kubenswrapper[4700]: I0227 17:19:28.766581 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzwqt\" (UniqueName: \"kubernetes.io/projected/f2325e29-9f4c-438f-9400-e643e2494808-kube-api-access-qzwqt\") pod \"openstack-galera-0\" (UID: \"f2325e29-9f4c-438f-9400-e643e2494808\") " pod="openstack/openstack-galera-0" Feb 27 17:19:28 crc kubenswrapper[4700]: I0227 17:19:28.781638 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"f2325e29-9f4c-438f-9400-e643e2494808\") " pod="openstack/openstack-galera-0" Feb 27 17:19:28 crc kubenswrapper[4700]: I0227 17:19:28.796592 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 27 17:19:28 crc kubenswrapper[4700]: I0227 17:19:28.997645 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0e1d8f9-0461-400d-841a-a2fe02b6fba9" path="/var/lib/kubelet/pods/e0e1d8f9-0461-400d-841a-a2fe02b6fba9/volumes" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.007761 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.010911 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.014501 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.014635 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-v8s4v" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.015095 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.016662 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.026564 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.165753 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223\") " pod="openstack/openstack-cell1-galera-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.165832 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223\") " pod="openstack/openstack-cell1-galera-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.165893 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223\") " pod="openstack/openstack-cell1-galera-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.166012 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223\") " pod="openstack/openstack-cell1-galera-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.166111 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dphd\" (UniqueName: \"kubernetes.io/projected/4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223-kube-api-access-2dphd\") pod \"openstack-cell1-galera-0\" (UID: \"4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223\") " pod="openstack/openstack-cell1-galera-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.166165 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223\") " pod="openstack/openstack-cell1-galera-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.166225 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223\") " pod="openstack/openstack-cell1-galera-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.166270 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223\") " pod="openstack/openstack-cell1-galera-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.260269 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.261794 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.263447 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-9dzhh" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.265305 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.265568 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.269241 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223\") " pod="openstack/openstack-cell1-galera-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.269368 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223\") " pod="openstack/openstack-cell1-galera-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.269406 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223\") " pod="openstack/openstack-cell1-galera-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.269510 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223\") " pod="openstack/openstack-cell1-galera-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.269543 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dphd\" (UniqueName: \"kubernetes.io/projected/4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223-kube-api-access-2dphd\") pod \"openstack-cell1-galera-0\" (UID: \"4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223\") " pod="openstack/openstack-cell1-galera-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.269574 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223\") " pod="openstack/openstack-cell1-galera-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.269618 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223\") " pod="openstack/openstack-cell1-galera-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.269667 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223\") " pod="openstack/openstack-cell1-galera-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.269710 4700 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/openstack-cell1-galera-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.270427 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223\") " pod="openstack/openstack-cell1-galera-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.270997 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223\") " pod="openstack/openstack-cell1-galera-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.271269 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223\") " pod="openstack/openstack-cell1-galera-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.272328 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223\") " pod="openstack/openstack-cell1-galera-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.275733 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223\") " pod="openstack/openstack-cell1-galera-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.276500 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223\") " pod="openstack/openstack-cell1-galera-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.295545 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.299655 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dphd\" (UniqueName: \"kubernetes.io/projected/4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223-kube-api-access-2dphd\") pod \"openstack-cell1-galera-0\" (UID: \"4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223\") " pod="openstack/openstack-cell1-galera-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.309778 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223\") " pod="openstack/openstack-cell1-galera-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.342012 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.372210 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9cf3890b-d331-4312-a6ce-a49945e61147-config-data\") pod \"memcached-0\" (UID: \"9cf3890b-d331-4312-a6ce-a49945e61147\") " pod="openstack/memcached-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.372265 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgv7g\" (UniqueName: \"kubernetes.io/projected/9cf3890b-d331-4312-a6ce-a49945e61147-kube-api-access-qgv7g\") pod \"memcached-0\" (UID: \"9cf3890b-d331-4312-a6ce-a49945e61147\") " pod="openstack/memcached-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.372302 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cf3890b-d331-4312-a6ce-a49945e61147-combined-ca-bundle\") pod \"memcached-0\" (UID: \"9cf3890b-d331-4312-a6ce-a49945e61147\") " pod="openstack/memcached-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.372344 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9cf3890b-d331-4312-a6ce-a49945e61147-kolla-config\") pod \"memcached-0\" (UID: \"9cf3890b-d331-4312-a6ce-a49945e61147\") " pod="openstack/memcached-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.372380 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/9cf3890b-d331-4312-a6ce-a49945e61147-memcached-tls-certs\") pod \"memcached-0\" (UID: \"9cf3890b-d331-4312-a6ce-a49945e61147\") " pod="openstack/memcached-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.473427 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9cf3890b-d331-4312-a6ce-a49945e61147-kolla-config\") pod \"memcached-0\" (UID: \"9cf3890b-d331-4312-a6ce-a49945e61147\") " pod="openstack/memcached-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.473497 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/9cf3890b-d331-4312-a6ce-a49945e61147-memcached-tls-certs\") pod \"memcached-0\" (UID: \"9cf3890b-d331-4312-a6ce-a49945e61147\") " pod="openstack/memcached-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.473547 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9cf3890b-d331-4312-a6ce-a49945e61147-config-data\") pod \"memcached-0\" (UID: \"9cf3890b-d331-4312-a6ce-a49945e61147\") " pod="openstack/memcached-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.473579 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgv7g\" (UniqueName: \"kubernetes.io/projected/9cf3890b-d331-4312-a6ce-a49945e61147-kube-api-access-qgv7g\") pod \"memcached-0\" (UID: \"9cf3890b-d331-4312-a6ce-a49945e61147\") " pod="openstack/memcached-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.473612 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cf3890b-d331-4312-a6ce-a49945e61147-combined-ca-bundle\") pod \"memcached-0\" (UID: \"9cf3890b-d331-4312-a6ce-a49945e61147\") " pod="openstack/memcached-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.475137 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9cf3890b-d331-4312-a6ce-a49945e61147-config-data\") pod \"memcached-0\" (UID: \"9cf3890b-d331-4312-a6ce-a49945e61147\") " pod="openstack/memcached-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.475310 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9cf3890b-d331-4312-a6ce-a49945e61147-kolla-config\") pod \"memcached-0\" (UID: \"9cf3890b-d331-4312-a6ce-a49945e61147\") " pod="openstack/memcached-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.478063 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/9cf3890b-d331-4312-a6ce-a49945e61147-memcached-tls-certs\") pod \"memcached-0\" (UID: \"9cf3890b-d331-4312-a6ce-a49945e61147\") " pod="openstack/memcached-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.479049 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cf3890b-d331-4312-a6ce-a49945e61147-combined-ca-bundle\") pod \"memcached-0\" (UID: \"9cf3890b-d331-4312-a6ce-a49945e61147\") " pod="openstack/memcached-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.497120 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgv7g\" (UniqueName: \"kubernetes.io/projected/9cf3890b-d331-4312-a6ce-a49945e61147-kube-api-access-qgv7g\") pod \"memcached-0\" (UID: \"9cf3890b-d331-4312-a6ce-a49945e61147\") " pod="openstack/memcached-0" Feb 27 17:19:30 crc kubenswrapper[4700]: I0227 17:19:30.643420 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 27 17:19:32 crc kubenswrapper[4700]: I0227 17:19:32.872315 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 27 17:19:32 crc kubenswrapper[4700]: I0227 17:19:32.874038 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 27 17:19:32 crc kubenswrapper[4700]: I0227 17:19:32.877494 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-mrnjg" Feb 27 17:19:32 crc kubenswrapper[4700]: I0227 17:19:32.899695 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 27 17:19:33 crc kubenswrapper[4700]: I0227 17:19:33.016367 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpmhn\" (UniqueName: \"kubernetes.io/projected/d204fe1c-3288-4837-992f-a04ddea6dbd3-kube-api-access-bpmhn\") pod \"kube-state-metrics-0\" (UID: \"d204fe1c-3288-4837-992f-a04ddea6dbd3\") " pod="openstack/kube-state-metrics-0" Feb 27 17:19:33 crc kubenswrapper[4700]: I0227 17:19:33.118354 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpmhn\" (UniqueName: \"kubernetes.io/projected/d204fe1c-3288-4837-992f-a04ddea6dbd3-kube-api-access-bpmhn\") pod \"kube-state-metrics-0\" (UID: \"d204fe1c-3288-4837-992f-a04ddea6dbd3\") " pod="openstack/kube-state-metrics-0" Feb 27 17:19:33 crc kubenswrapper[4700]: I0227 17:19:33.166070 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpmhn\" (UniqueName: \"kubernetes.io/projected/d204fe1c-3288-4837-992f-a04ddea6dbd3-kube-api-access-bpmhn\") pod \"kube-state-metrics-0\" (UID: \"d204fe1c-3288-4837-992f-a04ddea6dbd3\") " pod="openstack/kube-state-metrics-0" Feb 27 17:19:33 crc kubenswrapper[4700]: I0227 17:19:33.189929 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.231865 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.234258 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.238886 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.238907 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-2" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.238895 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-jdvhl" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.239004 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-1" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.238943 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.239752 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.239766 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.253895 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.265773 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.337799 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vhth\" (UniqueName: \"kubernetes.io/projected/9721431e-046b-4b3b-861e-22a65396981b-kube-api-access-4vhth\") pod \"prometheus-metric-storage-0\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.337886 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9721431e-046b-4b3b-861e-22a65396981b-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.337921 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9721431e-046b-4b3b-861e-22a65396981b-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.338189 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9721431e-046b-4b3b-861e-22a65396981b-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.338258 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/9721431e-046b-4b3b-861e-22a65396981b-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.338433 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/9721431e-046b-4b3b-861e-22a65396981b-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.338675 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/9721431e-046b-4b3b-861e-22a65396981b-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.338729 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9721431e-046b-4b3b-861e-22a65396981b-config\") pod \"prometheus-metric-storage-0\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.338814 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2\") pod \"prometheus-metric-storage-0\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.338914 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/9721431e-046b-4b3b-861e-22a65396981b-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.440577 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/9721431e-046b-4b3b-861e-22a65396981b-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.440655 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9721431e-046b-4b3b-861e-22a65396981b-config\") pod \"prometheus-metric-storage-0\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.440708 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2\") pod \"prometheus-metric-storage-0\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.440759 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/9721431e-046b-4b3b-861e-22a65396981b-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.440852 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vhth\" (UniqueName: \"kubernetes.io/projected/9721431e-046b-4b3b-861e-22a65396981b-kube-api-access-4vhth\") pod \"prometheus-metric-storage-0\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.440922 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9721431e-046b-4b3b-861e-22a65396981b-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.440955 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9721431e-046b-4b3b-861e-22a65396981b-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.441016 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9721431e-046b-4b3b-861e-22a65396981b-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.441047 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/9721431e-046b-4b3b-861e-22a65396981b-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.441107 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/9721431e-046b-4b3b-861e-22a65396981b-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.442132 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/9721431e-046b-4b3b-861e-22a65396981b-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.442171 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/9721431e-046b-4b3b-861e-22a65396981b-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.442580 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/9721431e-046b-4b3b-861e-22a65396981b-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.445745 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/9721431e-046b-4b3b-861e-22a65396981b-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.446721 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9721431e-046b-4b3b-861e-22a65396981b-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.446742 4700 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.447265 4700 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2\") pod \"prometheus-metric-storage-0\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/51157a65261d84782fc18d298283232db1d99eb5d30079bdb31ebdc3e04ec75d/globalmount\"" pod="openstack/prometheus-metric-storage-0" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.447121 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9721431e-046b-4b3b-861e-22a65396981b-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.451286 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9721431e-046b-4b3b-861e-22a65396981b-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.453543 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/9721431e-046b-4b3b-861e-22a65396981b-config\") pod \"prometheus-metric-storage-0\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.462210 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vhth\" (UniqueName: \"kubernetes.io/projected/9721431e-046b-4b3b-861e-22a65396981b-kube-api-access-4vhth\") pod \"prometheus-metric-storage-0\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.484353 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2\") pod \"prometheus-metric-storage-0\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:19:34 crc kubenswrapper[4700]: I0227 17:19:34.554062 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.608829 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-sfldb"] Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.609959 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-sfldb" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.623765 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-qqqm8"] Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.626045 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-qqqm8" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.627439 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-km9q4" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.627934 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.628084 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.642831 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-sfldb"] Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.652882 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-qqqm8"] Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.780782 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a-var-run-ovn\") pod \"ovn-controller-sfldb\" (UID: \"a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a\") " pod="openstack/ovn-controller-sfldb" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.780865 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c7fc37b2-4775-468e-b118-e399646a95e9-var-run\") pod \"ovn-controller-ovs-qqqm8\" (UID: \"c7fc37b2-4775-468e-b118-e399646a95e9\") " pod="openstack/ovn-controller-ovs-qqqm8" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.780892 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c7fc37b2-4775-468e-b118-e399646a95e9-scripts\") pod \"ovn-controller-ovs-qqqm8\" (UID: \"c7fc37b2-4775-468e-b118-e399646a95e9\") " pod="openstack/ovn-controller-ovs-qqqm8" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.781024 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/c7fc37b2-4775-468e-b118-e399646a95e9-var-log\") pod \"ovn-controller-ovs-qqqm8\" (UID: \"c7fc37b2-4775-468e-b118-e399646a95e9\") " pod="openstack/ovn-controller-ovs-qqqm8" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.781074 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a-scripts\") pod \"ovn-controller-sfldb\" (UID: \"a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a\") " pod="openstack/ovn-controller-sfldb" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.781123 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a-combined-ca-bundle\") pod \"ovn-controller-sfldb\" (UID: \"a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a\") " pod="openstack/ovn-controller-sfldb" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.781193 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a-var-log-ovn\") pod \"ovn-controller-sfldb\" (UID: \"a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a\") " pod="openstack/ovn-controller-sfldb" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.781232 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a-ovn-controller-tls-certs\") pod \"ovn-controller-sfldb\" (UID: \"a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a\") " pod="openstack/ovn-controller-sfldb" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.781277 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a-var-run\") pod \"ovn-controller-sfldb\" (UID: \"a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a\") " pod="openstack/ovn-controller-sfldb" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.781332 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/c7fc37b2-4775-468e-b118-e399646a95e9-var-lib\") pod \"ovn-controller-ovs-qqqm8\" (UID: \"c7fc37b2-4775-468e-b118-e399646a95e9\") " pod="openstack/ovn-controller-ovs-qqqm8" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.781368 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/c7fc37b2-4775-468e-b118-e399646a95e9-etc-ovs\") pod \"ovn-controller-ovs-qqqm8\" (UID: \"c7fc37b2-4775-468e-b118-e399646a95e9\") " pod="openstack/ovn-controller-ovs-qqqm8" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.781401 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tln2g\" (UniqueName: \"kubernetes.io/projected/a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a-kube-api-access-tln2g\") pod \"ovn-controller-sfldb\" (UID: \"a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a\") " pod="openstack/ovn-controller-sfldb" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.781483 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfz5w\" (UniqueName: \"kubernetes.io/projected/c7fc37b2-4775-468e-b118-e399646a95e9-kube-api-access-cfz5w\") pod \"ovn-controller-ovs-qqqm8\" (UID: \"c7fc37b2-4775-468e-b118-e399646a95e9\") " pod="openstack/ovn-controller-ovs-qqqm8" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.884884 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a-var-log-ovn\") pod \"ovn-controller-sfldb\" (UID: \"a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a\") " pod="openstack/ovn-controller-sfldb" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.886428 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a-ovn-controller-tls-certs\") pod \"ovn-controller-sfldb\" (UID: \"a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a\") " pod="openstack/ovn-controller-sfldb" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.886517 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a-var-run\") pod \"ovn-controller-sfldb\" (UID: \"a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a\") " pod="openstack/ovn-controller-sfldb" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.886598 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/c7fc37b2-4775-468e-b118-e399646a95e9-var-lib\") pod \"ovn-controller-ovs-qqqm8\" (UID: \"c7fc37b2-4775-468e-b118-e399646a95e9\") " pod="openstack/ovn-controller-ovs-qqqm8" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.886659 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/c7fc37b2-4775-468e-b118-e399646a95e9-etc-ovs\") pod \"ovn-controller-ovs-qqqm8\" (UID: \"c7fc37b2-4775-468e-b118-e399646a95e9\") " pod="openstack/ovn-controller-ovs-qqqm8" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.886692 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tln2g\" (UniqueName: \"kubernetes.io/projected/a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a-kube-api-access-tln2g\") pod \"ovn-controller-sfldb\" (UID: \"a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a\") " pod="openstack/ovn-controller-sfldb" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.886795 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a-var-run\") pod \"ovn-controller-sfldb\" (UID: \"a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a\") " pod="openstack/ovn-controller-sfldb" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.886793 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfz5w\" (UniqueName: \"kubernetes.io/projected/c7fc37b2-4775-468e-b118-e399646a95e9-kube-api-access-cfz5w\") pod \"ovn-controller-ovs-qqqm8\" (UID: \"c7fc37b2-4775-468e-b118-e399646a95e9\") " pod="openstack/ovn-controller-ovs-qqqm8" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.886921 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a-var-run-ovn\") pod \"ovn-controller-sfldb\" (UID: \"a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a\") " pod="openstack/ovn-controller-sfldb" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.886963 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c7fc37b2-4775-468e-b118-e399646a95e9-var-run\") pod \"ovn-controller-ovs-qqqm8\" (UID: \"c7fc37b2-4775-468e-b118-e399646a95e9\") " pod="openstack/ovn-controller-ovs-qqqm8" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.886981 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c7fc37b2-4775-468e-b118-e399646a95e9-scripts\") pod \"ovn-controller-ovs-qqqm8\" (UID: \"c7fc37b2-4775-468e-b118-e399646a95e9\") " pod="openstack/ovn-controller-ovs-qqqm8" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.887072 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/c7fc37b2-4775-468e-b118-e399646a95e9-var-log\") pod \"ovn-controller-ovs-qqqm8\" (UID: \"c7fc37b2-4775-468e-b118-e399646a95e9\") " pod="openstack/ovn-controller-ovs-qqqm8" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.887134 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a-scripts\") pod \"ovn-controller-sfldb\" (UID: \"a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a\") " pod="openstack/ovn-controller-sfldb" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.887196 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a-combined-ca-bundle\") pod \"ovn-controller-sfldb\" (UID: \"a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a\") " pod="openstack/ovn-controller-sfldb" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.887613 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/c7fc37b2-4775-468e-b118-e399646a95e9-var-lib\") pod \"ovn-controller-ovs-qqqm8\" (UID: \"c7fc37b2-4775-468e-b118-e399646a95e9\") " pod="openstack/ovn-controller-ovs-qqqm8" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.887852 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/c7fc37b2-4775-468e-b118-e399646a95e9-etc-ovs\") pod \"ovn-controller-ovs-qqqm8\" (UID: \"c7fc37b2-4775-468e-b118-e399646a95e9\") " pod="openstack/ovn-controller-ovs-qqqm8" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.887879 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c7fc37b2-4775-468e-b118-e399646a95e9-var-run\") pod \"ovn-controller-ovs-qqqm8\" (UID: \"c7fc37b2-4775-468e-b118-e399646a95e9\") " pod="openstack/ovn-controller-ovs-qqqm8" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.886683 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a-var-log-ovn\") pod \"ovn-controller-sfldb\" (UID: \"a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a\") " pod="openstack/ovn-controller-sfldb" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.888105 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a-var-run-ovn\") pod \"ovn-controller-sfldb\" (UID: \"a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a\") " pod="openstack/ovn-controller-sfldb" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.889664 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/c7fc37b2-4775-468e-b118-e399646a95e9-var-log\") pod \"ovn-controller-ovs-qqqm8\" (UID: \"c7fc37b2-4775-468e-b118-e399646a95e9\") " pod="openstack/ovn-controller-ovs-qqqm8" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.889948 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a-scripts\") pod \"ovn-controller-sfldb\" (UID: \"a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a\") " pod="openstack/ovn-controller-sfldb" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.890000 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c7fc37b2-4775-468e-b118-e399646a95e9-scripts\") pod \"ovn-controller-ovs-qqqm8\" (UID: \"c7fc37b2-4775-468e-b118-e399646a95e9\") " pod="openstack/ovn-controller-ovs-qqqm8" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.900998 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a-combined-ca-bundle\") pod \"ovn-controller-sfldb\" (UID: \"a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a\") " pod="openstack/ovn-controller-sfldb" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.904874 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfz5w\" (UniqueName: \"kubernetes.io/projected/c7fc37b2-4775-468e-b118-e399646a95e9-kube-api-access-cfz5w\") pod \"ovn-controller-ovs-qqqm8\" (UID: \"c7fc37b2-4775-468e-b118-e399646a95e9\") " pod="openstack/ovn-controller-ovs-qqqm8" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.907394 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tln2g\" (UniqueName: \"kubernetes.io/projected/a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a-kube-api-access-tln2g\") pod \"ovn-controller-sfldb\" (UID: \"a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a\") " pod="openstack/ovn-controller-sfldb" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.919352 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a-ovn-controller-tls-certs\") pod \"ovn-controller-sfldb\" (UID: \"a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a\") " pod="openstack/ovn-controller-sfldb" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.928763 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-sfldb" Feb 27 17:19:35 crc kubenswrapper[4700]: I0227 17:19:35.940413 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-qqqm8" Feb 27 17:19:36 crc kubenswrapper[4700]: I0227 17:19:36.244493 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 27 17:19:36 crc kubenswrapper[4700]: I0227 17:19:36.252622 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 27 17:19:36 crc kubenswrapper[4700]: I0227 17:19:36.257803 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Feb 27 17:19:36 crc kubenswrapper[4700]: I0227 17:19:36.261400 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-tmvzr" Feb 27 17:19:36 crc kubenswrapper[4700]: I0227 17:19:36.262982 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Feb 27 17:19:36 crc kubenswrapper[4700]: I0227 17:19:36.263219 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Feb 27 17:19:36 crc kubenswrapper[4700]: I0227 17:19:36.263366 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Feb 27 17:19:36 crc kubenswrapper[4700]: I0227 17:19:36.264305 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 27 17:19:36 crc kubenswrapper[4700]: I0227 17:19:36.284106 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bbc67fc-tr49q"] Feb 27 17:19:36 crc kubenswrapper[4700]: I0227 17:19:36.399499 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4b9e543-db0e-4b6a-bd32-9935f603556e-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a4b9e543-db0e-4b6a-bd32-9935f603556e\") " pod="openstack/ovsdbserver-nb-0" Feb 27 17:19:36 crc kubenswrapper[4700]: I0227 17:19:36.399559 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4b9e543-db0e-4b6a-bd32-9935f603556e-config\") pod \"ovsdbserver-nb-0\" (UID: \"a4b9e543-db0e-4b6a-bd32-9935f603556e\") " pod="openstack/ovsdbserver-nb-0" Feb 27 17:19:36 crc kubenswrapper[4700]: I0227 17:19:36.399605 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a4b9e543-db0e-4b6a-bd32-9935f603556e-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a4b9e543-db0e-4b6a-bd32-9935f603556e\") " pod="openstack/ovsdbserver-nb-0" Feb 27 17:19:36 crc kubenswrapper[4700]: I0227 17:19:36.399628 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4b9e543-db0e-4b6a-bd32-9935f603556e-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a4b9e543-db0e-4b6a-bd32-9935f603556e\") " pod="openstack/ovsdbserver-nb-0" Feb 27 17:19:36 crc kubenswrapper[4700]: I0227 17:19:36.399651 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4b9e543-db0e-4b6a-bd32-9935f603556e-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a4b9e543-db0e-4b6a-bd32-9935f603556e\") " pod="openstack/ovsdbserver-nb-0" Feb 27 17:19:36 crc kubenswrapper[4700]: I0227 17:19:36.399672 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwzx9\" (UniqueName: \"kubernetes.io/projected/a4b9e543-db0e-4b6a-bd32-9935f603556e-kube-api-access-qwzx9\") pod \"ovsdbserver-nb-0\" (UID: \"a4b9e543-db0e-4b6a-bd32-9935f603556e\") " pod="openstack/ovsdbserver-nb-0" Feb 27 17:19:36 crc kubenswrapper[4700]: I0227 17:19:36.399696 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a4b9e543-db0e-4b6a-bd32-9935f603556e-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a4b9e543-db0e-4b6a-bd32-9935f603556e\") " pod="openstack/ovsdbserver-nb-0" Feb 27 17:19:36 crc kubenswrapper[4700]: I0227 17:19:36.399712 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a4b9e543-db0e-4b6a-bd32-9935f603556e\") " pod="openstack/ovsdbserver-nb-0" Feb 27 17:19:36 crc kubenswrapper[4700]: I0227 17:19:36.501439 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a4b9e543-db0e-4b6a-bd32-9935f603556e-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a4b9e543-db0e-4b6a-bd32-9935f603556e\") " pod="openstack/ovsdbserver-nb-0" Feb 27 17:19:36 crc kubenswrapper[4700]: I0227 17:19:36.501501 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4b9e543-db0e-4b6a-bd32-9935f603556e-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a4b9e543-db0e-4b6a-bd32-9935f603556e\") " pod="openstack/ovsdbserver-nb-0" Feb 27 17:19:36 crc kubenswrapper[4700]: I0227 17:19:36.501529 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4b9e543-db0e-4b6a-bd32-9935f603556e-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a4b9e543-db0e-4b6a-bd32-9935f603556e\") " pod="openstack/ovsdbserver-nb-0" Feb 27 17:19:36 crc kubenswrapper[4700]: I0227 17:19:36.501555 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwzx9\" (UniqueName: \"kubernetes.io/projected/a4b9e543-db0e-4b6a-bd32-9935f603556e-kube-api-access-qwzx9\") pod \"ovsdbserver-nb-0\" (UID: \"a4b9e543-db0e-4b6a-bd32-9935f603556e\") " pod="openstack/ovsdbserver-nb-0" Feb 27 17:19:36 crc kubenswrapper[4700]: I0227 17:19:36.501588 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a4b9e543-db0e-4b6a-bd32-9935f603556e-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a4b9e543-db0e-4b6a-bd32-9935f603556e\") " pod="openstack/ovsdbserver-nb-0" Feb 27 17:19:36 crc kubenswrapper[4700]: I0227 17:19:36.501605 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a4b9e543-db0e-4b6a-bd32-9935f603556e\") " pod="openstack/ovsdbserver-nb-0" Feb 27 17:19:36 crc kubenswrapper[4700]: I0227 17:19:36.501649 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4b9e543-db0e-4b6a-bd32-9935f603556e-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a4b9e543-db0e-4b6a-bd32-9935f603556e\") " pod="openstack/ovsdbserver-nb-0" Feb 27 17:19:36 crc kubenswrapper[4700]: I0227 17:19:36.501679 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4b9e543-db0e-4b6a-bd32-9935f603556e-config\") pod \"ovsdbserver-nb-0\" (UID: \"a4b9e543-db0e-4b6a-bd32-9935f603556e\") " pod="openstack/ovsdbserver-nb-0" Feb 27 17:19:36 crc kubenswrapper[4700]: I0227 17:19:36.502002 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a4b9e543-db0e-4b6a-bd32-9935f603556e-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a4b9e543-db0e-4b6a-bd32-9935f603556e\") " pod="openstack/ovsdbserver-nb-0" Feb 27 17:19:36 crc kubenswrapper[4700]: I0227 17:19:36.502505 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4b9e543-db0e-4b6a-bd32-9935f603556e-config\") pod \"ovsdbserver-nb-0\" (UID: \"a4b9e543-db0e-4b6a-bd32-9935f603556e\") " pod="openstack/ovsdbserver-nb-0" Feb 27 17:19:36 crc kubenswrapper[4700]: I0227 17:19:36.502708 4700 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a4b9e543-db0e-4b6a-bd32-9935f603556e\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/ovsdbserver-nb-0" Feb 27 17:19:36 crc kubenswrapper[4700]: I0227 17:19:36.503442 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a4b9e543-db0e-4b6a-bd32-9935f603556e-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a4b9e543-db0e-4b6a-bd32-9935f603556e\") " pod="openstack/ovsdbserver-nb-0" Feb 27 17:19:36 crc kubenswrapper[4700]: I0227 17:19:36.511573 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4b9e543-db0e-4b6a-bd32-9935f603556e-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a4b9e543-db0e-4b6a-bd32-9935f603556e\") " pod="openstack/ovsdbserver-nb-0" Feb 27 17:19:36 crc kubenswrapper[4700]: I0227 17:19:36.512084 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4b9e543-db0e-4b6a-bd32-9935f603556e-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a4b9e543-db0e-4b6a-bd32-9935f603556e\") " pod="openstack/ovsdbserver-nb-0" Feb 27 17:19:36 crc kubenswrapper[4700]: I0227 17:19:36.514931 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4b9e543-db0e-4b6a-bd32-9935f603556e-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a4b9e543-db0e-4b6a-bd32-9935f603556e\") " pod="openstack/ovsdbserver-nb-0" Feb 27 17:19:36 crc kubenswrapper[4700]: I0227 17:19:36.518951 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwzx9\" (UniqueName: \"kubernetes.io/projected/a4b9e543-db0e-4b6a-bd32-9935f603556e-kube-api-access-qwzx9\") pod \"ovsdbserver-nb-0\" (UID: \"a4b9e543-db0e-4b6a-bd32-9935f603556e\") " pod="openstack/ovsdbserver-nb-0" Feb 27 17:19:36 crc kubenswrapper[4700]: I0227 17:19:36.525814 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a4b9e543-db0e-4b6a-bd32-9935f603556e\") " pod="openstack/ovsdbserver-nb-0" Feb 27 17:19:36 crc kubenswrapper[4700]: I0227 17:19:36.590828 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 27 17:19:36 crc kubenswrapper[4700]: E0227 17:19:36.829873 4700 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode548373a_d9bf_46bd_8efa_b4adf2dedc4b.slice/crio-e240c8fabd13b1b1dad1ae559c41c98d2ab7f927d48e1f1204cbcc97f5308a60\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode548373a_d9bf_46bd_8efa_b4adf2dedc4b.slice\": RecentStats: unable to find data in memory cache]" Feb 27 17:19:40 crc kubenswrapper[4700]: I0227 17:19:40.003558 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 27 17:19:40 crc kubenswrapper[4700]: I0227 17:19:40.005936 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 27 17:19:40 crc kubenswrapper[4700]: I0227 17:19:40.011775 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Feb 27 17:19:40 crc kubenswrapper[4700]: I0227 17:19:40.011965 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Feb 27 17:19:40 crc kubenswrapper[4700]: I0227 17:19:40.012288 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Feb 27 17:19:40 crc kubenswrapper[4700]: I0227 17:19:40.012543 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-z2bqn" Feb 27 17:19:40 crc kubenswrapper[4700]: I0227 17:19:40.022844 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 27 17:19:40 crc kubenswrapper[4700]: I0227 17:19:40.171220 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e\") " pod="openstack/ovsdbserver-sb-0" Feb 27 17:19:40 crc kubenswrapper[4700]: I0227 17:19:40.171277 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hc8wl\" (UniqueName: \"kubernetes.io/projected/f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e-kube-api-access-hc8wl\") pod \"ovsdbserver-sb-0\" (UID: \"f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e\") " pod="openstack/ovsdbserver-sb-0" Feb 27 17:19:40 crc kubenswrapper[4700]: I0227 17:19:40.171336 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e\") " pod="openstack/ovsdbserver-sb-0" Feb 27 17:19:40 crc kubenswrapper[4700]: I0227 17:19:40.171416 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e-config\") pod \"ovsdbserver-sb-0\" (UID: \"f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e\") " pod="openstack/ovsdbserver-sb-0" Feb 27 17:19:40 crc kubenswrapper[4700]: I0227 17:19:40.171537 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e\") " pod="openstack/ovsdbserver-sb-0" Feb 27 17:19:40 crc kubenswrapper[4700]: I0227 17:19:40.171568 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e\") " pod="openstack/ovsdbserver-sb-0" Feb 27 17:19:40 crc kubenswrapper[4700]: I0227 17:19:40.171592 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e\") " pod="openstack/ovsdbserver-sb-0" Feb 27 17:19:40 crc kubenswrapper[4700]: I0227 17:19:40.171742 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e\") " pod="openstack/ovsdbserver-sb-0" Feb 27 17:19:40 crc kubenswrapper[4700]: I0227 17:19:40.273794 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e\") " pod="openstack/ovsdbserver-sb-0" Feb 27 17:19:40 crc kubenswrapper[4700]: I0227 17:19:40.273849 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e-config\") pod \"ovsdbserver-sb-0\" (UID: \"f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e\") " pod="openstack/ovsdbserver-sb-0" Feb 27 17:19:40 crc kubenswrapper[4700]: I0227 17:19:40.273881 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e\") " pod="openstack/ovsdbserver-sb-0" Feb 27 17:19:40 crc kubenswrapper[4700]: I0227 17:19:40.273901 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e\") " pod="openstack/ovsdbserver-sb-0" Feb 27 17:19:40 crc kubenswrapper[4700]: I0227 17:19:40.273924 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e\") " pod="openstack/ovsdbserver-sb-0" Feb 27 17:19:40 crc kubenswrapper[4700]: I0227 17:19:40.273964 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e\") " pod="openstack/ovsdbserver-sb-0" Feb 27 17:19:40 crc kubenswrapper[4700]: I0227 17:19:40.274029 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e\") " pod="openstack/ovsdbserver-sb-0" Feb 27 17:19:40 crc kubenswrapper[4700]: I0227 17:19:40.274046 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hc8wl\" (UniqueName: \"kubernetes.io/projected/f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e-kube-api-access-hc8wl\") pod \"ovsdbserver-sb-0\" (UID: \"f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e\") " pod="openstack/ovsdbserver-sb-0" Feb 27 17:19:40 crc kubenswrapper[4700]: I0227 17:19:40.274160 4700 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/ovsdbserver-sb-0" Feb 27 17:19:40 crc kubenswrapper[4700]: I0227 17:19:40.274662 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e\") " pod="openstack/ovsdbserver-sb-0" Feb 27 17:19:40 crc kubenswrapper[4700]: I0227 17:19:40.275101 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e-config\") pod \"ovsdbserver-sb-0\" (UID: \"f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e\") " pod="openstack/ovsdbserver-sb-0" Feb 27 17:19:40 crc kubenswrapper[4700]: I0227 17:19:40.275171 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e\") " pod="openstack/ovsdbserver-sb-0" Feb 27 17:19:40 crc kubenswrapper[4700]: I0227 17:19:40.283041 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e\") " pod="openstack/ovsdbserver-sb-0" Feb 27 17:19:40 crc kubenswrapper[4700]: I0227 17:19:40.283055 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e\") " pod="openstack/ovsdbserver-sb-0" Feb 27 17:19:40 crc kubenswrapper[4700]: I0227 17:19:40.284451 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e\") " pod="openstack/ovsdbserver-sb-0" Feb 27 17:19:40 crc kubenswrapper[4700]: I0227 17:19:40.306387 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hc8wl\" (UniqueName: \"kubernetes.io/projected/f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e-kube-api-access-hc8wl\") pod \"ovsdbserver-sb-0\" (UID: \"f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e\") " pod="openstack/ovsdbserver-sb-0" Feb 27 17:19:40 crc kubenswrapper[4700]: I0227 17:19:40.309794 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e\") " pod="openstack/ovsdbserver-sb-0" Feb 27 17:19:40 crc kubenswrapper[4700]: I0227 17:19:40.336711 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 27 17:19:42 crc kubenswrapper[4700]: I0227 17:19:42.772311 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bbc67fc-tr49q" event={"ID":"4448288d-21f1-418b-95bf-9287e914f52b","Type":"ContainerStarted","Data":"3f0576771bc811588b00cc850d5921f2073af9431d691c3c088b57cb79943693"} Feb 27 17:19:43 crc kubenswrapper[4700]: E0227 17:19:43.043625 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.145:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Feb 27 17:19:43 crc kubenswrapper[4700]: E0227 17:19:43.044034 4700 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.145:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Feb 27 17:19:43 crc kubenswrapper[4700]: E0227 17:19:43.044429 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:38.102.83.145:5001/podified-master-centos10/openstack-neutron-server:watcher_latest,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7t4l4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-bd877c5c-t5pxf_openstack(cc996031-bc38-4f18-82b1-c7adf611b54b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 17:19:43 crc kubenswrapper[4700]: E0227 17:19:43.046815 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-bd877c5c-t5pxf" podUID="cc996031-bc38-4f18-82b1-c7adf611b54b" Feb 27 17:19:43 crc kubenswrapper[4700]: E0227 17:19:43.046820 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.145:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Feb 27 17:19:43 crc kubenswrapper[4700]: E0227 17:19:43.046887 4700 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.145:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Feb 27 17:19:43 crc kubenswrapper[4700]: E0227 17:19:43.047015 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:38.102.83.145:5001/podified-master-centos10/openstack-neutron-server:watcher_latest,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cgcmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-f4c94c77-hw5gz_openstack(26b9a2b0-0140-415e-911b-14fa2eb6c3ff): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 17:19:43 crc kubenswrapper[4700]: E0227 17:19:43.048176 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-f4c94c77-hw5gz" podUID="26b9a2b0-0140-415e-911b-14fa2eb6c3ff" Feb 27 17:19:43 crc kubenswrapper[4700]: I0227 17:19:43.651066 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/notifications-rabbitmq-server-0"] Feb 27 17:19:43 crc kubenswrapper[4700]: I0227 17:19:43.702511 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 27 17:19:43 crc kubenswrapper[4700]: I0227 17:19:43.709276 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 27 17:19:43 crc kubenswrapper[4700]: I0227 17:19:43.715079 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bd79c76b5-52bjb"] Feb 27 17:19:43 crc kubenswrapper[4700]: I0227 17:19:43.844572 4700 generic.go:334] "Generic (PLEG): container finished" podID="4448288d-21f1-418b-95bf-9287e914f52b" containerID="c49d783ad0bd870fae780433ddf4578161bc63d1f014a176c5133aaf46fb2365" exitCode=0 Feb 27 17:19:43 crc kubenswrapper[4700]: I0227 17:19:43.844843 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bbc67fc-tr49q" event={"ID":"4448288d-21f1-418b-95bf-9287e914f52b","Type":"ContainerDied","Data":"c49d783ad0bd870fae780433ddf4578161bc63d1f014a176c5133aaf46fb2365"} Feb 27 17:19:43 crc kubenswrapper[4700]: I0227 17:19:43.875558 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/notifications-rabbitmq-server-0" event={"ID":"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da","Type":"ContainerStarted","Data":"82add623798031a973ec4dbe5337072bc97bbd8325ceb63d1f5acccf0540d108"} Feb 27 17:19:43 crc kubenswrapper[4700]: I0227 17:19:43.892306 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 27 17:19:43 crc kubenswrapper[4700]: I0227 17:19:43.918810 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bd79c76b5-52bjb" event={"ID":"cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90","Type":"ContainerStarted","Data":"8f95e01899e24cde97f8711844fa6df575b6bd008ab9127233cfa4f928203396"} Feb 27 17:19:43 crc kubenswrapper[4700]: I0227 17:19:43.924298 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"458e6422-b856-4bcf-8308-4b6cb9ec8fef","Type":"ContainerStarted","Data":"5b6a1b53951776cb57deade96773bdd54ce581139e41fe6311edce8336aacd0d"} Feb 27 17:19:43 crc kubenswrapper[4700]: I0227 17:19:43.929311 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"efe7d637-da3a-4995-a40f-fae00257ac1d","Type":"ContainerStarted","Data":"d359c4ba60ca395c18f917a41cc24f1757315df110bc33704754d767fc0749b0"} Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.132570 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.159025 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.200477 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.249814 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Feb 27 17:19:44 crc kubenswrapper[4700]: E0227 17:19:44.251471 4700 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Feb 27 17:19:44 crc kubenswrapper[4700]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/4448288d-21f1-418b-95bf-9287e914f52b/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Feb 27 17:19:44 crc kubenswrapper[4700]: > podSandboxID="3f0576771bc811588b00cc850d5921f2073af9431d691c3c088b57cb79943693" Feb 27 17:19:44 crc kubenswrapper[4700]: E0227 17:19:44.251654 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 17:19:44 crc kubenswrapper[4700]: container &Container{Name:dnsmasq-dns,Image:38.102.83.145:5001/podified-master-centos10/openstack-neutron-server:watcher_latest,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kr8ft,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7bbc67fc-tr49q_openstack(4448288d-21f1-418b-95bf-9287e914f52b): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/4448288d-21f1-418b-95bf-9287e914f52b/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Feb 27 17:19:44 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 17:19:44 crc kubenswrapper[4700]: E0227 17:19:44.252724 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/4448288d-21f1-418b-95bf-9287e914f52b/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-7bbc67fc-tr49q" podUID="4448288d-21f1-418b-95bf-9287e914f52b" Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.279943 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-sfldb"] Feb 27 17:19:44 crc kubenswrapper[4700]: W0227 17:19:44.284037 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3c2abcc_cc58_4b41_ae6a_c1f67a10c32a.slice/crio-7c7056e3e4fad3cbd6cf5a3b9bdc3997949a250f828179bf86eab4422e0cbdd9 WatchSource:0}: Error finding container 7c7056e3e4fad3cbd6cf5a3b9bdc3997949a250f828179bf86eab4422e0cbdd9: Status 404 returned error can't find the container with id 7c7056e3e4fad3cbd6cf5a3b9bdc3997949a250f828179bf86eab4422e0cbdd9 Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.287814 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 27 17:19:44 crc kubenswrapper[4700]: W0227 17:19:44.289621 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9721431e_046b_4b3b_861e_22a65396981b.slice/crio-a1aeda0f76c4a3989dbd6ea69ae40b074a48598580bdf4195f0495834505d8a2 WatchSource:0}: Error finding container a1aeda0f76c4a3989dbd6ea69ae40b074a48598580bdf4195f0495834505d8a2: Status 404 returned error can't find the container with id a1aeda0f76c4a3989dbd6ea69ae40b074a48598580bdf4195f0495834505d8a2 Feb 27 17:19:44 crc kubenswrapper[4700]: W0227 17:19:44.290016 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd204fe1c_3288_4837_992f_a04ddea6dbd3.slice/crio-dd085c8bf25a8c9a5becee29a8178c0afe5712b089d9d7fffd72f63deae2b02f WatchSource:0}: Error finding container dd085c8bf25a8c9a5becee29a8178c0afe5712b089d9d7fffd72f63deae2b02f: Status 404 returned error can't find the container with id dd085c8bf25a8c9a5becee29a8178c0afe5712b089d9d7fffd72f63deae2b02f Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.361331 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-qqqm8"] Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.368008 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bd877c5c-t5pxf" Feb 27 17:19:44 crc kubenswrapper[4700]: W0227 17:19:44.397570 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc7fc37b2_4775_468e_b118_e399646a95e9.slice/crio-ba9a2a9fff7eeaeda512c1331a532b0adf8eed31022e44ca7716a5d51bce2634 WatchSource:0}: Error finding container ba9a2a9fff7eeaeda512c1331a532b0adf8eed31022e44ca7716a5d51bce2634: Status 404 returned error can't find the container with id ba9a2a9fff7eeaeda512c1331a532b0adf8eed31022e44ca7716a5d51bce2634 Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.417727 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f4c94c77-hw5gz" Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.561274 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7t4l4\" (UniqueName: \"kubernetes.io/projected/cc996031-bc38-4f18-82b1-c7adf611b54b-kube-api-access-7t4l4\") pod \"cc996031-bc38-4f18-82b1-c7adf611b54b\" (UID: \"cc996031-bc38-4f18-82b1-c7adf611b54b\") " Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.561486 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc996031-bc38-4f18-82b1-c7adf611b54b-config\") pod \"cc996031-bc38-4f18-82b1-c7adf611b54b\" (UID: \"cc996031-bc38-4f18-82b1-c7adf611b54b\") " Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.561511 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26b9a2b0-0140-415e-911b-14fa2eb6c3ff-config\") pod \"26b9a2b0-0140-415e-911b-14fa2eb6c3ff\" (UID: \"26b9a2b0-0140-415e-911b-14fa2eb6c3ff\") " Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.561534 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgcmf\" (UniqueName: \"kubernetes.io/projected/26b9a2b0-0140-415e-911b-14fa2eb6c3ff-kube-api-access-cgcmf\") pod \"26b9a2b0-0140-415e-911b-14fa2eb6c3ff\" (UID: \"26b9a2b0-0140-415e-911b-14fa2eb6c3ff\") " Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.561577 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26b9a2b0-0140-415e-911b-14fa2eb6c3ff-dns-svc\") pod \"26b9a2b0-0140-415e-911b-14fa2eb6c3ff\" (UID: \"26b9a2b0-0140-415e-911b-14fa2eb6c3ff\") " Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.562041 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc996031-bc38-4f18-82b1-c7adf611b54b-config" (OuterVolumeSpecName: "config") pod "cc996031-bc38-4f18-82b1-c7adf611b54b" (UID: "cc996031-bc38-4f18-82b1-c7adf611b54b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.562306 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26b9a2b0-0140-415e-911b-14fa2eb6c3ff-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "26b9a2b0-0140-415e-911b-14fa2eb6c3ff" (UID: "26b9a2b0-0140-415e-911b-14fa2eb6c3ff"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.562298 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26b9a2b0-0140-415e-911b-14fa2eb6c3ff-config" (OuterVolumeSpecName: "config") pod "26b9a2b0-0140-415e-911b-14fa2eb6c3ff" (UID: "26b9a2b0-0140-415e-911b-14fa2eb6c3ff"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.567182 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26b9a2b0-0140-415e-911b-14fa2eb6c3ff-kube-api-access-cgcmf" (OuterVolumeSpecName: "kube-api-access-cgcmf") pod "26b9a2b0-0140-415e-911b-14fa2eb6c3ff" (UID: "26b9a2b0-0140-415e-911b-14fa2eb6c3ff"). InnerVolumeSpecName "kube-api-access-cgcmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.567729 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc996031-bc38-4f18-82b1-c7adf611b54b-kube-api-access-7t4l4" (OuterVolumeSpecName: "kube-api-access-7t4l4") pod "cc996031-bc38-4f18-82b1-c7adf611b54b" (UID: "cc996031-bc38-4f18-82b1-c7adf611b54b"). InnerVolumeSpecName "kube-api-access-7t4l4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.663374 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc996031-bc38-4f18-82b1-c7adf611b54b-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.663406 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26b9a2b0-0140-415e-911b-14fa2eb6c3ff-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.663418 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgcmf\" (UniqueName: \"kubernetes.io/projected/26b9a2b0-0140-415e-911b-14fa2eb6c3ff-kube-api-access-cgcmf\") on node \"crc\" DevicePath \"\"" Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.663429 4700 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26b9a2b0-0140-415e-911b-14fa2eb6c3ff-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.663438 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7t4l4\" (UniqueName: \"kubernetes.io/projected/cc996031-bc38-4f18-82b1-c7adf611b54b-kube-api-access-7t4l4\") on node \"crc\" DevicePath \"\"" Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.765437 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 27 17:19:44 crc kubenswrapper[4700]: W0227 17:19:44.770299 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf9e933fb_76ab_4cf0_8168_86f0e3f0ea7e.slice/crio-701b1ed645b0e6facc3a739b03d223fba571e980f3adb96a66f280513fe268a5 WatchSource:0}: Error finding container 701b1ed645b0e6facc3a739b03d223fba571e980f3adb96a66f280513fe268a5: Status 404 returned error can't find the container with id 701b1ed645b0e6facc3a739b03d223fba571e980f3adb96a66f280513fe268a5 Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.943086 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e","Type":"ContainerStarted","Data":"701b1ed645b0e6facc3a739b03d223fba571e980f3adb96a66f280513fe268a5"} Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.944533 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-sfldb" event={"ID":"a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a","Type":"ContainerStarted","Data":"7c7056e3e4fad3cbd6cf5a3b9bdc3997949a250f828179bf86eab4422e0cbdd9"} Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.945601 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f4c94c77-hw5gz" event={"ID":"26b9a2b0-0140-415e-911b-14fa2eb6c3ff","Type":"ContainerDied","Data":"47c09d8aaf6ff9dfae75ce882fb913c1d192c6cec549f4112383b37064a71ae9"} Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.945661 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f4c94c77-hw5gz" Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.949149 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"9cf3890b-d331-4312-a6ce-a49945e61147","Type":"ContainerStarted","Data":"eae2e486ad2470fb48bb26030cdf06ac0c0cc494715c5e8b5fa85ed135d84e51"} Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.950667 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9721431e-046b-4b3b-861e-22a65396981b","Type":"ContainerStarted","Data":"a1aeda0f76c4a3989dbd6ea69ae40b074a48598580bdf4195f0495834505d8a2"} Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.953133 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a4b9e543-db0e-4b6a-bd32-9935f603556e","Type":"ContainerStarted","Data":"02a64e9573f0b20211d288c4b9099f46410eeaadf1e6a3659a19c462db8c550b"} Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.954316 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d204fe1c-3288-4837-992f-a04ddea6dbd3","Type":"ContainerStarted","Data":"dd085c8bf25a8c9a5becee29a8178c0afe5712b089d9d7fffd72f63deae2b02f"} Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.955777 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bd877c5c-t5pxf" event={"ID":"cc996031-bc38-4f18-82b1-c7adf611b54b","Type":"ContainerDied","Data":"c8d2af4f3291aef0078855d4beb941c6075ddea942bc81014f37930128b12f54"} Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.955851 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bd877c5c-t5pxf" Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.959846 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f2325e29-9f4c-438f-9400-e643e2494808","Type":"ContainerStarted","Data":"dfb09b46dd99cea54fc957bd35e6a0ba8b7e7a98c8e7fa29c897416977c7b054"} Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.962389 4700 generic.go:334] "Generic (PLEG): container finished" podID="cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90" containerID="938e831da689d378d7521ce51730c1b6ba7a0eae98f18eae14142909ce26a608" exitCode=0 Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.963105 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bd79c76b5-52bjb" event={"ID":"cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90","Type":"ContainerDied","Data":"938e831da689d378d7521ce51730c1b6ba7a0eae98f18eae14142909ce26a608"} Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.966656 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223","Type":"ContainerStarted","Data":"8f8708d2b15678deef6477dd345ccf3a3d9dda386516132bb8e84f079e634f4c"} Feb 27 17:19:44 crc kubenswrapper[4700]: I0227 17:19:44.968025 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-qqqm8" event={"ID":"c7fc37b2-4775-468e-b118-e399646a95e9","Type":"ContainerStarted","Data":"ba9a2a9fff7eeaeda512c1331a532b0adf8eed31022e44ca7716a5d51bce2634"} Feb 27 17:19:45 crc kubenswrapper[4700]: I0227 17:19:45.041578 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f4c94c77-hw5gz"] Feb 27 17:19:45 crc kubenswrapper[4700]: I0227 17:19:45.053661 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f4c94c77-hw5gz"] Feb 27 17:19:45 crc kubenswrapper[4700]: I0227 17:19:45.087411 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bd877c5c-t5pxf"] Feb 27 17:19:45 crc kubenswrapper[4700]: I0227 17:19:45.095081 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bd877c5c-t5pxf"] Feb 27 17:19:47 crc kubenswrapper[4700]: I0227 17:19:47.009361 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26b9a2b0-0140-415e-911b-14fa2eb6c3ff" path="/var/lib/kubelet/pods/26b9a2b0-0140-415e-911b-14fa2eb6c3ff/volumes" Feb 27 17:19:47 crc kubenswrapper[4700]: I0227 17:19:47.010098 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc996031-bc38-4f18-82b1-c7adf611b54b" path="/var/lib/kubelet/pods/cc996031-bc38-4f18-82b1-c7adf611b54b/volumes" Feb 27 17:19:58 crc kubenswrapper[4700]: E0227 17:19:58.746244 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.145:5001/podified-master-centos10/openstack-rabbitmq:watcher_latest" Feb 27 17:19:58 crc kubenswrapper[4700]: E0227 17:19:58.747065 4700 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.145:5001/podified-master-centos10/openstack-rabbitmq:watcher_latest" Feb 27 17:19:58 crc kubenswrapper[4700]: E0227 17:19:58.747261 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:38.102.83.145:5001/podified-master-centos10/openstack-rabbitmq:watcher_latest,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rw6kl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(efe7d637-da3a-4995-a40f-fae00257ac1d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 17:19:58 crc kubenswrapper[4700]: E0227 17:19:58.748806 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="efe7d637-da3a-4995-a40f-fae00257ac1d" Feb 27 17:19:58 crc kubenswrapper[4700]: E0227 17:19:58.756901 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.145:5001/podified-master-centos10/openstack-rabbitmq:watcher_latest" Feb 27 17:19:58 crc kubenswrapper[4700]: E0227 17:19:58.756953 4700 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.145:5001/podified-master-centos10/openstack-rabbitmq:watcher_latest" Feb 27 17:19:58 crc kubenswrapper[4700]: E0227 17:19:58.757097 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:38.102.83.145:5001/podified-master-centos10/openstack-rabbitmq:watcher_latest,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dbjsk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(458e6422-b856-4bcf-8308-4b6cb9ec8fef): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 17:19:58 crc kubenswrapper[4700]: E0227 17:19:58.758385 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="458e6422-b856-4bcf-8308-4b6cb9ec8fef" Feb 27 17:19:59 crc kubenswrapper[4700]: E0227 17:19:59.047319 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.145:5001/podified-master-centos10/openstack-ovn-nb-db-server:watcher_latest" Feb 27 17:19:59 crc kubenswrapper[4700]: E0227 17:19:59.047383 4700 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.145:5001/podified-master-centos10/openstack-ovn-nb-db-server:watcher_latest" Feb 27 17:19:59 crc kubenswrapper[4700]: E0227 17:19:59.047606 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovsdbserver-nb,Image:38.102.83.145:5001/podified-master-centos10/openstack-ovn-nb-db-server:watcher_latest,Command:[/usr/bin/dumb-init],Args:[/usr/local/bin/container-scripts/setup.sh],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n6fh55h68bh696h5c4h95hb7h78h5d6h646h5c6h645h7ch5c4h5bh578hc8hb7h7hb4h646h98h678hcbh655h5d9h7dh65dh675h55dh689h9dq,ValueFrom:nil,},EnvVar{Name:OVN_LOGDIR,Value:/tmp,ValueFrom:nil,},EnvVar{Name:OVN_RUNDIR,Value:/tmp,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovndbcluster-nb-etc-ovn,ReadOnly:false,MountPath:/etc/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qwzx9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/cleanup.sh],},HTTPGet:nil,TCPSocket:nil,Sleep:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:20,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-nb-0_openstack(a4b9e543-db0e-4b6a-bd32-9935f603556e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 17:19:59 crc kubenswrapper[4700]: E0227 17:19:59.068716 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.145:5001/podified-master-centos10/openstack-mariadb:watcher_latest" Feb 27 17:19:59 crc kubenswrapper[4700]: E0227 17:19:59.068775 4700 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.145:5001/podified-master-centos10/openstack-mariadb:watcher_latest" Feb 27 17:19:59 crc kubenswrapper[4700]: E0227 17:19:59.068937 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:38.102.83.145:5001/podified-master-centos10/openstack-mariadb:watcher_latest,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2dphd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-cell1-galera-0_openstack(4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 17:19:59 crc kubenswrapper[4700]: E0227 17:19:59.070310 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-cell1-galera-0" podUID="4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223" Feb 27 17:19:59 crc kubenswrapper[4700]: E0227 17:19:59.076793 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.145:5001/podified-master-centos10/openstack-rabbitmq:watcher_latest" Feb 27 17:19:59 crc kubenswrapper[4700]: E0227 17:19:59.076850 4700 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.145:5001/podified-master-centos10/openstack-rabbitmq:watcher_latest" Feb 27 17:19:59 crc kubenswrapper[4700]: E0227 17:19:59.076990 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:38.102.83.145:5001/podified-master-centos10/openstack-rabbitmq:watcher_latest,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-clzkh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod notifications-rabbitmq-server-0_openstack(fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 17:19:59 crc kubenswrapper[4700]: E0227 17:19:59.078435 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/notifications-rabbitmq-server-0" podUID="fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da" Feb 27 17:19:59 crc kubenswrapper[4700]: E0227 17:19:59.121211 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.145:5001/podified-master-centos10/openstack-rabbitmq:watcher_latest\\\"\"" pod="openstack/rabbitmq-server-0" podUID="efe7d637-da3a-4995-a40f-fae00257ac1d" Feb 27 17:19:59 crc kubenswrapper[4700]: E0227 17:19:59.121217 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.145:5001/podified-master-centos10/openstack-rabbitmq:watcher_latest\\\"\"" pod="openstack/notifications-rabbitmq-server-0" podUID="fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da" Feb 27 17:19:59 crc kubenswrapper[4700]: E0227 17:19:59.121666 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.145:5001/podified-master-centos10/openstack-mariadb:watcher_latest\\\"\"" pod="openstack/openstack-cell1-galera-0" podUID="4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223" Feb 27 17:19:59 crc kubenswrapper[4700]: E0227 17:19:59.121617 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.145:5001/podified-master-centos10/openstack-rabbitmq:watcher_latest\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="458e6422-b856-4bcf-8308-4b6cb9ec8fef" Feb 27 17:20:00 crc kubenswrapper[4700]: I0227 17:20:00.134918 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536880-7gbpd"] Feb 27 17:20:00 crc kubenswrapper[4700]: I0227 17:20:00.136793 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536880-7gbpd" Feb 27 17:20:00 crc kubenswrapper[4700]: I0227 17:20:00.142651 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:20:00 crc kubenswrapper[4700]: I0227 17:20:00.143053 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:20:00 crc kubenswrapper[4700]: I0227 17:20:00.143517 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d62kc" Feb 27 17:20:00 crc kubenswrapper[4700]: I0227 17:20:00.161273 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfr6w\" (UniqueName: \"kubernetes.io/projected/4392a4d6-4e34-4821-8350-a82f8c06fc8d-kube-api-access-vfr6w\") pod \"auto-csr-approver-29536880-7gbpd\" (UID: \"4392a4d6-4e34-4821-8350-a82f8c06fc8d\") " pod="openshift-infra/auto-csr-approver-29536880-7gbpd" Feb 27 17:20:00 crc kubenswrapper[4700]: I0227 17:20:00.167288 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536880-7gbpd"] Feb 27 17:20:00 crc kubenswrapper[4700]: I0227 17:20:00.272962 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfr6w\" (UniqueName: \"kubernetes.io/projected/4392a4d6-4e34-4821-8350-a82f8c06fc8d-kube-api-access-vfr6w\") pod \"auto-csr-approver-29536880-7gbpd\" (UID: \"4392a4d6-4e34-4821-8350-a82f8c06fc8d\") " pod="openshift-infra/auto-csr-approver-29536880-7gbpd" Feb 27 17:20:00 crc kubenswrapper[4700]: I0227 17:20:00.292907 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfr6w\" (UniqueName: \"kubernetes.io/projected/4392a4d6-4e34-4821-8350-a82f8c06fc8d-kube-api-access-vfr6w\") pod \"auto-csr-approver-29536880-7gbpd\" (UID: \"4392a4d6-4e34-4821-8350-a82f8c06fc8d\") " pod="openshift-infra/auto-csr-approver-29536880-7gbpd" Feb 27 17:20:00 crc kubenswrapper[4700]: I0227 17:20:00.456091 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536880-7gbpd" Feb 27 17:20:00 crc kubenswrapper[4700]: E0227 17:20:00.858780 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Feb 27 17:20:00 crc kubenswrapper[4700]: E0227 17:20:00.858846 4700 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Feb 27 17:20:00 crc kubenswrapper[4700]: E0227 17:20:00.858984 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-state-metrics,Image:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,Command:[],Args:[--resources=pods --namespaces=openstack],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},ContainerPort{Name:telemetry,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bpmhn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{0 8080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod kube-state-metrics-0_openstack(d204fe1c-3288-4837-992f-a04ddea6dbd3): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 27 17:20:00 crc kubenswrapper[4700]: E0227 17:20:00.860182 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/kube-state-metrics-0" podUID="d204fe1c-3288-4837-992f-a04ddea6dbd3" Feb 27 17:20:01 crc kubenswrapper[4700]: I0227 17:20:01.148842 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bbc67fc-tr49q" event={"ID":"4448288d-21f1-418b-95bf-9287e914f52b","Type":"ContainerStarted","Data":"36a7354b8a1d35885156629f2499a02f10cf0989a40e38c49fba3fae4670d9aa"} Feb 27 17:20:01 crc kubenswrapper[4700]: E0227 17:20:01.150196 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0\\\"\"" pod="openstack/kube-state-metrics-0" podUID="d204fe1c-3288-4837-992f-a04ddea6dbd3" Feb 27 17:20:01 crc kubenswrapper[4700]: I0227 17:20:01.197280 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7bbc67fc-tr49q" podStartSLOduration=35.251008666 podStartE2EDuration="36.197257266s" podCreationTimestamp="2026-02-27 17:19:25 +0000 UTC" firstStartedPulling="2026-02-27 17:19:42.208542732 +0000 UTC m=+1142.193855499" lastFinishedPulling="2026-02-27 17:19:43.154791352 +0000 UTC m=+1143.140104099" observedRunningTime="2026-02-27 17:20:01.19102374 +0000 UTC m=+1161.176336497" watchObservedRunningTime="2026-02-27 17:20:01.197257266 +0000 UTC m=+1161.182570043" Feb 27 17:20:02 crc kubenswrapper[4700]: I0227 17:20:02.173201 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7bbc67fc-tr49q" Feb 27 17:20:03 crc kubenswrapper[4700]: I0227 17:20:03.181443 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bd79c76b5-52bjb" event={"ID":"cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90","Type":"ContainerStarted","Data":"db1c6a8ab62953cd151732875f286cc94432bc85b8e15ce6b67ca51d01abecc4"} Feb 27 17:20:03 crc kubenswrapper[4700]: I0227 17:20:03.317352 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bd79c76b5-52bjb" podStartSLOduration=37.317335936 podStartE2EDuration="37.317335936s" podCreationTimestamp="2026-02-27 17:19:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:20:03.198077822 +0000 UTC m=+1163.183390569" watchObservedRunningTime="2026-02-27 17:20:03.317335936 +0000 UTC m=+1163.302648673" Feb 27 17:20:03 crc kubenswrapper[4700]: I0227 17:20:03.318720 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536880-7gbpd"] Feb 27 17:20:03 crc kubenswrapper[4700]: W0227 17:20:03.352383 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4392a4d6_4e34_4821_8350_a82f8c06fc8d.slice/crio-02b5f453382eb502d0659244f72e4887c32f43588108e113e745dfaa46575d42 WatchSource:0}: Error finding container 02b5f453382eb502d0659244f72e4887c32f43588108e113e745dfaa46575d42: Status 404 returned error can't find the container with id 02b5f453382eb502d0659244f72e4887c32f43588108e113e745dfaa46575d42 Feb 27 17:20:03 crc kubenswrapper[4700]: I0227 17:20:03.357392 4700 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 17:20:03 crc kubenswrapper[4700]: E0227 17:20:03.513572 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-nb-0" podUID="a4b9e543-db0e-4b6a-bd32-9935f603556e" Feb 27 17:20:04 crc kubenswrapper[4700]: I0227 17:20:04.191640 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a4b9e543-db0e-4b6a-bd32-9935f603556e","Type":"ContainerStarted","Data":"d5965271ece4736f1476986c87048523b3dc924e282c8d5e7a851740c7ff02c5"} Feb 27 17:20:04 crc kubenswrapper[4700]: E0227 17:20:04.193567 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.145:5001/podified-master-centos10/openstack-ovn-nb-db-server:watcher_latest\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="a4b9e543-db0e-4b6a-bd32-9935f603556e" Feb 27 17:20:04 crc kubenswrapper[4700]: I0227 17:20:04.194252 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e","Type":"ContainerStarted","Data":"ea79ffb4319c5fbb6fe2d1b0760123f9cad189fe40820ece15d4688c76ae7f67"} Feb 27 17:20:04 crc kubenswrapper[4700]: I0227 17:20:04.194277 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e","Type":"ContainerStarted","Data":"0078fc8c9c42bc880abd02f23b2bd9f45cefa78f6556ceae21a7c4bba39d9414"} Feb 27 17:20:04 crc kubenswrapper[4700]: I0227 17:20:04.195836 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536880-7gbpd" event={"ID":"4392a4d6-4e34-4821-8350-a82f8c06fc8d","Type":"ContainerStarted","Data":"02b5f453382eb502d0659244f72e4887c32f43588108e113e745dfaa46575d42"} Feb 27 17:20:04 crc kubenswrapper[4700]: I0227 17:20:04.197691 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f2325e29-9f4c-438f-9400-e643e2494808","Type":"ContainerStarted","Data":"9ff3a5239c9efca695db63b517f3295d9533dc87d43cf34a5d63ab1cd090d6e8"} Feb 27 17:20:04 crc kubenswrapper[4700]: I0227 17:20:04.199976 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-sfldb" event={"ID":"a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a","Type":"ContainerStarted","Data":"8a64210b36e5339a12ab8f68ab3ff7d507087d8c31048dcbd88d308f8633c9b5"} Feb 27 17:20:04 crc kubenswrapper[4700]: I0227 17:20:04.200260 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-sfldb" Feb 27 17:20:04 crc kubenswrapper[4700]: I0227 17:20:04.202646 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"9cf3890b-d331-4312-a6ce-a49945e61147","Type":"ContainerStarted","Data":"cdebbc00a629d371f4d10115088d71ad7b7d8cd5084f718f8021c4c371a65996"} Feb 27 17:20:04 crc kubenswrapper[4700]: I0227 17:20:04.203422 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Feb 27 17:20:04 crc kubenswrapper[4700]: I0227 17:20:04.207423 4700 generic.go:334] "Generic (PLEG): container finished" podID="c7fc37b2-4775-468e-b118-e399646a95e9" containerID="5a67b4422a553e1daf9e018110b708a9ae40532416f19270dda25bea39fafa76" exitCode=0 Feb 27 17:20:04 crc kubenswrapper[4700]: I0227 17:20:04.207638 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-qqqm8" event={"ID":"c7fc37b2-4775-468e-b118-e399646a95e9","Type":"ContainerDied","Data":"5a67b4422a553e1daf9e018110b708a9ae40532416f19270dda25bea39fafa76"} Feb 27 17:20:04 crc kubenswrapper[4700]: I0227 17:20:04.207695 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bd79c76b5-52bjb" Feb 27 17:20:04 crc kubenswrapper[4700]: I0227 17:20:04.295408 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=19.009666355 podStartE2EDuration="34.29538597s" podCreationTimestamp="2026-02-27 17:19:30 +0000 UTC" firstStartedPulling="2026-02-27 17:19:44.251420543 +0000 UTC m=+1144.236733290" lastFinishedPulling="2026-02-27 17:19:59.537140158 +0000 UTC m=+1159.522452905" observedRunningTime="2026-02-27 17:20:04.260014672 +0000 UTC m=+1164.245327529" watchObservedRunningTime="2026-02-27 17:20:04.29538597 +0000 UTC m=+1164.280698727" Feb 27 17:20:04 crc kubenswrapper[4700]: I0227 17:20:04.318098 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-sfldb" podStartSLOduration=14.063052262 podStartE2EDuration="29.318080412s" podCreationTimestamp="2026-02-27 17:19:35 +0000 UTC" firstStartedPulling="2026-02-27 17:19:44.290789258 +0000 UTC m=+1144.276102005" lastFinishedPulling="2026-02-27 17:19:59.545817388 +0000 UTC m=+1159.531130155" observedRunningTime="2026-02-27 17:20:04.309454203 +0000 UTC m=+1164.294766960" watchObservedRunningTime="2026-02-27 17:20:04.318080412 +0000 UTC m=+1164.303393179" Feb 27 17:20:04 crc kubenswrapper[4700]: I0227 17:20:04.345855 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Feb 27 17:20:04 crc kubenswrapper[4700]: I0227 17:20:04.377743 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=11.389970083 podStartE2EDuration="26.377721844s" podCreationTimestamp="2026-02-27 17:19:38 +0000 UTC" firstStartedPulling="2026-02-27 17:19:44.780011595 +0000 UTC m=+1144.765324342" lastFinishedPulling="2026-02-27 17:19:59.767763316 +0000 UTC m=+1159.753076103" observedRunningTime="2026-02-27 17:20:04.374289233 +0000 UTC m=+1164.359601990" watchObservedRunningTime="2026-02-27 17:20:04.377721844 +0000 UTC m=+1164.363034611" Feb 27 17:20:05 crc kubenswrapper[4700]: I0227 17:20:05.219495 4700 generic.go:334] "Generic (PLEG): container finished" podID="4392a4d6-4e34-4821-8350-a82f8c06fc8d" containerID="812a90f84653ff6064934e96d31046dc02493a88565d50828f107fbab0eaf0c1" exitCode=0 Feb 27 17:20:05 crc kubenswrapper[4700]: I0227 17:20:05.219734 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536880-7gbpd" event={"ID":"4392a4d6-4e34-4821-8350-a82f8c06fc8d","Type":"ContainerDied","Data":"812a90f84653ff6064934e96d31046dc02493a88565d50828f107fbab0eaf0c1"} Feb 27 17:20:05 crc kubenswrapper[4700]: I0227 17:20:05.224319 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-qqqm8" event={"ID":"c7fc37b2-4775-468e-b118-e399646a95e9","Type":"ContainerStarted","Data":"80e097aec682af1e75c0bc24840db57105e915163d74e66a6660002f2dacfcf1"} Feb 27 17:20:05 crc kubenswrapper[4700]: I0227 17:20:05.224381 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-qqqm8" event={"ID":"c7fc37b2-4775-468e-b118-e399646a95e9","Type":"ContainerStarted","Data":"a9b4b587a952398839529ea6a91a97d13adf97a540315570dd33fce70227e3e8"} Feb 27 17:20:05 crc kubenswrapper[4700]: E0227 17:20:05.226988 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.145:5001/podified-master-centos10/openstack-ovn-nb-db-server:watcher_latest\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="a4b9e543-db0e-4b6a-bd32-9935f603556e" Feb 27 17:20:05 crc kubenswrapper[4700]: I0227 17:20:05.321782 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-qqqm8" podStartSLOduration=15.184291665 podStartE2EDuration="30.321751637s" podCreationTimestamp="2026-02-27 17:19:35 +0000 UTC" firstStartedPulling="2026-02-27 17:19:44.399683266 +0000 UTC m=+1144.384996013" lastFinishedPulling="2026-02-27 17:19:59.537143228 +0000 UTC m=+1159.522455985" observedRunningTime="2026-02-27 17:20:05.310533939 +0000 UTC m=+1165.295846726" watchObservedRunningTime="2026-02-27 17:20:05.321751637 +0000 UTC m=+1165.307064424" Feb 27 17:20:05 crc kubenswrapper[4700]: I0227 17:20:05.339492 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Feb 27 17:20:05 crc kubenswrapper[4700]: I0227 17:20:05.942051 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-qqqm8" Feb 27 17:20:05 crc kubenswrapper[4700]: I0227 17:20:05.942121 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-qqqm8" Feb 27 17:20:06 crc kubenswrapper[4700]: I0227 17:20:06.002608 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7bbc67fc-tr49q" Feb 27 17:20:06 crc kubenswrapper[4700]: I0227 17:20:06.233179 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9721431e-046b-4b3b-861e-22a65396981b","Type":"ContainerStarted","Data":"750401c839d4786183e640b5a5874d5afec2bc8dbd5e838d8a1d75f3a9e675a8"} Feb 27 17:20:06 crc kubenswrapper[4700]: I0227 17:20:06.592424 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536880-7gbpd" Feb 27 17:20:06 crc kubenswrapper[4700]: I0227 17:20:06.633828 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfr6w\" (UniqueName: \"kubernetes.io/projected/4392a4d6-4e34-4821-8350-a82f8c06fc8d-kube-api-access-vfr6w\") pod \"4392a4d6-4e34-4821-8350-a82f8c06fc8d\" (UID: \"4392a4d6-4e34-4821-8350-a82f8c06fc8d\") " Feb 27 17:20:06 crc kubenswrapper[4700]: I0227 17:20:06.640544 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4392a4d6-4e34-4821-8350-a82f8c06fc8d-kube-api-access-vfr6w" (OuterVolumeSpecName: "kube-api-access-vfr6w") pod "4392a4d6-4e34-4821-8350-a82f8c06fc8d" (UID: "4392a4d6-4e34-4821-8350-a82f8c06fc8d"). InnerVolumeSpecName "kube-api-access-vfr6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:20:06 crc kubenswrapper[4700]: I0227 17:20:06.735663 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfr6w\" (UniqueName: \"kubernetes.io/projected/4392a4d6-4e34-4821-8350-a82f8c06fc8d-kube-api-access-vfr6w\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:07 crc kubenswrapper[4700]: I0227 17:20:07.243408 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536880-7gbpd" event={"ID":"4392a4d6-4e34-4821-8350-a82f8c06fc8d","Type":"ContainerDied","Data":"02b5f453382eb502d0659244f72e4887c32f43588108e113e745dfaa46575d42"} Feb 27 17:20:07 crc kubenswrapper[4700]: I0227 17:20:07.243504 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02b5f453382eb502d0659244f72e4887c32f43588108e113e745dfaa46575d42" Feb 27 17:20:07 crc kubenswrapper[4700]: I0227 17:20:07.243718 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536880-7gbpd" Feb 27 17:20:07 crc kubenswrapper[4700]: I0227 17:20:07.680097 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536874-g79g6"] Feb 27 17:20:07 crc kubenswrapper[4700]: I0227 17:20:07.687559 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536874-g79g6"] Feb 27 17:20:07 crc kubenswrapper[4700]: I0227 17:20:07.863387 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.309986 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.616666 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bd79c76b5-52bjb"] Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.616930 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bd79c76b5-52bjb" podUID="cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90" containerName="dnsmasq-dns" containerID="cri-o://db1c6a8ab62953cd151732875f286cc94432bc85b8e15ce6b67ca51d01abecc4" gracePeriod=10 Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.619772 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6bd79c76b5-52bjb" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.671605 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-7r49n"] Feb 27 17:20:08 crc kubenswrapper[4700]: E0227 17:20:08.671996 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4392a4d6-4e34-4821-8350-a82f8c06fc8d" containerName="oc" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.672009 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="4392a4d6-4e34-4821-8350-a82f8c06fc8d" containerName="oc" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.672157 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="4392a4d6-4e34-4821-8350-a82f8c06fc8d" containerName="oc" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.672752 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-7r49n" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.676358 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.692179 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-857d474bb7-cd9zz"] Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.693542 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-857d474bb7-cd9zz" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.699165 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.724649 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-857d474bb7-cd9zz"] Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.735052 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-7r49n"] Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.769469 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a439f84-1aec-4aa7-bd1d-6e0683425b5b-dns-svc\") pod \"dnsmasq-dns-857d474bb7-cd9zz\" (UID: \"5a439f84-1aec-4aa7-bd1d-6e0683425b5b\") " pod="openstack/dnsmasq-dns-857d474bb7-cd9zz" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.769967 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a439f84-1aec-4aa7-bd1d-6e0683425b5b-config\") pod \"dnsmasq-dns-857d474bb7-cd9zz\" (UID: \"5a439f84-1aec-4aa7-bd1d-6e0683425b5b\") " pod="openstack/dnsmasq-dns-857d474bb7-cd9zz" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.770058 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5a439f84-1aec-4aa7-bd1d-6e0683425b5b-ovsdbserver-sb\") pod \"dnsmasq-dns-857d474bb7-cd9zz\" (UID: \"5a439f84-1aec-4aa7-bd1d-6e0683425b5b\") " pod="openstack/dnsmasq-dns-857d474bb7-cd9zz" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.770147 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/faaab80d-004f-48ab-b216-e687a0769a7b-ovn-rundir\") pod \"ovn-controller-metrics-7r49n\" (UID: \"faaab80d-004f-48ab-b216-e687a0769a7b\") " pod="openstack/ovn-controller-metrics-7r49n" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.770276 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqrs9\" (UniqueName: \"kubernetes.io/projected/faaab80d-004f-48ab-b216-e687a0769a7b-kube-api-access-bqrs9\") pod \"ovn-controller-metrics-7r49n\" (UID: \"faaab80d-004f-48ab-b216-e687a0769a7b\") " pod="openstack/ovn-controller-metrics-7r49n" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.770360 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrvhp\" (UniqueName: \"kubernetes.io/projected/5a439f84-1aec-4aa7-bd1d-6e0683425b5b-kube-api-access-hrvhp\") pod \"dnsmasq-dns-857d474bb7-cd9zz\" (UID: \"5a439f84-1aec-4aa7-bd1d-6e0683425b5b\") " pod="openstack/dnsmasq-dns-857d474bb7-cd9zz" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.770446 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/faaab80d-004f-48ab-b216-e687a0769a7b-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-7r49n\" (UID: \"faaab80d-004f-48ab-b216-e687a0769a7b\") " pod="openstack/ovn-controller-metrics-7r49n" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.770551 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/faaab80d-004f-48ab-b216-e687a0769a7b-combined-ca-bundle\") pod \"ovn-controller-metrics-7r49n\" (UID: \"faaab80d-004f-48ab-b216-e687a0769a7b\") " pod="openstack/ovn-controller-metrics-7r49n" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.770625 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/faaab80d-004f-48ab-b216-e687a0769a7b-ovs-rundir\") pod \"ovn-controller-metrics-7r49n\" (UID: \"faaab80d-004f-48ab-b216-e687a0769a7b\") " pod="openstack/ovn-controller-metrics-7r49n" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.770718 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/faaab80d-004f-48ab-b216-e687a0769a7b-config\") pod \"ovn-controller-metrics-7r49n\" (UID: \"faaab80d-004f-48ab-b216-e687a0769a7b\") " pod="openstack/ovn-controller-metrics-7r49n" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.871817 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a439f84-1aec-4aa7-bd1d-6e0683425b5b-config\") pod \"dnsmasq-dns-857d474bb7-cd9zz\" (UID: \"5a439f84-1aec-4aa7-bd1d-6e0683425b5b\") " pod="openstack/dnsmasq-dns-857d474bb7-cd9zz" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.871865 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5a439f84-1aec-4aa7-bd1d-6e0683425b5b-ovsdbserver-sb\") pod \"dnsmasq-dns-857d474bb7-cd9zz\" (UID: \"5a439f84-1aec-4aa7-bd1d-6e0683425b5b\") " pod="openstack/dnsmasq-dns-857d474bb7-cd9zz" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.871890 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/faaab80d-004f-48ab-b216-e687a0769a7b-ovn-rundir\") pod \"ovn-controller-metrics-7r49n\" (UID: \"faaab80d-004f-48ab-b216-e687a0769a7b\") " pod="openstack/ovn-controller-metrics-7r49n" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.871939 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqrs9\" (UniqueName: \"kubernetes.io/projected/faaab80d-004f-48ab-b216-e687a0769a7b-kube-api-access-bqrs9\") pod \"ovn-controller-metrics-7r49n\" (UID: \"faaab80d-004f-48ab-b216-e687a0769a7b\") " pod="openstack/ovn-controller-metrics-7r49n" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.871959 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrvhp\" (UniqueName: \"kubernetes.io/projected/5a439f84-1aec-4aa7-bd1d-6e0683425b5b-kube-api-access-hrvhp\") pod \"dnsmasq-dns-857d474bb7-cd9zz\" (UID: \"5a439f84-1aec-4aa7-bd1d-6e0683425b5b\") " pod="openstack/dnsmasq-dns-857d474bb7-cd9zz" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.871980 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/faaab80d-004f-48ab-b216-e687a0769a7b-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-7r49n\" (UID: \"faaab80d-004f-48ab-b216-e687a0769a7b\") " pod="openstack/ovn-controller-metrics-7r49n" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.871999 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/faaab80d-004f-48ab-b216-e687a0769a7b-combined-ca-bundle\") pod \"ovn-controller-metrics-7r49n\" (UID: \"faaab80d-004f-48ab-b216-e687a0769a7b\") " pod="openstack/ovn-controller-metrics-7r49n" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.872018 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/faaab80d-004f-48ab-b216-e687a0769a7b-ovs-rundir\") pod \"ovn-controller-metrics-7r49n\" (UID: \"faaab80d-004f-48ab-b216-e687a0769a7b\") " pod="openstack/ovn-controller-metrics-7r49n" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.872041 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/faaab80d-004f-48ab-b216-e687a0769a7b-config\") pod \"ovn-controller-metrics-7r49n\" (UID: \"faaab80d-004f-48ab-b216-e687a0769a7b\") " pod="openstack/ovn-controller-metrics-7r49n" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.872065 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a439f84-1aec-4aa7-bd1d-6e0683425b5b-dns-svc\") pod \"dnsmasq-dns-857d474bb7-cd9zz\" (UID: \"5a439f84-1aec-4aa7-bd1d-6e0683425b5b\") " pod="openstack/dnsmasq-dns-857d474bb7-cd9zz" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.872802 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5a439f84-1aec-4aa7-bd1d-6e0683425b5b-ovsdbserver-sb\") pod \"dnsmasq-dns-857d474bb7-cd9zz\" (UID: \"5a439f84-1aec-4aa7-bd1d-6e0683425b5b\") " pod="openstack/dnsmasq-dns-857d474bb7-cd9zz" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.872985 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a439f84-1aec-4aa7-bd1d-6e0683425b5b-config\") pod \"dnsmasq-dns-857d474bb7-cd9zz\" (UID: \"5a439f84-1aec-4aa7-bd1d-6e0683425b5b\") " pod="openstack/dnsmasq-dns-857d474bb7-cd9zz" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.873097 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/faaab80d-004f-48ab-b216-e687a0769a7b-ovn-rundir\") pod \"ovn-controller-metrics-7r49n\" (UID: \"faaab80d-004f-48ab-b216-e687a0769a7b\") " pod="openstack/ovn-controller-metrics-7r49n" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.873158 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/faaab80d-004f-48ab-b216-e687a0769a7b-ovs-rundir\") pod \"ovn-controller-metrics-7r49n\" (UID: \"faaab80d-004f-48ab-b216-e687a0769a7b\") " pod="openstack/ovn-controller-metrics-7r49n" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.877048 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a439f84-1aec-4aa7-bd1d-6e0683425b5b-dns-svc\") pod \"dnsmasq-dns-857d474bb7-cd9zz\" (UID: \"5a439f84-1aec-4aa7-bd1d-6e0683425b5b\") " pod="openstack/dnsmasq-dns-857d474bb7-cd9zz" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.877295 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/faaab80d-004f-48ab-b216-e687a0769a7b-config\") pod \"ovn-controller-metrics-7r49n\" (UID: \"faaab80d-004f-48ab-b216-e687a0769a7b\") " pod="openstack/ovn-controller-metrics-7r49n" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.878817 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/faaab80d-004f-48ab-b216-e687a0769a7b-combined-ca-bundle\") pod \"ovn-controller-metrics-7r49n\" (UID: \"faaab80d-004f-48ab-b216-e687a0769a7b\") " pod="openstack/ovn-controller-metrics-7r49n" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.882050 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/faaab80d-004f-48ab-b216-e687a0769a7b-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-7r49n\" (UID: \"faaab80d-004f-48ab-b216-e687a0769a7b\") " pod="openstack/ovn-controller-metrics-7r49n" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.900575 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqrs9\" (UniqueName: \"kubernetes.io/projected/faaab80d-004f-48ab-b216-e687a0769a7b-kube-api-access-bqrs9\") pod \"ovn-controller-metrics-7r49n\" (UID: \"faaab80d-004f-48ab-b216-e687a0769a7b\") " pod="openstack/ovn-controller-metrics-7r49n" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.901423 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrvhp\" (UniqueName: \"kubernetes.io/projected/5a439f84-1aec-4aa7-bd1d-6e0683425b5b-kube-api-access-hrvhp\") pod \"dnsmasq-dns-857d474bb7-cd9zz\" (UID: \"5a439f84-1aec-4aa7-bd1d-6e0683425b5b\") " pod="openstack/dnsmasq-dns-857d474bb7-cd9zz" Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.979308 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-857d474bb7-cd9zz"] Feb 27 17:20:08 crc kubenswrapper[4700]: I0227 17:20:08.979914 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-857d474bb7-cd9zz" Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.014407 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e2f6b19-ca84-4f18-b030-420a69fc6dce" path="/var/lib/kubelet/pods/5e2f6b19-ca84-4f18-b030-420a69fc6dce/volumes" Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.015065 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bb977cfbf-vkv4n"] Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.016282 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bb977cfbf-vkv4n" Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.022379 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.023265 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-7r49n" Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.036944 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bb977cfbf-vkv4n"] Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.075933 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23ac2574-09ac-4796-89b9-8581f4a73a00-config\") pod \"dnsmasq-dns-5bb977cfbf-vkv4n\" (UID: \"23ac2574-09ac-4796-89b9-8581f4a73a00\") " pod="openstack/dnsmasq-dns-5bb977cfbf-vkv4n" Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.076036 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/23ac2574-09ac-4796-89b9-8581f4a73a00-ovsdbserver-sb\") pod \"dnsmasq-dns-5bb977cfbf-vkv4n\" (UID: \"23ac2574-09ac-4796-89b9-8581f4a73a00\") " pod="openstack/dnsmasq-dns-5bb977cfbf-vkv4n" Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.076057 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/23ac2574-09ac-4796-89b9-8581f4a73a00-ovsdbserver-nb\") pod \"dnsmasq-dns-5bb977cfbf-vkv4n\" (UID: \"23ac2574-09ac-4796-89b9-8581f4a73a00\") " pod="openstack/dnsmasq-dns-5bb977cfbf-vkv4n" Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.076157 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/23ac2574-09ac-4796-89b9-8581f4a73a00-dns-svc\") pod \"dnsmasq-dns-5bb977cfbf-vkv4n\" (UID: \"23ac2574-09ac-4796-89b9-8581f4a73a00\") " pod="openstack/dnsmasq-dns-5bb977cfbf-vkv4n" Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.076186 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7fcs\" (UniqueName: \"kubernetes.io/projected/23ac2574-09ac-4796-89b9-8581f4a73a00-kube-api-access-q7fcs\") pod \"dnsmasq-dns-5bb977cfbf-vkv4n\" (UID: \"23ac2574-09ac-4796-89b9-8581f4a73a00\") " pod="openstack/dnsmasq-dns-5bb977cfbf-vkv4n" Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.132776 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bd79c76b5-52bjb" Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.179968 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90-dns-svc\") pod \"cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90\" (UID: \"cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90\") " Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.180094 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-td4pt\" (UniqueName: \"kubernetes.io/projected/cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90-kube-api-access-td4pt\") pod \"cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90\" (UID: \"cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90\") " Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.180199 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90-config\") pod \"cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90\" (UID: \"cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90\") " Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.180369 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/23ac2574-09ac-4796-89b9-8581f4a73a00-dns-svc\") pod \"dnsmasq-dns-5bb977cfbf-vkv4n\" (UID: \"23ac2574-09ac-4796-89b9-8581f4a73a00\") " pod="openstack/dnsmasq-dns-5bb977cfbf-vkv4n" Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.180409 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7fcs\" (UniqueName: \"kubernetes.io/projected/23ac2574-09ac-4796-89b9-8581f4a73a00-kube-api-access-q7fcs\") pod \"dnsmasq-dns-5bb977cfbf-vkv4n\" (UID: \"23ac2574-09ac-4796-89b9-8581f4a73a00\") " pod="openstack/dnsmasq-dns-5bb977cfbf-vkv4n" Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.180448 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23ac2574-09ac-4796-89b9-8581f4a73a00-config\") pod \"dnsmasq-dns-5bb977cfbf-vkv4n\" (UID: \"23ac2574-09ac-4796-89b9-8581f4a73a00\") " pod="openstack/dnsmasq-dns-5bb977cfbf-vkv4n" Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.180503 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/23ac2574-09ac-4796-89b9-8581f4a73a00-ovsdbserver-sb\") pod \"dnsmasq-dns-5bb977cfbf-vkv4n\" (UID: \"23ac2574-09ac-4796-89b9-8581f4a73a00\") " pod="openstack/dnsmasq-dns-5bb977cfbf-vkv4n" Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.180520 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/23ac2574-09ac-4796-89b9-8581f4a73a00-ovsdbserver-nb\") pod \"dnsmasq-dns-5bb977cfbf-vkv4n\" (UID: \"23ac2574-09ac-4796-89b9-8581f4a73a00\") " pod="openstack/dnsmasq-dns-5bb977cfbf-vkv4n" Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.181308 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/23ac2574-09ac-4796-89b9-8581f4a73a00-ovsdbserver-nb\") pod \"dnsmasq-dns-5bb977cfbf-vkv4n\" (UID: \"23ac2574-09ac-4796-89b9-8581f4a73a00\") " pod="openstack/dnsmasq-dns-5bb977cfbf-vkv4n" Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.188270 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/23ac2574-09ac-4796-89b9-8581f4a73a00-dns-svc\") pod \"dnsmasq-dns-5bb977cfbf-vkv4n\" (UID: \"23ac2574-09ac-4796-89b9-8581f4a73a00\") " pod="openstack/dnsmasq-dns-5bb977cfbf-vkv4n" Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.190585 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/23ac2574-09ac-4796-89b9-8581f4a73a00-ovsdbserver-sb\") pod \"dnsmasq-dns-5bb977cfbf-vkv4n\" (UID: \"23ac2574-09ac-4796-89b9-8581f4a73a00\") " pod="openstack/dnsmasq-dns-5bb977cfbf-vkv4n" Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.191212 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23ac2574-09ac-4796-89b9-8581f4a73a00-config\") pod \"dnsmasq-dns-5bb977cfbf-vkv4n\" (UID: \"23ac2574-09ac-4796-89b9-8581f4a73a00\") " pod="openstack/dnsmasq-dns-5bb977cfbf-vkv4n" Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.203194 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90-kube-api-access-td4pt" (OuterVolumeSpecName: "kube-api-access-td4pt") pod "cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90" (UID: "cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90"). InnerVolumeSpecName "kube-api-access-td4pt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.242070 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7fcs\" (UniqueName: \"kubernetes.io/projected/23ac2574-09ac-4796-89b9-8581f4a73a00-kube-api-access-q7fcs\") pod \"dnsmasq-dns-5bb977cfbf-vkv4n\" (UID: \"23ac2574-09ac-4796-89b9-8581f4a73a00\") " pod="openstack/dnsmasq-dns-5bb977cfbf-vkv4n" Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.287843 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-td4pt\" (UniqueName: \"kubernetes.io/projected/cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90-kube-api-access-td4pt\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.300405 4700 generic.go:334] "Generic (PLEG): container finished" podID="cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90" containerID="db1c6a8ab62953cd151732875f286cc94432bc85b8e15ce6b67ca51d01abecc4" exitCode=0 Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.300524 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bd79c76b5-52bjb" Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.300565 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bd79c76b5-52bjb" event={"ID":"cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90","Type":"ContainerDied","Data":"db1c6a8ab62953cd151732875f286cc94432bc85b8e15ce6b67ca51d01abecc4"} Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.300597 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bd79c76b5-52bjb" event={"ID":"cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90","Type":"ContainerDied","Data":"8f95e01899e24cde97f8711844fa6df575b6bd008ab9127233cfa4f928203396"} Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.300614 4700 scope.go:117] "RemoveContainer" containerID="db1c6a8ab62953cd151732875f286cc94432bc85b8e15ce6b67ca51d01abecc4" Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.314128 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90" (UID: "cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.330131 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90-config" (OuterVolumeSpecName: "config") pod "cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90" (UID: "cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.353606 4700 scope.go:117] "RemoveContainer" containerID="938e831da689d378d7521ce51730c1b6ba7a0eae98f18eae14142909ce26a608" Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.392885 4700 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.392918 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.406467 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bb977cfbf-vkv4n" Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.421519 4700 scope.go:117] "RemoveContainer" containerID="db1c6a8ab62953cd151732875f286cc94432bc85b8e15ce6b67ca51d01abecc4" Feb 27 17:20:09 crc kubenswrapper[4700]: E0227 17:20:09.422237 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db1c6a8ab62953cd151732875f286cc94432bc85b8e15ce6b67ca51d01abecc4\": container with ID starting with db1c6a8ab62953cd151732875f286cc94432bc85b8e15ce6b67ca51d01abecc4 not found: ID does not exist" containerID="db1c6a8ab62953cd151732875f286cc94432bc85b8e15ce6b67ca51d01abecc4" Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.422273 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db1c6a8ab62953cd151732875f286cc94432bc85b8e15ce6b67ca51d01abecc4"} err="failed to get container status \"db1c6a8ab62953cd151732875f286cc94432bc85b8e15ce6b67ca51d01abecc4\": rpc error: code = NotFound desc = could not find container \"db1c6a8ab62953cd151732875f286cc94432bc85b8e15ce6b67ca51d01abecc4\": container with ID starting with db1c6a8ab62953cd151732875f286cc94432bc85b8e15ce6b67ca51d01abecc4 not found: ID does not exist" Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.422296 4700 scope.go:117] "RemoveContainer" containerID="938e831da689d378d7521ce51730c1b6ba7a0eae98f18eae14142909ce26a608" Feb 27 17:20:09 crc kubenswrapper[4700]: E0227 17:20:09.424600 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"938e831da689d378d7521ce51730c1b6ba7a0eae98f18eae14142909ce26a608\": container with ID starting with 938e831da689d378d7521ce51730c1b6ba7a0eae98f18eae14142909ce26a608 not found: ID does not exist" containerID="938e831da689d378d7521ce51730c1b6ba7a0eae98f18eae14142909ce26a608" Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.424634 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"938e831da689d378d7521ce51730c1b6ba7a0eae98f18eae14142909ce26a608"} err="failed to get container status \"938e831da689d378d7521ce51730c1b6ba7a0eae98f18eae14142909ce26a608\": rpc error: code = NotFound desc = could not find container \"938e831da689d378d7521ce51730c1b6ba7a0eae98f18eae14142909ce26a608\": container with ID starting with 938e831da689d378d7521ce51730c1b6ba7a0eae98f18eae14142909ce26a608 not found: ID does not exist" Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.632522 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bd79c76b5-52bjb"] Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.639917 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bd79c76b5-52bjb"] Feb 27 17:20:09 crc kubenswrapper[4700]: W0227 17:20:09.673140 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfaaab80d_004f_48ab_b216_e687a0769a7b.slice/crio-bfa4220537f5648e13ace6693d1b9f5302c496fe874cb1813f67585b097e34f3 WatchSource:0}: Error finding container bfa4220537f5648e13ace6693d1b9f5302c496fe874cb1813f67585b097e34f3: Status 404 returned error can't find the container with id bfa4220537f5648e13ace6693d1b9f5302c496fe874cb1813f67585b097e34f3 Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.674491 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-7r49n"] Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.745110 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-857d474bb7-cd9zz"] Feb 27 17:20:09 crc kubenswrapper[4700]: W0227 17:20:09.768628 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a439f84_1aec_4aa7_bd1d_6e0683425b5b.slice/crio-61dba040d54f2ed01a2ddb3d3964e00120ae55895d113b37610c302f1eac877e WatchSource:0}: Error finding container 61dba040d54f2ed01a2ddb3d3964e00120ae55895d113b37610c302f1eac877e: Status 404 returned error can't find the container with id 61dba040d54f2ed01a2ddb3d3964e00120ae55895d113b37610c302f1eac877e Feb 27 17:20:09 crc kubenswrapper[4700]: I0227 17:20:09.885117 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bb977cfbf-vkv4n"] Feb 27 17:20:10 crc kubenswrapper[4700]: I0227 17:20:10.312898 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223","Type":"ContainerStarted","Data":"b154a054c6a45d944ce284b5f64f55a21d5f1658815d5dcbff5c63fee94cc5c1"} Feb 27 17:20:10 crc kubenswrapper[4700]: I0227 17:20:10.315896 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-7r49n" event={"ID":"faaab80d-004f-48ab-b216-e687a0769a7b","Type":"ContainerStarted","Data":"d22ac7cb3d094f7309cbc7314ab1cae781cdbff407c0b3cf4e55f2004fcf1fab"} Feb 27 17:20:10 crc kubenswrapper[4700]: I0227 17:20:10.315930 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-7r49n" event={"ID":"faaab80d-004f-48ab-b216-e687a0769a7b","Type":"ContainerStarted","Data":"bfa4220537f5648e13ace6693d1b9f5302c496fe874cb1813f67585b097e34f3"} Feb 27 17:20:10 crc kubenswrapper[4700]: I0227 17:20:10.317889 4700 generic.go:334] "Generic (PLEG): container finished" podID="23ac2574-09ac-4796-89b9-8581f4a73a00" containerID="949ac513e407dc81bdc02c7f5191885f58423e2b712e2b0e10597a037d0022af" exitCode=0 Feb 27 17:20:10 crc kubenswrapper[4700]: I0227 17:20:10.317923 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bb977cfbf-vkv4n" event={"ID":"23ac2574-09ac-4796-89b9-8581f4a73a00","Type":"ContainerDied","Data":"949ac513e407dc81bdc02c7f5191885f58423e2b712e2b0e10597a037d0022af"} Feb 27 17:20:10 crc kubenswrapper[4700]: I0227 17:20:10.317957 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bb977cfbf-vkv4n" event={"ID":"23ac2574-09ac-4796-89b9-8581f4a73a00","Type":"ContainerStarted","Data":"28c1761a698b7b74b43183d2f04d531b3d18ff3cd276ef569c86254b0049c2dc"} Feb 27 17:20:10 crc kubenswrapper[4700]: I0227 17:20:10.320742 4700 generic.go:334] "Generic (PLEG): container finished" podID="5a439f84-1aec-4aa7-bd1d-6e0683425b5b" containerID="00154ebd0614fd1ce33fae4283bedbed150ed30005fb10e797aa90fa9e2cea8d" exitCode=0 Feb 27 17:20:10 crc kubenswrapper[4700]: I0227 17:20:10.320797 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-857d474bb7-cd9zz" event={"ID":"5a439f84-1aec-4aa7-bd1d-6e0683425b5b","Type":"ContainerDied","Data":"00154ebd0614fd1ce33fae4283bedbed150ed30005fb10e797aa90fa9e2cea8d"} Feb 27 17:20:10 crc kubenswrapper[4700]: I0227 17:20:10.320844 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-857d474bb7-cd9zz" event={"ID":"5a439f84-1aec-4aa7-bd1d-6e0683425b5b","Type":"ContainerStarted","Data":"61dba040d54f2ed01a2ddb3d3964e00120ae55895d113b37610c302f1eac877e"} Feb 27 17:20:10 crc kubenswrapper[4700]: I0227 17:20:10.392237 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-7r49n" podStartSLOduration=2.392220253 podStartE2EDuration="2.392220253s" podCreationTimestamp="2026-02-27 17:20:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:20:10.388158466 +0000 UTC m=+1170.373471213" watchObservedRunningTime="2026-02-27 17:20:10.392220253 +0000 UTC m=+1170.377533000" Feb 27 17:20:10 crc kubenswrapper[4700]: I0227 17:20:10.593016 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-857d474bb7-cd9zz" Feb 27 17:20:10 crc kubenswrapper[4700]: I0227 17:20:10.645635 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Feb 27 17:20:10 crc kubenswrapper[4700]: I0227 17:20:10.716001 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrvhp\" (UniqueName: \"kubernetes.io/projected/5a439f84-1aec-4aa7-bd1d-6e0683425b5b-kube-api-access-hrvhp\") pod \"5a439f84-1aec-4aa7-bd1d-6e0683425b5b\" (UID: \"5a439f84-1aec-4aa7-bd1d-6e0683425b5b\") " Feb 27 17:20:10 crc kubenswrapper[4700]: I0227 17:20:10.716102 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a439f84-1aec-4aa7-bd1d-6e0683425b5b-dns-svc\") pod \"5a439f84-1aec-4aa7-bd1d-6e0683425b5b\" (UID: \"5a439f84-1aec-4aa7-bd1d-6e0683425b5b\") " Feb 27 17:20:10 crc kubenswrapper[4700]: I0227 17:20:10.716129 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a439f84-1aec-4aa7-bd1d-6e0683425b5b-config\") pod \"5a439f84-1aec-4aa7-bd1d-6e0683425b5b\" (UID: \"5a439f84-1aec-4aa7-bd1d-6e0683425b5b\") " Feb 27 17:20:10 crc kubenswrapper[4700]: I0227 17:20:10.716218 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5a439f84-1aec-4aa7-bd1d-6e0683425b5b-ovsdbserver-sb\") pod \"5a439f84-1aec-4aa7-bd1d-6e0683425b5b\" (UID: \"5a439f84-1aec-4aa7-bd1d-6e0683425b5b\") " Feb 27 17:20:10 crc kubenswrapper[4700]: I0227 17:20:10.720947 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a439f84-1aec-4aa7-bd1d-6e0683425b5b-kube-api-access-hrvhp" (OuterVolumeSpecName: "kube-api-access-hrvhp") pod "5a439f84-1aec-4aa7-bd1d-6e0683425b5b" (UID: "5a439f84-1aec-4aa7-bd1d-6e0683425b5b"). InnerVolumeSpecName "kube-api-access-hrvhp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:20:10 crc kubenswrapper[4700]: I0227 17:20:10.743982 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a439f84-1aec-4aa7-bd1d-6e0683425b5b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5a439f84-1aec-4aa7-bd1d-6e0683425b5b" (UID: "5a439f84-1aec-4aa7-bd1d-6e0683425b5b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:20:10 crc kubenswrapper[4700]: I0227 17:20:10.745722 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a439f84-1aec-4aa7-bd1d-6e0683425b5b-config" (OuterVolumeSpecName: "config") pod "5a439f84-1aec-4aa7-bd1d-6e0683425b5b" (UID: "5a439f84-1aec-4aa7-bd1d-6e0683425b5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:20:10 crc kubenswrapper[4700]: I0227 17:20:10.747801 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a439f84-1aec-4aa7-bd1d-6e0683425b5b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5a439f84-1aec-4aa7-bd1d-6e0683425b5b" (UID: "5a439f84-1aec-4aa7-bd1d-6e0683425b5b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:20:10 crc kubenswrapper[4700]: I0227 17:20:10.818410 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrvhp\" (UniqueName: \"kubernetes.io/projected/5a439f84-1aec-4aa7-bd1d-6e0683425b5b-kube-api-access-hrvhp\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:10 crc kubenswrapper[4700]: I0227 17:20:10.818741 4700 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a439f84-1aec-4aa7-bd1d-6e0683425b5b-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:10 crc kubenswrapper[4700]: I0227 17:20:10.818751 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a439f84-1aec-4aa7-bd1d-6e0683425b5b-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:10 crc kubenswrapper[4700]: I0227 17:20:10.818759 4700 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5a439f84-1aec-4aa7-bd1d-6e0683425b5b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:10 crc kubenswrapper[4700]: I0227 17:20:10.998193 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90" path="/var/lib/kubelet/pods/cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90/volumes" Feb 27 17:20:11 crc kubenswrapper[4700]: I0227 17:20:11.333187 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-857d474bb7-cd9zz" event={"ID":"5a439f84-1aec-4aa7-bd1d-6e0683425b5b","Type":"ContainerDied","Data":"61dba040d54f2ed01a2ddb3d3964e00120ae55895d113b37610c302f1eac877e"} Feb 27 17:20:11 crc kubenswrapper[4700]: I0227 17:20:11.333501 4700 scope.go:117] "RemoveContainer" containerID="00154ebd0614fd1ce33fae4283bedbed150ed30005fb10e797aa90fa9e2cea8d" Feb 27 17:20:11 crc kubenswrapper[4700]: I0227 17:20:11.333200 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-857d474bb7-cd9zz" Feb 27 17:20:11 crc kubenswrapper[4700]: I0227 17:20:11.339755 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bb977cfbf-vkv4n" event={"ID":"23ac2574-09ac-4796-89b9-8581f4a73a00","Type":"ContainerStarted","Data":"d16fdec76bd77fa9512b26c575ef9918040eb9229e456c3d2caec054ec4d8fdf"} Feb 27 17:20:11 crc kubenswrapper[4700]: I0227 17:20:11.373341 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-857d474bb7-cd9zz"] Feb 27 17:20:11 crc kubenswrapper[4700]: I0227 17:20:11.380505 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-857d474bb7-cd9zz"] Feb 27 17:20:11 crc kubenswrapper[4700]: I0227 17:20:11.404403 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5bb977cfbf-vkv4n" podStartSLOduration=3.404361556 podStartE2EDuration="3.404361556s" podCreationTimestamp="2026-02-27 17:20:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:20:11.401744456 +0000 UTC m=+1171.387057203" watchObservedRunningTime="2026-02-27 17:20:11.404361556 +0000 UTC m=+1171.389674303" Feb 27 17:20:12 crc kubenswrapper[4700]: I0227 17:20:12.354108 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5bb977cfbf-vkv4n" Feb 27 17:20:12 crc kubenswrapper[4700]: I0227 17:20:12.990971 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a439f84-1aec-4aa7-bd1d-6e0683425b5b" path="/var/lib/kubelet/pods/5a439f84-1aec-4aa7-bd1d-6e0683425b5b/volumes" Feb 27 17:20:13 crc kubenswrapper[4700]: I0227 17:20:13.376239 4700 generic.go:334] "Generic (PLEG): container finished" podID="f2325e29-9f4c-438f-9400-e643e2494808" containerID="9ff3a5239c9efca695db63b517f3295d9533dc87d43cf34a5d63ab1cd090d6e8" exitCode=0 Feb 27 17:20:13 crc kubenswrapper[4700]: I0227 17:20:13.376688 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f2325e29-9f4c-438f-9400-e643e2494808","Type":"ContainerDied","Data":"9ff3a5239c9efca695db63b517f3295d9533dc87d43cf34a5d63ab1cd090d6e8"} Feb 27 17:20:13 crc kubenswrapper[4700]: I0227 17:20:13.402028 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"458e6422-b856-4bcf-8308-4b6cb9ec8fef","Type":"ContainerStarted","Data":"3c345536df56b1f58f851e4bdb0af75eeb80bcf3825e3bbbaf62f032bc472793"} Feb 27 17:20:13 crc kubenswrapper[4700]: I0227 17:20:13.417682 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"efe7d637-da3a-4995-a40f-fae00257ac1d","Type":"ContainerStarted","Data":"eb50098b16c98521dac5370ccb4478720cceb5dc15fabc0e5682e5ccc113eb78"} Feb 27 17:20:13 crc kubenswrapper[4700]: I0227 17:20:13.420017 4700 generic.go:334] "Generic (PLEG): container finished" podID="9721431e-046b-4b3b-861e-22a65396981b" containerID="750401c839d4786183e640b5a5874d5afec2bc8dbd5e838d8a1d75f3a9e675a8" exitCode=0 Feb 27 17:20:13 crc kubenswrapper[4700]: I0227 17:20:13.420109 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9721431e-046b-4b3b-861e-22a65396981b","Type":"ContainerDied","Data":"750401c839d4786183e640b5a5874d5afec2bc8dbd5e838d8a1d75f3a9e675a8"} Feb 27 17:20:13 crc kubenswrapper[4700]: I0227 17:20:13.465779 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bb977cfbf-vkv4n"] Feb 27 17:20:13 crc kubenswrapper[4700]: I0227 17:20:13.521534 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6dc5d795ff-nbxlk"] Feb 27 17:20:13 crc kubenswrapper[4700]: E0227 17:20:13.521897 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90" containerName="init" Feb 27 17:20:13 crc kubenswrapper[4700]: I0227 17:20:13.521908 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90" containerName="init" Feb 27 17:20:13 crc kubenswrapper[4700]: E0227 17:20:13.521934 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a439f84-1aec-4aa7-bd1d-6e0683425b5b" containerName="init" Feb 27 17:20:13 crc kubenswrapper[4700]: I0227 17:20:13.521940 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a439f84-1aec-4aa7-bd1d-6e0683425b5b" containerName="init" Feb 27 17:20:13 crc kubenswrapper[4700]: E0227 17:20:13.521954 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90" containerName="dnsmasq-dns" Feb 27 17:20:13 crc kubenswrapper[4700]: I0227 17:20:13.521961 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90" containerName="dnsmasq-dns" Feb 27 17:20:13 crc kubenswrapper[4700]: I0227 17:20:13.522111 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a439f84-1aec-4aa7-bd1d-6e0683425b5b" containerName="init" Feb 27 17:20:13 crc kubenswrapper[4700]: I0227 17:20:13.522123 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc40f003-ee37-4e6c-a5ed-1b1ce2dfbf90" containerName="dnsmasq-dns" Feb 27 17:20:13 crc kubenswrapper[4700]: I0227 17:20:13.523197 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6dc5d795ff-nbxlk" Feb 27 17:20:13 crc kubenswrapper[4700]: I0227 17:20:13.555434 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6dc5d795ff-nbxlk"] Feb 27 17:20:13 crc kubenswrapper[4700]: I0227 17:20:13.685945 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da70ae3e-1bde-47f1-ab27-70ac8726f421-config\") pod \"dnsmasq-dns-6dc5d795ff-nbxlk\" (UID: \"da70ae3e-1bde-47f1-ab27-70ac8726f421\") " pod="openstack/dnsmasq-dns-6dc5d795ff-nbxlk" Feb 27 17:20:13 crc kubenswrapper[4700]: I0227 17:20:13.686320 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjq6m\" (UniqueName: \"kubernetes.io/projected/da70ae3e-1bde-47f1-ab27-70ac8726f421-kube-api-access-fjq6m\") pod \"dnsmasq-dns-6dc5d795ff-nbxlk\" (UID: \"da70ae3e-1bde-47f1-ab27-70ac8726f421\") " pod="openstack/dnsmasq-dns-6dc5d795ff-nbxlk" Feb 27 17:20:13 crc kubenswrapper[4700]: I0227 17:20:13.686410 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/da70ae3e-1bde-47f1-ab27-70ac8726f421-ovsdbserver-nb\") pod \"dnsmasq-dns-6dc5d795ff-nbxlk\" (UID: \"da70ae3e-1bde-47f1-ab27-70ac8726f421\") " pod="openstack/dnsmasq-dns-6dc5d795ff-nbxlk" Feb 27 17:20:13 crc kubenswrapper[4700]: I0227 17:20:13.686484 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da70ae3e-1bde-47f1-ab27-70ac8726f421-dns-svc\") pod \"dnsmasq-dns-6dc5d795ff-nbxlk\" (UID: \"da70ae3e-1bde-47f1-ab27-70ac8726f421\") " pod="openstack/dnsmasq-dns-6dc5d795ff-nbxlk" Feb 27 17:20:13 crc kubenswrapper[4700]: I0227 17:20:13.686522 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/da70ae3e-1bde-47f1-ab27-70ac8726f421-ovsdbserver-sb\") pod \"dnsmasq-dns-6dc5d795ff-nbxlk\" (UID: \"da70ae3e-1bde-47f1-ab27-70ac8726f421\") " pod="openstack/dnsmasq-dns-6dc5d795ff-nbxlk" Feb 27 17:20:13 crc kubenswrapper[4700]: I0227 17:20:13.788364 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da70ae3e-1bde-47f1-ab27-70ac8726f421-config\") pod \"dnsmasq-dns-6dc5d795ff-nbxlk\" (UID: \"da70ae3e-1bde-47f1-ab27-70ac8726f421\") " pod="openstack/dnsmasq-dns-6dc5d795ff-nbxlk" Feb 27 17:20:13 crc kubenswrapper[4700]: I0227 17:20:13.788425 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjq6m\" (UniqueName: \"kubernetes.io/projected/da70ae3e-1bde-47f1-ab27-70ac8726f421-kube-api-access-fjq6m\") pod \"dnsmasq-dns-6dc5d795ff-nbxlk\" (UID: \"da70ae3e-1bde-47f1-ab27-70ac8726f421\") " pod="openstack/dnsmasq-dns-6dc5d795ff-nbxlk" Feb 27 17:20:13 crc kubenswrapper[4700]: I0227 17:20:13.788491 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/da70ae3e-1bde-47f1-ab27-70ac8726f421-ovsdbserver-nb\") pod \"dnsmasq-dns-6dc5d795ff-nbxlk\" (UID: \"da70ae3e-1bde-47f1-ab27-70ac8726f421\") " pod="openstack/dnsmasq-dns-6dc5d795ff-nbxlk" Feb 27 17:20:13 crc kubenswrapper[4700]: I0227 17:20:13.788530 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da70ae3e-1bde-47f1-ab27-70ac8726f421-dns-svc\") pod \"dnsmasq-dns-6dc5d795ff-nbxlk\" (UID: \"da70ae3e-1bde-47f1-ab27-70ac8726f421\") " pod="openstack/dnsmasq-dns-6dc5d795ff-nbxlk" Feb 27 17:20:13 crc kubenswrapper[4700]: I0227 17:20:13.788546 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/da70ae3e-1bde-47f1-ab27-70ac8726f421-ovsdbserver-sb\") pod \"dnsmasq-dns-6dc5d795ff-nbxlk\" (UID: \"da70ae3e-1bde-47f1-ab27-70ac8726f421\") " pod="openstack/dnsmasq-dns-6dc5d795ff-nbxlk" Feb 27 17:20:13 crc kubenswrapper[4700]: I0227 17:20:13.789481 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/da70ae3e-1bde-47f1-ab27-70ac8726f421-ovsdbserver-sb\") pod \"dnsmasq-dns-6dc5d795ff-nbxlk\" (UID: \"da70ae3e-1bde-47f1-ab27-70ac8726f421\") " pod="openstack/dnsmasq-dns-6dc5d795ff-nbxlk" Feb 27 17:20:13 crc kubenswrapper[4700]: I0227 17:20:13.789650 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da70ae3e-1bde-47f1-ab27-70ac8726f421-dns-svc\") pod \"dnsmasq-dns-6dc5d795ff-nbxlk\" (UID: \"da70ae3e-1bde-47f1-ab27-70ac8726f421\") " pod="openstack/dnsmasq-dns-6dc5d795ff-nbxlk" Feb 27 17:20:13 crc kubenswrapper[4700]: I0227 17:20:13.790215 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da70ae3e-1bde-47f1-ab27-70ac8726f421-config\") pod \"dnsmasq-dns-6dc5d795ff-nbxlk\" (UID: \"da70ae3e-1bde-47f1-ab27-70ac8726f421\") " pod="openstack/dnsmasq-dns-6dc5d795ff-nbxlk" Feb 27 17:20:13 crc kubenswrapper[4700]: I0227 17:20:13.790301 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/da70ae3e-1bde-47f1-ab27-70ac8726f421-ovsdbserver-nb\") pod \"dnsmasq-dns-6dc5d795ff-nbxlk\" (UID: \"da70ae3e-1bde-47f1-ab27-70ac8726f421\") " pod="openstack/dnsmasq-dns-6dc5d795ff-nbxlk" Feb 27 17:20:13 crc kubenswrapper[4700]: I0227 17:20:13.808184 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjq6m\" (UniqueName: \"kubernetes.io/projected/da70ae3e-1bde-47f1-ab27-70ac8726f421-kube-api-access-fjq6m\") pod \"dnsmasq-dns-6dc5d795ff-nbxlk\" (UID: \"da70ae3e-1bde-47f1-ab27-70ac8726f421\") " pod="openstack/dnsmasq-dns-6dc5d795ff-nbxlk" Feb 27 17:20:13 crc kubenswrapper[4700]: I0227 17:20:13.864647 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6dc5d795ff-nbxlk" Feb 27 17:20:14 crc kubenswrapper[4700]: I0227 17:20:14.348834 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6dc5d795ff-nbxlk"] Feb 27 17:20:14 crc kubenswrapper[4700]: I0227 17:20:14.430047 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f2325e29-9f4c-438f-9400-e643e2494808","Type":"ContainerStarted","Data":"f8660853a9b459f4edea188e62a9bfe130d2d44f449e125d3c870f88b697b028"} Feb 27 17:20:14 crc kubenswrapper[4700]: I0227 17:20:14.433619 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5bb977cfbf-vkv4n" podUID="23ac2574-09ac-4796-89b9-8581f4a73a00" containerName="dnsmasq-dns" containerID="cri-o://d16fdec76bd77fa9512b26c575ef9918040eb9229e456c3d2caec054ec4d8fdf" gracePeriod=10 Feb 27 17:20:14 crc kubenswrapper[4700]: I0227 17:20:14.433904 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6dc5d795ff-nbxlk" event={"ID":"da70ae3e-1bde-47f1-ab27-70ac8726f421","Type":"ContainerStarted","Data":"503169e0dc123056d4a4f574b64a53114435a1eb573a85590a860a288b099253"} Feb 27 17:20:14 crc kubenswrapper[4700]: I0227 17:20:14.458790 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=31.740860151 podStartE2EDuration="47.458771202s" podCreationTimestamp="2026-02-27 17:19:27 +0000 UTC" firstStartedPulling="2026-02-27 17:19:43.93918875 +0000 UTC m=+1143.924501497" lastFinishedPulling="2026-02-27 17:19:59.657099791 +0000 UTC m=+1159.642412548" observedRunningTime="2026-02-27 17:20:14.453938024 +0000 UTC m=+1174.439250781" watchObservedRunningTime="2026-02-27 17:20:14.458771202 +0000 UTC m=+1174.444083949" Feb 27 17:20:14 crc kubenswrapper[4700]: I0227 17:20:14.611591 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Feb 27 17:20:14 crc kubenswrapper[4700]: I0227 17:20:14.618768 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 27 17:20:14 crc kubenswrapper[4700]: I0227 17:20:14.620676 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-cq7wr" Feb 27 17:20:14 crc kubenswrapper[4700]: I0227 17:20:14.631752 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Feb 27 17:20:14 crc kubenswrapper[4700]: I0227 17:20:14.631770 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Feb 27 17:20:14 crc kubenswrapper[4700]: I0227 17:20:14.631821 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Feb 27 17:20:14 crc kubenswrapper[4700]: I0227 17:20:14.675922 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Feb 27 17:20:14 crc kubenswrapper[4700]: I0227 17:20:14.704835 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/9df57c73-8d37-4f7d-b62f-209210dbba42-cache\") pod \"swift-storage-0\" (UID: \"9df57c73-8d37-4f7d-b62f-209210dbba42\") " pod="openstack/swift-storage-0" Feb 27 17:20:14 crc kubenswrapper[4700]: I0227 17:20:14.704919 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9df57c73-8d37-4f7d-b62f-209210dbba42-etc-swift\") pod \"swift-storage-0\" (UID: \"9df57c73-8d37-4f7d-b62f-209210dbba42\") " pod="openstack/swift-storage-0" Feb 27 17:20:14 crc kubenswrapper[4700]: I0227 17:20:14.705113 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/9df57c73-8d37-4f7d-b62f-209210dbba42-lock\") pod \"swift-storage-0\" (UID: \"9df57c73-8d37-4f7d-b62f-209210dbba42\") " pod="openstack/swift-storage-0" Feb 27 17:20:14 crc kubenswrapper[4700]: I0227 17:20:14.705160 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85rhv\" (UniqueName: \"kubernetes.io/projected/9df57c73-8d37-4f7d-b62f-209210dbba42-kube-api-access-85rhv\") pod \"swift-storage-0\" (UID: \"9df57c73-8d37-4f7d-b62f-209210dbba42\") " pod="openstack/swift-storage-0" Feb 27 17:20:14 crc kubenswrapper[4700]: I0227 17:20:14.705266 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"9df57c73-8d37-4f7d-b62f-209210dbba42\") " pod="openstack/swift-storage-0" Feb 27 17:20:14 crc kubenswrapper[4700]: I0227 17:20:14.705391 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9df57c73-8d37-4f7d-b62f-209210dbba42-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"9df57c73-8d37-4f7d-b62f-209210dbba42\") " pod="openstack/swift-storage-0" Feb 27 17:20:14 crc kubenswrapper[4700]: I0227 17:20:14.807091 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/9df57c73-8d37-4f7d-b62f-209210dbba42-lock\") pod \"swift-storage-0\" (UID: \"9df57c73-8d37-4f7d-b62f-209210dbba42\") " pod="openstack/swift-storage-0" Feb 27 17:20:14 crc kubenswrapper[4700]: I0227 17:20:14.807141 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85rhv\" (UniqueName: \"kubernetes.io/projected/9df57c73-8d37-4f7d-b62f-209210dbba42-kube-api-access-85rhv\") pod \"swift-storage-0\" (UID: \"9df57c73-8d37-4f7d-b62f-209210dbba42\") " pod="openstack/swift-storage-0" Feb 27 17:20:14 crc kubenswrapper[4700]: I0227 17:20:14.807198 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"9df57c73-8d37-4f7d-b62f-209210dbba42\") " pod="openstack/swift-storage-0" Feb 27 17:20:14 crc kubenswrapper[4700]: I0227 17:20:14.807242 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9df57c73-8d37-4f7d-b62f-209210dbba42-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"9df57c73-8d37-4f7d-b62f-209210dbba42\") " pod="openstack/swift-storage-0" Feb 27 17:20:14 crc kubenswrapper[4700]: I0227 17:20:14.807276 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/9df57c73-8d37-4f7d-b62f-209210dbba42-cache\") pod \"swift-storage-0\" (UID: \"9df57c73-8d37-4f7d-b62f-209210dbba42\") " pod="openstack/swift-storage-0" Feb 27 17:20:14 crc kubenswrapper[4700]: I0227 17:20:14.807324 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9df57c73-8d37-4f7d-b62f-209210dbba42-etc-swift\") pod \"swift-storage-0\" (UID: \"9df57c73-8d37-4f7d-b62f-209210dbba42\") " pod="openstack/swift-storage-0" Feb 27 17:20:14 crc kubenswrapper[4700]: E0227 17:20:14.807498 4700 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 27 17:20:14 crc kubenswrapper[4700]: E0227 17:20:14.807514 4700 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 27 17:20:14 crc kubenswrapper[4700]: E0227 17:20:14.807559 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9df57c73-8d37-4f7d-b62f-209210dbba42-etc-swift podName:9df57c73-8d37-4f7d-b62f-209210dbba42 nodeName:}" failed. No retries permitted until 2026-02-27 17:20:15.307543787 +0000 UTC m=+1175.292856534 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9df57c73-8d37-4f7d-b62f-209210dbba42-etc-swift") pod "swift-storage-0" (UID: "9df57c73-8d37-4f7d-b62f-209210dbba42") : configmap "swift-ring-files" not found Feb 27 17:20:14 crc kubenswrapper[4700]: I0227 17:20:14.807896 4700 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"9df57c73-8d37-4f7d-b62f-209210dbba42\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/swift-storage-0" Feb 27 17:20:14 crc kubenswrapper[4700]: I0227 17:20:14.808144 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/9df57c73-8d37-4f7d-b62f-209210dbba42-lock\") pod \"swift-storage-0\" (UID: \"9df57c73-8d37-4f7d-b62f-209210dbba42\") " pod="openstack/swift-storage-0" Feb 27 17:20:14 crc kubenswrapper[4700]: I0227 17:20:14.808564 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/9df57c73-8d37-4f7d-b62f-209210dbba42-cache\") pod \"swift-storage-0\" (UID: \"9df57c73-8d37-4f7d-b62f-209210dbba42\") " pod="openstack/swift-storage-0" Feb 27 17:20:14 crc kubenswrapper[4700]: I0227 17:20:14.811942 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9df57c73-8d37-4f7d-b62f-209210dbba42-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"9df57c73-8d37-4f7d-b62f-209210dbba42\") " pod="openstack/swift-storage-0" Feb 27 17:20:14 crc kubenswrapper[4700]: I0227 17:20:14.828541 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"9df57c73-8d37-4f7d-b62f-209210dbba42\") " pod="openstack/swift-storage-0" Feb 27 17:20:14 crc kubenswrapper[4700]: I0227 17:20:14.834396 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85rhv\" (UniqueName: \"kubernetes.io/projected/9df57c73-8d37-4f7d-b62f-209210dbba42-kube-api-access-85rhv\") pod \"swift-storage-0\" (UID: \"9df57c73-8d37-4f7d-b62f-209210dbba42\") " pod="openstack/swift-storage-0" Feb 27 17:20:15 crc kubenswrapper[4700]: I0227 17:20:15.315186 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9df57c73-8d37-4f7d-b62f-209210dbba42-etc-swift\") pod \"swift-storage-0\" (UID: \"9df57c73-8d37-4f7d-b62f-209210dbba42\") " pod="openstack/swift-storage-0" Feb 27 17:20:15 crc kubenswrapper[4700]: E0227 17:20:15.315392 4700 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 27 17:20:15 crc kubenswrapper[4700]: E0227 17:20:15.315430 4700 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 27 17:20:15 crc kubenswrapper[4700]: E0227 17:20:15.315527 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9df57c73-8d37-4f7d-b62f-209210dbba42-etc-swift podName:9df57c73-8d37-4f7d-b62f-209210dbba42 nodeName:}" failed. No retries permitted until 2026-02-27 17:20:16.315502793 +0000 UTC m=+1176.300815540 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9df57c73-8d37-4f7d-b62f-209210dbba42-etc-swift") pod "swift-storage-0" (UID: "9df57c73-8d37-4f7d-b62f-209210dbba42") : configmap "swift-ring-files" not found Feb 27 17:20:15 crc kubenswrapper[4700]: I0227 17:20:15.443047 4700 generic.go:334] "Generic (PLEG): container finished" podID="23ac2574-09ac-4796-89b9-8581f4a73a00" containerID="d16fdec76bd77fa9512b26c575ef9918040eb9229e456c3d2caec054ec4d8fdf" exitCode=0 Feb 27 17:20:15 crc kubenswrapper[4700]: I0227 17:20:15.443141 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bb977cfbf-vkv4n" event={"ID":"23ac2574-09ac-4796-89b9-8581f4a73a00","Type":"ContainerDied","Data":"d16fdec76bd77fa9512b26c575ef9918040eb9229e456c3d2caec054ec4d8fdf"} Feb 27 17:20:15 crc kubenswrapper[4700]: I0227 17:20:15.445447 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/notifications-rabbitmq-server-0" event={"ID":"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da","Type":"ContainerStarted","Data":"2f40fb47f39b8deb36a5e78ca5f657280f1b38d8420607c6ccf0836addcac851"} Feb 27 17:20:15 crc kubenswrapper[4700]: I0227 17:20:15.969151 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bb977cfbf-vkv4n" Feb 27 17:20:16 crc kubenswrapper[4700]: I0227 17:20:16.026951 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/23ac2574-09ac-4796-89b9-8581f4a73a00-ovsdbserver-sb\") pod \"23ac2574-09ac-4796-89b9-8581f4a73a00\" (UID: \"23ac2574-09ac-4796-89b9-8581f4a73a00\") " Feb 27 17:20:16 crc kubenswrapper[4700]: I0227 17:20:16.027025 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23ac2574-09ac-4796-89b9-8581f4a73a00-config\") pod \"23ac2574-09ac-4796-89b9-8581f4a73a00\" (UID: \"23ac2574-09ac-4796-89b9-8581f4a73a00\") " Feb 27 17:20:16 crc kubenswrapper[4700]: I0227 17:20:16.027096 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/23ac2574-09ac-4796-89b9-8581f4a73a00-dns-svc\") pod \"23ac2574-09ac-4796-89b9-8581f4a73a00\" (UID: \"23ac2574-09ac-4796-89b9-8581f4a73a00\") " Feb 27 17:20:16 crc kubenswrapper[4700]: I0227 17:20:16.027208 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/23ac2574-09ac-4796-89b9-8581f4a73a00-ovsdbserver-nb\") pod \"23ac2574-09ac-4796-89b9-8581f4a73a00\" (UID: \"23ac2574-09ac-4796-89b9-8581f4a73a00\") " Feb 27 17:20:16 crc kubenswrapper[4700]: I0227 17:20:16.027256 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7fcs\" (UniqueName: \"kubernetes.io/projected/23ac2574-09ac-4796-89b9-8581f4a73a00-kube-api-access-q7fcs\") pod \"23ac2574-09ac-4796-89b9-8581f4a73a00\" (UID: \"23ac2574-09ac-4796-89b9-8581f4a73a00\") " Feb 27 17:20:16 crc kubenswrapper[4700]: I0227 17:20:16.033365 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23ac2574-09ac-4796-89b9-8581f4a73a00-kube-api-access-q7fcs" (OuterVolumeSpecName: "kube-api-access-q7fcs") pod "23ac2574-09ac-4796-89b9-8581f4a73a00" (UID: "23ac2574-09ac-4796-89b9-8581f4a73a00"). InnerVolumeSpecName "kube-api-access-q7fcs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:20:16 crc kubenswrapper[4700]: I0227 17:20:16.075223 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23ac2574-09ac-4796-89b9-8581f4a73a00-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "23ac2574-09ac-4796-89b9-8581f4a73a00" (UID: "23ac2574-09ac-4796-89b9-8581f4a73a00"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:20:16 crc kubenswrapper[4700]: I0227 17:20:16.079215 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23ac2574-09ac-4796-89b9-8581f4a73a00-config" (OuterVolumeSpecName: "config") pod "23ac2574-09ac-4796-89b9-8581f4a73a00" (UID: "23ac2574-09ac-4796-89b9-8581f4a73a00"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:20:16 crc kubenswrapper[4700]: I0227 17:20:16.080469 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23ac2574-09ac-4796-89b9-8581f4a73a00-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "23ac2574-09ac-4796-89b9-8581f4a73a00" (UID: "23ac2574-09ac-4796-89b9-8581f4a73a00"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:20:16 crc kubenswrapper[4700]: I0227 17:20:16.088774 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23ac2574-09ac-4796-89b9-8581f4a73a00-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "23ac2574-09ac-4796-89b9-8581f4a73a00" (UID: "23ac2574-09ac-4796-89b9-8581f4a73a00"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:20:16 crc kubenswrapper[4700]: I0227 17:20:16.132510 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7fcs\" (UniqueName: \"kubernetes.io/projected/23ac2574-09ac-4796-89b9-8581f4a73a00-kube-api-access-q7fcs\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:16 crc kubenswrapper[4700]: I0227 17:20:16.132543 4700 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/23ac2574-09ac-4796-89b9-8581f4a73a00-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:16 crc kubenswrapper[4700]: I0227 17:20:16.132556 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23ac2574-09ac-4796-89b9-8581f4a73a00-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:16 crc kubenswrapper[4700]: I0227 17:20:16.132566 4700 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/23ac2574-09ac-4796-89b9-8581f4a73a00-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:16 crc kubenswrapper[4700]: I0227 17:20:16.132576 4700 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/23ac2574-09ac-4796-89b9-8581f4a73a00-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:16 crc kubenswrapper[4700]: I0227 17:20:16.335287 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9df57c73-8d37-4f7d-b62f-209210dbba42-etc-swift\") pod \"swift-storage-0\" (UID: \"9df57c73-8d37-4f7d-b62f-209210dbba42\") " pod="openstack/swift-storage-0" Feb 27 17:20:16 crc kubenswrapper[4700]: E0227 17:20:16.335895 4700 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 27 17:20:16 crc kubenswrapper[4700]: E0227 17:20:16.336025 4700 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 27 17:20:16 crc kubenswrapper[4700]: E0227 17:20:16.336151 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9df57c73-8d37-4f7d-b62f-209210dbba42-etc-swift podName:9df57c73-8d37-4f7d-b62f-209210dbba42 nodeName:}" failed. No retries permitted until 2026-02-27 17:20:18.336131541 +0000 UTC m=+1178.321444288 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9df57c73-8d37-4f7d-b62f-209210dbba42-etc-swift") pod "swift-storage-0" (UID: "9df57c73-8d37-4f7d-b62f-209210dbba42") : configmap "swift-ring-files" not found Feb 27 17:20:16 crc kubenswrapper[4700]: I0227 17:20:16.455165 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bb977cfbf-vkv4n" event={"ID":"23ac2574-09ac-4796-89b9-8581f4a73a00","Type":"ContainerDied","Data":"28c1761a698b7b74b43183d2f04d531b3d18ff3cd276ef569c86254b0049c2dc"} Feb 27 17:20:16 crc kubenswrapper[4700]: I0227 17:20:16.455214 4700 scope.go:117] "RemoveContainer" containerID="d16fdec76bd77fa9512b26c575ef9918040eb9229e456c3d2caec054ec4d8fdf" Feb 27 17:20:16 crc kubenswrapper[4700]: I0227 17:20:16.455183 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bb977cfbf-vkv4n" Feb 27 17:20:16 crc kubenswrapper[4700]: I0227 17:20:16.456644 4700 generic.go:334] "Generic (PLEG): container finished" podID="da70ae3e-1bde-47f1-ab27-70ac8726f421" containerID="e638eee0c53829a6e26465fe8359d3f713d701fa523d9c136f61d6c0fd0d2444" exitCode=0 Feb 27 17:20:16 crc kubenswrapper[4700]: I0227 17:20:16.456870 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6dc5d795ff-nbxlk" event={"ID":"da70ae3e-1bde-47f1-ab27-70ac8726f421","Type":"ContainerDied","Data":"e638eee0c53829a6e26465fe8359d3f713d701fa523d9c136f61d6c0fd0d2444"} Feb 27 17:20:16 crc kubenswrapper[4700]: I0227 17:20:16.481470 4700 scope.go:117] "RemoveContainer" containerID="949ac513e407dc81bdc02c7f5191885f58423e2b712e2b0e10597a037d0022af" Feb 27 17:20:16 crc kubenswrapper[4700]: I0227 17:20:16.544167 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bb977cfbf-vkv4n"] Feb 27 17:20:16 crc kubenswrapper[4700]: I0227 17:20:16.552865 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bb977cfbf-vkv4n"] Feb 27 17:20:16 crc kubenswrapper[4700]: I0227 17:20:16.998078 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23ac2574-09ac-4796-89b9-8581f4a73a00" path="/var/lib/kubelet/pods/23ac2574-09ac-4796-89b9-8581f4a73a00/volumes" Feb 27 17:20:17 crc kubenswrapper[4700]: I0227 17:20:17.467050 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6dc5d795ff-nbxlk" event={"ID":"da70ae3e-1bde-47f1-ab27-70ac8726f421","Type":"ContainerStarted","Data":"6fb9866278becdccf3970d53fb04f5f1456c5ecd00d4a819521b7670a61aecd3"} Feb 27 17:20:17 crc kubenswrapper[4700]: I0227 17:20:17.467280 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6dc5d795ff-nbxlk" Feb 27 17:20:17 crc kubenswrapper[4700]: I0227 17:20:17.492104 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6dc5d795ff-nbxlk" podStartSLOduration=4.492084739 podStartE2EDuration="4.492084739s" podCreationTimestamp="2026-02-27 17:20:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:20:17.48833914 +0000 UTC m=+1177.473651897" watchObservedRunningTime="2026-02-27 17:20:17.492084739 +0000 UTC m=+1177.477397496" Feb 27 17:20:18 crc kubenswrapper[4700]: I0227 17:20:18.379934 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9df57c73-8d37-4f7d-b62f-209210dbba42-etc-swift\") pod \"swift-storage-0\" (UID: \"9df57c73-8d37-4f7d-b62f-209210dbba42\") " pod="openstack/swift-storage-0" Feb 27 17:20:18 crc kubenswrapper[4700]: E0227 17:20:18.380925 4700 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 27 17:20:18 crc kubenswrapper[4700]: E0227 17:20:18.380976 4700 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 27 17:20:18 crc kubenswrapper[4700]: E0227 17:20:18.381064 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9df57c73-8d37-4f7d-b62f-209210dbba42-etc-swift podName:9df57c73-8d37-4f7d-b62f-209210dbba42 nodeName:}" failed. No retries permitted until 2026-02-27 17:20:22.381037904 +0000 UTC m=+1182.366350691 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9df57c73-8d37-4f7d-b62f-209210dbba42-etc-swift") pod "swift-storage-0" (UID: "9df57c73-8d37-4f7d-b62f-209210dbba42") : configmap "swift-ring-files" not found Feb 27 17:20:18 crc kubenswrapper[4700]: I0227 17:20:18.491342 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d204fe1c-3288-4837-992f-a04ddea6dbd3","Type":"ContainerStarted","Data":"faca510343c0aafdfbd58f2b0053ee3d479c24c17c45411fbcb42ad20c39d844"} Feb 27 17:20:18 crc kubenswrapper[4700]: I0227 17:20:18.491916 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 27 17:20:18 crc kubenswrapper[4700]: I0227 17:20:18.499762 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-m7qsc"] Feb 27 17:20:18 crc kubenswrapper[4700]: E0227 17:20:18.500267 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23ac2574-09ac-4796-89b9-8581f4a73a00" containerName="dnsmasq-dns" Feb 27 17:20:18 crc kubenswrapper[4700]: I0227 17:20:18.500296 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="23ac2574-09ac-4796-89b9-8581f4a73a00" containerName="dnsmasq-dns" Feb 27 17:20:18 crc kubenswrapper[4700]: E0227 17:20:18.500341 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23ac2574-09ac-4796-89b9-8581f4a73a00" containerName="init" Feb 27 17:20:18 crc kubenswrapper[4700]: I0227 17:20:18.500357 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="23ac2574-09ac-4796-89b9-8581f4a73a00" containerName="init" Feb 27 17:20:18 crc kubenswrapper[4700]: I0227 17:20:18.500767 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="23ac2574-09ac-4796-89b9-8581f4a73a00" containerName="dnsmasq-dns" Feb 27 17:20:18 crc kubenswrapper[4700]: I0227 17:20:18.501688 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-m7qsc" Feb 27 17:20:18 crc kubenswrapper[4700]: I0227 17:20:18.505604 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Feb 27 17:20:18 crc kubenswrapper[4700]: I0227 17:20:18.505944 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Feb 27 17:20:18 crc kubenswrapper[4700]: I0227 17:20:18.506170 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Feb 27 17:20:18 crc kubenswrapper[4700]: I0227 17:20:18.521548 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=13.453785661 podStartE2EDuration="46.521530309s" podCreationTimestamp="2026-02-27 17:19:32 +0000 UTC" firstStartedPulling="2026-02-27 17:19:44.296343795 +0000 UTC m=+1144.281656552" lastFinishedPulling="2026-02-27 17:20:17.364088453 +0000 UTC m=+1177.349401200" observedRunningTime="2026-02-27 17:20:18.513456476 +0000 UTC m=+1178.498769263" watchObservedRunningTime="2026-02-27 17:20:18.521530309 +0000 UTC m=+1178.506843066" Feb 27 17:20:18 crc kubenswrapper[4700]: I0227 17:20:18.557449 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-m7qsc"] Feb 27 17:20:18 crc kubenswrapper[4700]: I0227 17:20:18.584698 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/89befb8e-a997-46a9-9db2-24f0dc5bf10d-etc-swift\") pod \"swift-ring-rebalance-m7qsc\" (UID: \"89befb8e-a997-46a9-9db2-24f0dc5bf10d\") " pod="openstack/swift-ring-rebalance-m7qsc" Feb 27 17:20:18 crc kubenswrapper[4700]: I0227 17:20:18.584734 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89befb8e-a997-46a9-9db2-24f0dc5bf10d-combined-ca-bundle\") pod \"swift-ring-rebalance-m7qsc\" (UID: \"89befb8e-a997-46a9-9db2-24f0dc5bf10d\") " pod="openstack/swift-ring-rebalance-m7qsc" Feb 27 17:20:18 crc kubenswrapper[4700]: I0227 17:20:18.584755 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89befb8e-a997-46a9-9db2-24f0dc5bf10d-scripts\") pod \"swift-ring-rebalance-m7qsc\" (UID: \"89befb8e-a997-46a9-9db2-24f0dc5bf10d\") " pod="openstack/swift-ring-rebalance-m7qsc" Feb 27 17:20:18 crc kubenswrapper[4700]: I0227 17:20:18.584773 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/89befb8e-a997-46a9-9db2-24f0dc5bf10d-dispersionconf\") pod \"swift-ring-rebalance-m7qsc\" (UID: \"89befb8e-a997-46a9-9db2-24f0dc5bf10d\") " pod="openstack/swift-ring-rebalance-m7qsc" Feb 27 17:20:18 crc kubenswrapper[4700]: I0227 17:20:18.584805 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npln6\" (UniqueName: \"kubernetes.io/projected/89befb8e-a997-46a9-9db2-24f0dc5bf10d-kube-api-access-npln6\") pod \"swift-ring-rebalance-m7qsc\" (UID: \"89befb8e-a997-46a9-9db2-24f0dc5bf10d\") " pod="openstack/swift-ring-rebalance-m7qsc" Feb 27 17:20:18 crc kubenswrapper[4700]: I0227 17:20:18.584887 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/89befb8e-a997-46a9-9db2-24f0dc5bf10d-ring-data-devices\") pod \"swift-ring-rebalance-m7qsc\" (UID: \"89befb8e-a997-46a9-9db2-24f0dc5bf10d\") " pod="openstack/swift-ring-rebalance-m7qsc" Feb 27 17:20:18 crc kubenswrapper[4700]: I0227 17:20:18.584949 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/89befb8e-a997-46a9-9db2-24f0dc5bf10d-swiftconf\") pod \"swift-ring-rebalance-m7qsc\" (UID: \"89befb8e-a997-46a9-9db2-24f0dc5bf10d\") " pod="openstack/swift-ring-rebalance-m7qsc" Feb 27 17:20:18 crc kubenswrapper[4700]: I0227 17:20:18.686175 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/89befb8e-a997-46a9-9db2-24f0dc5bf10d-ring-data-devices\") pod \"swift-ring-rebalance-m7qsc\" (UID: \"89befb8e-a997-46a9-9db2-24f0dc5bf10d\") " pod="openstack/swift-ring-rebalance-m7qsc" Feb 27 17:20:18 crc kubenswrapper[4700]: I0227 17:20:18.686254 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/89befb8e-a997-46a9-9db2-24f0dc5bf10d-swiftconf\") pod \"swift-ring-rebalance-m7qsc\" (UID: \"89befb8e-a997-46a9-9db2-24f0dc5bf10d\") " pod="openstack/swift-ring-rebalance-m7qsc" Feb 27 17:20:18 crc kubenswrapper[4700]: I0227 17:20:18.686304 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/89befb8e-a997-46a9-9db2-24f0dc5bf10d-etc-swift\") pod \"swift-ring-rebalance-m7qsc\" (UID: \"89befb8e-a997-46a9-9db2-24f0dc5bf10d\") " pod="openstack/swift-ring-rebalance-m7qsc" Feb 27 17:20:18 crc kubenswrapper[4700]: I0227 17:20:18.686323 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89befb8e-a997-46a9-9db2-24f0dc5bf10d-combined-ca-bundle\") pod \"swift-ring-rebalance-m7qsc\" (UID: \"89befb8e-a997-46a9-9db2-24f0dc5bf10d\") " pod="openstack/swift-ring-rebalance-m7qsc" Feb 27 17:20:18 crc kubenswrapper[4700]: I0227 17:20:18.686339 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89befb8e-a997-46a9-9db2-24f0dc5bf10d-scripts\") pod \"swift-ring-rebalance-m7qsc\" (UID: \"89befb8e-a997-46a9-9db2-24f0dc5bf10d\") " pod="openstack/swift-ring-rebalance-m7qsc" Feb 27 17:20:18 crc kubenswrapper[4700]: I0227 17:20:18.686355 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/89befb8e-a997-46a9-9db2-24f0dc5bf10d-dispersionconf\") pod \"swift-ring-rebalance-m7qsc\" (UID: \"89befb8e-a997-46a9-9db2-24f0dc5bf10d\") " pod="openstack/swift-ring-rebalance-m7qsc" Feb 27 17:20:18 crc kubenswrapper[4700]: I0227 17:20:18.686379 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npln6\" (UniqueName: \"kubernetes.io/projected/89befb8e-a997-46a9-9db2-24f0dc5bf10d-kube-api-access-npln6\") pod \"swift-ring-rebalance-m7qsc\" (UID: \"89befb8e-a997-46a9-9db2-24f0dc5bf10d\") " pod="openstack/swift-ring-rebalance-m7qsc" Feb 27 17:20:18 crc kubenswrapper[4700]: I0227 17:20:18.687161 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/89befb8e-a997-46a9-9db2-24f0dc5bf10d-etc-swift\") pod \"swift-ring-rebalance-m7qsc\" (UID: \"89befb8e-a997-46a9-9db2-24f0dc5bf10d\") " pod="openstack/swift-ring-rebalance-m7qsc" Feb 27 17:20:18 crc kubenswrapper[4700]: I0227 17:20:18.687231 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/89befb8e-a997-46a9-9db2-24f0dc5bf10d-ring-data-devices\") pod \"swift-ring-rebalance-m7qsc\" (UID: \"89befb8e-a997-46a9-9db2-24f0dc5bf10d\") " pod="openstack/swift-ring-rebalance-m7qsc" Feb 27 17:20:18 crc kubenswrapper[4700]: I0227 17:20:18.687699 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89befb8e-a997-46a9-9db2-24f0dc5bf10d-scripts\") pod \"swift-ring-rebalance-m7qsc\" (UID: \"89befb8e-a997-46a9-9db2-24f0dc5bf10d\") " pod="openstack/swift-ring-rebalance-m7qsc" Feb 27 17:20:18 crc kubenswrapper[4700]: I0227 17:20:18.691549 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/89befb8e-a997-46a9-9db2-24f0dc5bf10d-dispersionconf\") pod \"swift-ring-rebalance-m7qsc\" (UID: \"89befb8e-a997-46a9-9db2-24f0dc5bf10d\") " pod="openstack/swift-ring-rebalance-m7qsc" Feb 27 17:20:18 crc kubenswrapper[4700]: I0227 17:20:18.692090 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/89befb8e-a997-46a9-9db2-24f0dc5bf10d-swiftconf\") pod \"swift-ring-rebalance-m7qsc\" (UID: \"89befb8e-a997-46a9-9db2-24f0dc5bf10d\") " pod="openstack/swift-ring-rebalance-m7qsc" Feb 27 17:20:18 crc kubenswrapper[4700]: I0227 17:20:18.693310 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89befb8e-a997-46a9-9db2-24f0dc5bf10d-combined-ca-bundle\") pod \"swift-ring-rebalance-m7qsc\" (UID: \"89befb8e-a997-46a9-9db2-24f0dc5bf10d\") " pod="openstack/swift-ring-rebalance-m7qsc" Feb 27 17:20:18 crc kubenswrapper[4700]: I0227 17:20:18.704962 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npln6\" (UniqueName: \"kubernetes.io/projected/89befb8e-a997-46a9-9db2-24f0dc5bf10d-kube-api-access-npln6\") pod \"swift-ring-rebalance-m7qsc\" (UID: \"89befb8e-a997-46a9-9db2-24f0dc5bf10d\") " pod="openstack/swift-ring-rebalance-m7qsc" Feb 27 17:20:18 crc kubenswrapper[4700]: I0227 17:20:18.797454 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Feb 27 17:20:18 crc kubenswrapper[4700]: I0227 17:20:18.797882 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Feb 27 17:20:18 crc kubenswrapper[4700]: I0227 17:20:18.833342 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-m7qsc" Feb 27 17:20:19 crc kubenswrapper[4700]: I0227 17:20:19.369264 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-m7qsc"] Feb 27 17:20:19 crc kubenswrapper[4700]: I0227 17:20:19.504191 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a4b9e543-db0e-4b6a-bd32-9935f603556e","Type":"ContainerStarted","Data":"0952ed3959c98c001227b724897efb41d45a44903ee797902c3659e1031f67cb"} Feb 27 17:20:19 crc kubenswrapper[4700]: I0227 17:20:19.505789 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-m7qsc" event={"ID":"89befb8e-a997-46a9-9db2-24f0dc5bf10d","Type":"ContainerStarted","Data":"e201d4853198997660f8cf9b9866a504737185cd637f0fc081d2984aa5794595"} Feb 27 17:20:19 crc kubenswrapper[4700]: I0227 17:20:19.526983 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=9.726596767 podStartE2EDuration="44.526966926s" podCreationTimestamp="2026-02-27 17:19:35 +0000 UTC" firstStartedPulling="2026-02-27 17:19:44.311285901 +0000 UTC m=+1144.296598648" lastFinishedPulling="2026-02-27 17:20:19.11165606 +0000 UTC m=+1179.096968807" observedRunningTime="2026-02-27 17:20:19.522057146 +0000 UTC m=+1179.507369893" watchObservedRunningTime="2026-02-27 17:20:19.526966926 +0000 UTC m=+1179.512279673" Feb 27 17:20:20 crc kubenswrapper[4700]: I0227 17:20:20.519007 4700 generic.go:334] "Generic (PLEG): container finished" podID="4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223" containerID="b154a054c6a45d944ce284b5f64f55a21d5f1658815d5dcbff5c63fee94cc5c1" exitCode=0 Feb 27 17:20:20 crc kubenswrapper[4700]: I0227 17:20:20.519058 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223","Type":"ContainerDied","Data":"b154a054c6a45d944ce284b5f64f55a21d5f1658815d5dcbff5c63fee94cc5c1"} Feb 27 17:20:21 crc kubenswrapper[4700]: I0227 17:20:21.591328 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Feb 27 17:20:21 crc kubenswrapper[4700]: I0227 17:20:21.591773 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Feb 27 17:20:22 crc kubenswrapper[4700]: I0227 17:20:22.457028 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9df57c73-8d37-4f7d-b62f-209210dbba42-etc-swift\") pod \"swift-storage-0\" (UID: \"9df57c73-8d37-4f7d-b62f-209210dbba42\") " pod="openstack/swift-storage-0" Feb 27 17:20:22 crc kubenswrapper[4700]: E0227 17:20:22.457709 4700 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 27 17:20:22 crc kubenswrapper[4700]: E0227 17:20:22.457728 4700 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 27 17:20:22 crc kubenswrapper[4700]: E0227 17:20:22.457780 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9df57c73-8d37-4f7d-b62f-209210dbba42-etc-swift podName:9df57c73-8d37-4f7d-b62f-209210dbba42 nodeName:}" failed. No retries permitted until 2026-02-27 17:20:30.457762371 +0000 UTC m=+1190.443075138 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9df57c73-8d37-4f7d-b62f-209210dbba42-etc-swift") pod "swift-storage-0" (UID: "9df57c73-8d37-4f7d-b62f-209210dbba42") : configmap "swift-ring-files" not found Feb 27 17:20:22 crc kubenswrapper[4700]: I0227 17:20:22.490437 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Feb 27 17:20:22 crc kubenswrapper[4700]: I0227 17:20:22.617845 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Feb 27 17:20:23 crc kubenswrapper[4700]: I0227 17:20:23.193441 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 27 17:20:23 crc kubenswrapper[4700]: I0227 17:20:23.453566 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-db-create-gx4c9"] Feb 27 17:20:23 crc kubenswrapper[4700]: I0227 17:20:23.454635 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-gx4c9" Feb 27 17:20:23 crc kubenswrapper[4700]: I0227 17:20:23.462607 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-0076-account-create-update-hwxxq"] Feb 27 17:20:23 crc kubenswrapper[4700]: I0227 17:20:23.463663 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-0076-account-create-update-hwxxq" Feb 27 17:20:23 crc kubenswrapper[4700]: I0227 17:20:23.469641 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-create-gx4c9"] Feb 27 17:20:23 crc kubenswrapper[4700]: I0227 17:20:23.489788 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-db-secret" Feb 27 17:20:23 crc kubenswrapper[4700]: I0227 17:20:23.497665 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-0076-account-create-update-hwxxq"] Feb 27 17:20:23 crc kubenswrapper[4700]: I0227 17:20:23.590134 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnp75\" (UniqueName: \"kubernetes.io/projected/18a5e20e-5c10-4966-9adc-c10670ddd4e3-kube-api-access-mnp75\") pod \"watcher-0076-account-create-update-hwxxq\" (UID: \"18a5e20e-5c10-4966-9adc-c10670ddd4e3\") " pod="openstack/watcher-0076-account-create-update-hwxxq" Feb 27 17:20:23 crc kubenswrapper[4700]: I0227 17:20:23.590298 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d0612657-780a-4388-9b79-ae606ed152e3-operator-scripts\") pod \"watcher-db-create-gx4c9\" (UID: \"d0612657-780a-4388-9b79-ae606ed152e3\") " pod="openstack/watcher-db-create-gx4c9" Feb 27 17:20:23 crc kubenswrapper[4700]: I0227 17:20:23.590422 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vr7fj\" (UniqueName: \"kubernetes.io/projected/d0612657-780a-4388-9b79-ae606ed152e3-kube-api-access-vr7fj\") pod \"watcher-db-create-gx4c9\" (UID: \"d0612657-780a-4388-9b79-ae606ed152e3\") " pod="openstack/watcher-db-create-gx4c9" Feb 27 17:20:23 crc kubenswrapper[4700]: I0227 17:20:23.590561 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18a5e20e-5c10-4966-9adc-c10670ddd4e3-operator-scripts\") pod \"watcher-0076-account-create-update-hwxxq\" (UID: \"18a5e20e-5c10-4966-9adc-c10670ddd4e3\") " pod="openstack/watcher-0076-account-create-update-hwxxq" Feb 27 17:20:23 crc kubenswrapper[4700]: I0227 17:20:23.692616 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vr7fj\" (UniqueName: \"kubernetes.io/projected/d0612657-780a-4388-9b79-ae606ed152e3-kube-api-access-vr7fj\") pod \"watcher-db-create-gx4c9\" (UID: \"d0612657-780a-4388-9b79-ae606ed152e3\") " pod="openstack/watcher-db-create-gx4c9" Feb 27 17:20:23 crc kubenswrapper[4700]: I0227 17:20:23.692723 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18a5e20e-5c10-4966-9adc-c10670ddd4e3-operator-scripts\") pod \"watcher-0076-account-create-update-hwxxq\" (UID: \"18a5e20e-5c10-4966-9adc-c10670ddd4e3\") " pod="openstack/watcher-0076-account-create-update-hwxxq" Feb 27 17:20:23 crc kubenswrapper[4700]: I0227 17:20:23.692799 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnp75\" (UniqueName: \"kubernetes.io/projected/18a5e20e-5c10-4966-9adc-c10670ddd4e3-kube-api-access-mnp75\") pod \"watcher-0076-account-create-update-hwxxq\" (UID: \"18a5e20e-5c10-4966-9adc-c10670ddd4e3\") " pod="openstack/watcher-0076-account-create-update-hwxxq" Feb 27 17:20:23 crc kubenswrapper[4700]: I0227 17:20:23.692874 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d0612657-780a-4388-9b79-ae606ed152e3-operator-scripts\") pod \"watcher-db-create-gx4c9\" (UID: \"d0612657-780a-4388-9b79-ae606ed152e3\") " pod="openstack/watcher-db-create-gx4c9" Feb 27 17:20:23 crc kubenswrapper[4700]: I0227 17:20:23.693826 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d0612657-780a-4388-9b79-ae606ed152e3-operator-scripts\") pod \"watcher-db-create-gx4c9\" (UID: \"d0612657-780a-4388-9b79-ae606ed152e3\") " pod="openstack/watcher-db-create-gx4c9" Feb 27 17:20:23 crc kubenswrapper[4700]: I0227 17:20:23.694846 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18a5e20e-5c10-4966-9adc-c10670ddd4e3-operator-scripts\") pod \"watcher-0076-account-create-update-hwxxq\" (UID: \"18a5e20e-5c10-4966-9adc-c10670ddd4e3\") " pod="openstack/watcher-0076-account-create-update-hwxxq" Feb 27 17:20:23 crc kubenswrapper[4700]: I0227 17:20:23.711038 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vr7fj\" (UniqueName: \"kubernetes.io/projected/d0612657-780a-4388-9b79-ae606ed152e3-kube-api-access-vr7fj\") pod \"watcher-db-create-gx4c9\" (UID: \"d0612657-780a-4388-9b79-ae606ed152e3\") " pod="openstack/watcher-db-create-gx4c9" Feb 27 17:20:23 crc kubenswrapper[4700]: I0227 17:20:23.728734 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnp75\" (UniqueName: \"kubernetes.io/projected/18a5e20e-5c10-4966-9adc-c10670ddd4e3-kube-api-access-mnp75\") pod \"watcher-0076-account-create-update-hwxxq\" (UID: \"18a5e20e-5c10-4966-9adc-c10670ddd4e3\") " pod="openstack/watcher-0076-account-create-update-hwxxq" Feb 27 17:20:23 crc kubenswrapper[4700]: I0227 17:20:23.817369 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-gx4c9" Feb 27 17:20:23 crc kubenswrapper[4700]: I0227 17:20:23.831494 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-0076-account-create-update-hwxxq" Feb 27 17:20:23 crc kubenswrapper[4700]: I0227 17:20:23.866537 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6dc5d795ff-nbxlk" Feb 27 17:20:23 crc kubenswrapper[4700]: I0227 17:20:23.928208 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bbc67fc-tr49q"] Feb 27 17:20:23 crc kubenswrapper[4700]: I0227 17:20:23.928479 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7bbc67fc-tr49q" podUID="4448288d-21f1-418b-95bf-9287e914f52b" containerName="dnsmasq-dns" containerID="cri-o://36a7354b8a1d35885156629f2499a02f10cf0989a40e38c49fba3fae4670d9aa" gracePeriod=10 Feb 27 17:20:24 crc kubenswrapper[4700]: I0227 17:20:24.571023 4700 generic.go:334] "Generic (PLEG): container finished" podID="4448288d-21f1-418b-95bf-9287e914f52b" containerID="36a7354b8a1d35885156629f2499a02f10cf0989a40e38c49fba3fae4670d9aa" exitCode=0 Feb 27 17:20:24 crc kubenswrapper[4700]: I0227 17:20:24.571072 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bbc67fc-tr49q" event={"ID":"4448288d-21f1-418b-95bf-9287e914f52b","Type":"ContainerDied","Data":"36a7354b8a1d35885156629f2499a02f10cf0989a40e38c49fba3fae4670d9aa"} Feb 27 17:20:24 crc kubenswrapper[4700]: I0227 17:20:24.666588 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Feb 27 17:20:24 crc kubenswrapper[4700]: I0227 17:20:24.732310 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Feb 27 17:20:24 crc kubenswrapper[4700]: I0227 17:20:24.899797 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Feb 27 17:20:24 crc kubenswrapper[4700]: I0227 17:20:24.901897 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 27 17:20:24 crc kubenswrapper[4700]: I0227 17:20:24.905583 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Feb 27 17:20:24 crc kubenswrapper[4700]: I0227 17:20:24.905729 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Feb 27 17:20:24 crc kubenswrapper[4700]: I0227 17:20:24.905914 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-g9cdg" Feb 27 17:20:24 crc kubenswrapper[4700]: I0227 17:20:24.906022 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Feb 27 17:20:24 crc kubenswrapper[4700]: I0227 17:20:24.930359 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.042155 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbhmc\" (UniqueName: \"kubernetes.io/projected/616dbda2-825d-4a95-a3fc-96c44dbd49aa-kube-api-access-jbhmc\") pod \"ovn-northd-0\" (UID: \"616dbda2-825d-4a95-a3fc-96c44dbd49aa\") " pod="openstack/ovn-northd-0" Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.042201 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/616dbda2-825d-4a95-a3fc-96c44dbd49aa-scripts\") pod \"ovn-northd-0\" (UID: \"616dbda2-825d-4a95-a3fc-96c44dbd49aa\") " pod="openstack/ovn-northd-0" Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.042243 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/616dbda2-825d-4a95-a3fc-96c44dbd49aa-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"616dbda2-825d-4a95-a3fc-96c44dbd49aa\") " pod="openstack/ovn-northd-0" Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.042590 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/616dbda2-825d-4a95-a3fc-96c44dbd49aa-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"616dbda2-825d-4a95-a3fc-96c44dbd49aa\") " pod="openstack/ovn-northd-0" Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.042671 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/616dbda2-825d-4a95-a3fc-96c44dbd49aa-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"616dbda2-825d-4a95-a3fc-96c44dbd49aa\") " pod="openstack/ovn-northd-0" Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.042692 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/616dbda2-825d-4a95-a3fc-96c44dbd49aa-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"616dbda2-825d-4a95-a3fc-96c44dbd49aa\") " pod="openstack/ovn-northd-0" Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.042707 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/616dbda2-825d-4a95-a3fc-96c44dbd49aa-config\") pod \"ovn-northd-0\" (UID: \"616dbda2-825d-4a95-a3fc-96c44dbd49aa\") " pod="openstack/ovn-northd-0" Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.144153 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbhmc\" (UniqueName: \"kubernetes.io/projected/616dbda2-825d-4a95-a3fc-96c44dbd49aa-kube-api-access-jbhmc\") pod \"ovn-northd-0\" (UID: \"616dbda2-825d-4a95-a3fc-96c44dbd49aa\") " pod="openstack/ovn-northd-0" Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.144374 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/616dbda2-825d-4a95-a3fc-96c44dbd49aa-scripts\") pod \"ovn-northd-0\" (UID: \"616dbda2-825d-4a95-a3fc-96c44dbd49aa\") " pod="openstack/ovn-northd-0" Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.144412 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/616dbda2-825d-4a95-a3fc-96c44dbd49aa-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"616dbda2-825d-4a95-a3fc-96c44dbd49aa\") " pod="openstack/ovn-northd-0" Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.144443 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/616dbda2-825d-4a95-a3fc-96c44dbd49aa-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"616dbda2-825d-4a95-a3fc-96c44dbd49aa\") " pod="openstack/ovn-northd-0" Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.145700 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/616dbda2-825d-4a95-a3fc-96c44dbd49aa-scripts\") pod \"ovn-northd-0\" (UID: \"616dbda2-825d-4a95-a3fc-96c44dbd49aa\") " pod="openstack/ovn-northd-0" Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.146276 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/616dbda2-825d-4a95-a3fc-96c44dbd49aa-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"616dbda2-825d-4a95-a3fc-96c44dbd49aa\") " pod="openstack/ovn-northd-0" Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.146311 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/616dbda2-825d-4a95-a3fc-96c44dbd49aa-config\") pod \"ovn-northd-0\" (UID: \"616dbda2-825d-4a95-a3fc-96c44dbd49aa\") " pod="openstack/ovn-northd-0" Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.146331 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/616dbda2-825d-4a95-a3fc-96c44dbd49aa-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"616dbda2-825d-4a95-a3fc-96c44dbd49aa\") " pod="openstack/ovn-northd-0" Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.147313 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/616dbda2-825d-4a95-a3fc-96c44dbd49aa-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"616dbda2-825d-4a95-a3fc-96c44dbd49aa\") " pod="openstack/ovn-northd-0" Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.147543 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/616dbda2-825d-4a95-a3fc-96c44dbd49aa-config\") pod \"ovn-northd-0\" (UID: \"616dbda2-825d-4a95-a3fc-96c44dbd49aa\") " pod="openstack/ovn-northd-0" Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.150973 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/616dbda2-825d-4a95-a3fc-96c44dbd49aa-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"616dbda2-825d-4a95-a3fc-96c44dbd49aa\") " pod="openstack/ovn-northd-0" Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.150995 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/616dbda2-825d-4a95-a3fc-96c44dbd49aa-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"616dbda2-825d-4a95-a3fc-96c44dbd49aa\") " pod="openstack/ovn-northd-0" Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.159296 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbhmc\" (UniqueName: \"kubernetes.io/projected/616dbda2-825d-4a95-a3fc-96c44dbd49aa-kube-api-access-jbhmc\") pod \"ovn-northd-0\" (UID: \"616dbda2-825d-4a95-a3fc-96c44dbd49aa\") " pod="openstack/ovn-northd-0" Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.162631 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/616dbda2-825d-4a95-a3fc-96c44dbd49aa-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"616dbda2-825d-4a95-a3fc-96c44dbd49aa\") " pod="openstack/ovn-northd-0" Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.172904 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bbc67fc-tr49q" Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.247128 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4448288d-21f1-418b-95bf-9287e914f52b-config\") pod \"4448288d-21f1-418b-95bf-9287e914f52b\" (UID: \"4448288d-21f1-418b-95bf-9287e914f52b\") " Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.247218 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4448288d-21f1-418b-95bf-9287e914f52b-dns-svc\") pod \"4448288d-21f1-418b-95bf-9287e914f52b\" (UID: \"4448288d-21f1-418b-95bf-9287e914f52b\") " Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.247268 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kr8ft\" (UniqueName: \"kubernetes.io/projected/4448288d-21f1-418b-95bf-9287e914f52b-kube-api-access-kr8ft\") pod \"4448288d-21f1-418b-95bf-9287e914f52b\" (UID: \"4448288d-21f1-418b-95bf-9287e914f52b\") " Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.251135 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4448288d-21f1-418b-95bf-9287e914f52b-kube-api-access-kr8ft" (OuterVolumeSpecName: "kube-api-access-kr8ft") pod "4448288d-21f1-418b-95bf-9287e914f52b" (UID: "4448288d-21f1-418b-95bf-9287e914f52b"). InnerVolumeSpecName "kube-api-access-kr8ft". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.284671 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.297991 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4448288d-21f1-418b-95bf-9287e914f52b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4448288d-21f1-418b-95bf-9287e914f52b" (UID: "4448288d-21f1-418b-95bf-9287e914f52b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.325142 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4448288d-21f1-418b-95bf-9287e914f52b-config" (OuterVolumeSpecName: "config") pod "4448288d-21f1-418b-95bf-9287e914f52b" (UID: "4448288d-21f1-418b-95bf-9287e914f52b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.351225 4700 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4448288d-21f1-418b-95bf-9287e914f52b-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.351526 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kr8ft\" (UniqueName: \"kubernetes.io/projected/4448288d-21f1-418b-95bf-9287e914f52b-kube-api-access-kr8ft\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.351536 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4448288d-21f1-418b-95bf-9287e914f52b-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.372677 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-0076-account-create-update-hwxxq"] Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.390123 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-create-gx4c9"] Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.591534 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223","Type":"ContainerStarted","Data":"19915243585f8f1aafa4cfea99f178e07a0057788c4d3be10a6786672bb97676"} Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.594238 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-gx4c9" event={"ID":"d0612657-780a-4388-9b79-ae606ed152e3","Type":"ContainerStarted","Data":"ad3c29216c80b98f47106f0f67d9b0209aa5fa69cd6364f79536996e0d69f347"} Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.596833 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-m7qsc" event={"ID":"89befb8e-a997-46a9-9db2-24f0dc5bf10d","Type":"ContainerStarted","Data":"efa7faa3a869c2e7843333aaff00f538d61d9b36d5342a1745b26e101297ad21"} Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.598378 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bbc67fc-tr49q" event={"ID":"4448288d-21f1-418b-95bf-9287e914f52b","Type":"ContainerDied","Data":"3f0576771bc811588b00cc850d5921f2073af9431d691c3c088b57cb79943693"} Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.598561 4700 scope.go:117] "RemoveContainer" containerID="36a7354b8a1d35885156629f2499a02f10cf0989a40e38c49fba3fae4670d9aa" Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.598528 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bbc67fc-tr49q" Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.599756 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-0076-account-create-update-hwxxq" event={"ID":"18a5e20e-5c10-4966-9adc-c10670ddd4e3","Type":"ContainerStarted","Data":"3f9756986364e23164a6f6076bd87bafe9c8e575f52868f1f6c43dd7e4903bfd"} Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.602580 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9721431e-046b-4b3b-861e-22a65396981b","Type":"ContainerStarted","Data":"7b9277446cc5fc2593bdc3b3cb6e1b7c283d1c39909d8b7b4ad6eef57493d935"} Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.666156 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=-9223371979.188635 podStartE2EDuration="57.666140949s" podCreationTimestamp="2026-02-27 17:19:28 +0000 UTC" firstStartedPulling="2026-02-27 17:19:44.285036425 +0000 UTC m=+1144.270349172" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:20:25.620141372 +0000 UTC m=+1185.605454109" watchObservedRunningTime="2026-02-27 17:20:25.666140949 +0000 UTC m=+1185.651453696" Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.669892 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-m7qsc" podStartSLOduration=2.215213881 podStartE2EDuration="7.669875448s" podCreationTimestamp="2026-02-27 17:20:18 +0000 UTC" firstStartedPulling="2026-02-27 17:20:19.378365155 +0000 UTC m=+1179.363677902" lastFinishedPulling="2026-02-27 17:20:24.833026722 +0000 UTC m=+1184.818339469" observedRunningTime="2026-02-27 17:20:25.665175853 +0000 UTC m=+1185.650488600" watchObservedRunningTime="2026-02-27 17:20:25.669875448 +0000 UTC m=+1185.655188195" Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.695866 4700 scope.go:117] "RemoveContainer" containerID="c49d783ad0bd870fae780433ddf4578161bc63d1f014a176c5133aaf46fb2365" Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.698725 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bbc67fc-tr49q"] Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.704901 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7bbc67fc-tr49q"] Feb 27 17:20:25 crc kubenswrapper[4700]: I0227 17:20:25.831363 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 27 17:20:25 crc kubenswrapper[4700]: W0227 17:20:25.834944 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod616dbda2_825d_4a95_a3fc_96c44dbd49aa.slice/crio-e1b315ff0d594ecd230dcbdbb9d9b2b1ec2dabdd6c9ecf85c6888afa956c01aa WatchSource:0}: Error finding container e1b315ff0d594ecd230dcbdbb9d9b2b1ec2dabdd6c9ecf85c6888afa956c01aa: Status 404 returned error can't find the container with id e1b315ff0d594ecd230dcbdbb9d9b2b1ec2dabdd6c9ecf85c6888afa956c01aa Feb 27 17:20:26 crc kubenswrapper[4700]: I0227 17:20:26.619021 4700 generic.go:334] "Generic (PLEG): container finished" podID="d0612657-780a-4388-9b79-ae606ed152e3" containerID="d493357cc638860117531b3c4db652280791a199e851e8bf7c70f533308e5f06" exitCode=0 Feb 27 17:20:26 crc kubenswrapper[4700]: I0227 17:20:26.619387 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-gx4c9" event={"ID":"d0612657-780a-4388-9b79-ae606ed152e3","Type":"ContainerDied","Data":"d493357cc638860117531b3c4db652280791a199e851e8bf7c70f533308e5f06"} Feb 27 17:20:26 crc kubenswrapper[4700]: I0227 17:20:26.629308 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"616dbda2-825d-4a95-a3fc-96c44dbd49aa","Type":"ContainerStarted","Data":"e1b315ff0d594ecd230dcbdbb9d9b2b1ec2dabdd6c9ecf85c6888afa956c01aa"} Feb 27 17:20:26 crc kubenswrapper[4700]: I0227 17:20:26.631234 4700 generic.go:334] "Generic (PLEG): container finished" podID="18a5e20e-5c10-4966-9adc-c10670ddd4e3" containerID="f90f21dfd58e8098c3e9c424632b9b719bc730c56acfc19e81fcfda0344ecaeb" exitCode=0 Feb 27 17:20:26 crc kubenswrapper[4700]: I0227 17:20:26.632117 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-0076-account-create-update-hwxxq" event={"ID":"18a5e20e-5c10-4966-9adc-c10670ddd4e3","Type":"ContainerDied","Data":"f90f21dfd58e8098c3e9c424632b9b719bc730c56acfc19e81fcfda0344ecaeb"} Feb 27 17:20:26 crc kubenswrapper[4700]: I0227 17:20:26.994393 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4448288d-21f1-418b-95bf-9287e914f52b" path="/var/lib/kubelet/pods/4448288d-21f1-418b-95bf-9287e914f52b/volumes" Feb 27 17:20:27 crc kubenswrapper[4700]: I0227 17:20:27.452260 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-rtn6x"] Feb 27 17:20:27 crc kubenswrapper[4700]: E0227 17:20:27.452787 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4448288d-21f1-418b-95bf-9287e914f52b" containerName="dnsmasq-dns" Feb 27 17:20:27 crc kubenswrapper[4700]: I0227 17:20:27.452811 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="4448288d-21f1-418b-95bf-9287e914f52b" containerName="dnsmasq-dns" Feb 27 17:20:27 crc kubenswrapper[4700]: E0227 17:20:27.452828 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4448288d-21f1-418b-95bf-9287e914f52b" containerName="init" Feb 27 17:20:27 crc kubenswrapper[4700]: I0227 17:20:27.452840 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="4448288d-21f1-418b-95bf-9287e914f52b" containerName="init" Feb 27 17:20:27 crc kubenswrapper[4700]: I0227 17:20:27.453045 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="4448288d-21f1-418b-95bf-9287e914f52b" containerName="dnsmasq-dns" Feb 27 17:20:27 crc kubenswrapper[4700]: I0227 17:20:27.453917 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-rtn6x" Feb 27 17:20:27 crc kubenswrapper[4700]: I0227 17:20:27.456541 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Feb 27 17:20:27 crc kubenswrapper[4700]: I0227 17:20:27.464114 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-rtn6x"] Feb 27 17:20:27 crc kubenswrapper[4700]: I0227 17:20:27.509316 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnkw4\" (UniqueName: \"kubernetes.io/projected/a28896eb-9f45-43fa-a078-8ae9175cb25d-kube-api-access-pnkw4\") pod \"root-account-create-update-rtn6x\" (UID: \"a28896eb-9f45-43fa-a078-8ae9175cb25d\") " pod="openstack/root-account-create-update-rtn6x" Feb 27 17:20:27 crc kubenswrapper[4700]: I0227 17:20:27.509369 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a28896eb-9f45-43fa-a078-8ae9175cb25d-operator-scripts\") pod \"root-account-create-update-rtn6x\" (UID: \"a28896eb-9f45-43fa-a078-8ae9175cb25d\") " pod="openstack/root-account-create-update-rtn6x" Feb 27 17:20:27 crc kubenswrapper[4700]: I0227 17:20:27.610861 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnkw4\" (UniqueName: \"kubernetes.io/projected/a28896eb-9f45-43fa-a078-8ae9175cb25d-kube-api-access-pnkw4\") pod \"root-account-create-update-rtn6x\" (UID: \"a28896eb-9f45-43fa-a078-8ae9175cb25d\") " pod="openstack/root-account-create-update-rtn6x" Feb 27 17:20:27 crc kubenswrapper[4700]: I0227 17:20:27.610932 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a28896eb-9f45-43fa-a078-8ae9175cb25d-operator-scripts\") pod \"root-account-create-update-rtn6x\" (UID: \"a28896eb-9f45-43fa-a078-8ae9175cb25d\") " pod="openstack/root-account-create-update-rtn6x" Feb 27 17:20:27 crc kubenswrapper[4700]: I0227 17:20:27.612294 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a28896eb-9f45-43fa-a078-8ae9175cb25d-operator-scripts\") pod \"root-account-create-update-rtn6x\" (UID: \"a28896eb-9f45-43fa-a078-8ae9175cb25d\") " pod="openstack/root-account-create-update-rtn6x" Feb 27 17:20:27 crc kubenswrapper[4700]: I0227 17:20:27.635751 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnkw4\" (UniqueName: \"kubernetes.io/projected/a28896eb-9f45-43fa-a078-8ae9175cb25d-kube-api-access-pnkw4\") pod \"root-account-create-update-rtn6x\" (UID: \"a28896eb-9f45-43fa-a078-8ae9175cb25d\") " pod="openstack/root-account-create-update-rtn6x" Feb 27 17:20:27 crc kubenswrapper[4700]: I0227 17:20:27.645297 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"616dbda2-825d-4a95-a3fc-96c44dbd49aa","Type":"ContainerStarted","Data":"bc1bc0b285a72d86957d30ef825213fe45d6e28c56c4e122b6ecc1057d5d5ada"} Feb 27 17:20:27 crc kubenswrapper[4700]: I0227 17:20:27.645370 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"616dbda2-825d-4a95-a3fc-96c44dbd49aa","Type":"ContainerStarted","Data":"2d0117da6ff272438b74de7205398789dcd35f6d6b0a6e4f44af9945595bdc72"} Feb 27 17:20:27 crc kubenswrapper[4700]: I0227 17:20:27.645501 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Feb 27 17:20:27 crc kubenswrapper[4700]: I0227 17:20:27.650046 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9721431e-046b-4b3b-861e-22a65396981b","Type":"ContainerStarted","Data":"4016795d07ac4553ff298150bb0cf07997d9258aee1bdce657c1c5422071402b"} Feb 27 17:20:27 crc kubenswrapper[4700]: I0227 17:20:27.670720 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.085666448 podStartE2EDuration="3.670697984s" podCreationTimestamp="2026-02-27 17:20:24 +0000 UTC" firstStartedPulling="2026-02-27 17:20:25.836835044 +0000 UTC m=+1185.822147781" lastFinishedPulling="2026-02-27 17:20:26.42186657 +0000 UTC m=+1186.407179317" observedRunningTime="2026-02-27 17:20:27.6648667 +0000 UTC m=+1187.650179477" watchObservedRunningTime="2026-02-27 17:20:27.670697984 +0000 UTC m=+1187.656010761" Feb 27 17:20:27 crc kubenswrapper[4700]: I0227 17:20:27.780028 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-rtn6x" Feb 27 17:20:28 crc kubenswrapper[4700]: I0227 17:20:28.116034 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-0076-account-create-update-hwxxq" Feb 27 17:20:28 crc kubenswrapper[4700]: I0227 17:20:28.124953 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-gx4c9" Feb 27 17:20:28 crc kubenswrapper[4700]: I0227 17:20:28.223454 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnp75\" (UniqueName: \"kubernetes.io/projected/18a5e20e-5c10-4966-9adc-c10670ddd4e3-kube-api-access-mnp75\") pod \"18a5e20e-5c10-4966-9adc-c10670ddd4e3\" (UID: \"18a5e20e-5c10-4966-9adc-c10670ddd4e3\") " Feb 27 17:20:28 crc kubenswrapper[4700]: I0227 17:20:28.223650 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18a5e20e-5c10-4966-9adc-c10670ddd4e3-operator-scripts\") pod \"18a5e20e-5c10-4966-9adc-c10670ddd4e3\" (UID: \"18a5e20e-5c10-4966-9adc-c10670ddd4e3\") " Feb 27 17:20:28 crc kubenswrapper[4700]: I0227 17:20:28.224649 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18a5e20e-5c10-4966-9adc-c10670ddd4e3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "18a5e20e-5c10-4966-9adc-c10670ddd4e3" (UID: "18a5e20e-5c10-4966-9adc-c10670ddd4e3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:20:28 crc kubenswrapper[4700]: I0227 17:20:28.231752 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18a5e20e-5c10-4966-9adc-c10670ddd4e3-kube-api-access-mnp75" (OuterVolumeSpecName: "kube-api-access-mnp75") pod "18a5e20e-5c10-4966-9adc-c10670ddd4e3" (UID: "18a5e20e-5c10-4966-9adc-c10670ddd4e3"). InnerVolumeSpecName "kube-api-access-mnp75". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:20:28 crc kubenswrapper[4700]: I0227 17:20:28.310694 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-rtn6x"] Feb 27 17:20:28 crc kubenswrapper[4700]: W0227 17:20:28.313531 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda28896eb_9f45_43fa_a078_8ae9175cb25d.slice/crio-20b3e99076a1920df2f7dcc64116aa8eb129c6501f838c6690a19dcde220b103 WatchSource:0}: Error finding container 20b3e99076a1920df2f7dcc64116aa8eb129c6501f838c6690a19dcde220b103: Status 404 returned error can't find the container with id 20b3e99076a1920df2f7dcc64116aa8eb129c6501f838c6690a19dcde220b103 Feb 27 17:20:28 crc kubenswrapper[4700]: I0227 17:20:28.325347 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vr7fj\" (UniqueName: \"kubernetes.io/projected/d0612657-780a-4388-9b79-ae606ed152e3-kube-api-access-vr7fj\") pod \"d0612657-780a-4388-9b79-ae606ed152e3\" (UID: \"d0612657-780a-4388-9b79-ae606ed152e3\") " Feb 27 17:20:28 crc kubenswrapper[4700]: I0227 17:20:28.325614 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d0612657-780a-4388-9b79-ae606ed152e3-operator-scripts\") pod \"d0612657-780a-4388-9b79-ae606ed152e3\" (UID: \"d0612657-780a-4388-9b79-ae606ed152e3\") " Feb 27 17:20:28 crc kubenswrapper[4700]: I0227 17:20:28.326202 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0612657-780a-4388-9b79-ae606ed152e3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d0612657-780a-4388-9b79-ae606ed152e3" (UID: "d0612657-780a-4388-9b79-ae606ed152e3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:20:28 crc kubenswrapper[4700]: I0227 17:20:28.326547 4700 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d0612657-780a-4388-9b79-ae606ed152e3-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:28 crc kubenswrapper[4700]: I0227 17:20:28.326589 4700 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18a5e20e-5c10-4966-9adc-c10670ddd4e3-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:28 crc kubenswrapper[4700]: I0227 17:20:28.326614 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnp75\" (UniqueName: \"kubernetes.io/projected/18a5e20e-5c10-4966-9adc-c10670ddd4e3-kube-api-access-mnp75\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:28 crc kubenswrapper[4700]: I0227 17:20:28.330353 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0612657-780a-4388-9b79-ae606ed152e3-kube-api-access-vr7fj" (OuterVolumeSpecName: "kube-api-access-vr7fj") pod "d0612657-780a-4388-9b79-ae606ed152e3" (UID: "d0612657-780a-4388-9b79-ae606ed152e3"). InnerVolumeSpecName "kube-api-access-vr7fj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:20:28 crc kubenswrapper[4700]: I0227 17:20:28.427971 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vr7fj\" (UniqueName: \"kubernetes.io/projected/d0612657-780a-4388-9b79-ae606ed152e3-kube-api-access-vr7fj\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:28 crc kubenswrapper[4700]: I0227 17:20:28.664037 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-gx4c9" event={"ID":"d0612657-780a-4388-9b79-ae606ed152e3","Type":"ContainerDied","Data":"ad3c29216c80b98f47106f0f67d9b0209aa5fa69cd6364f79536996e0d69f347"} Feb 27 17:20:28 crc kubenswrapper[4700]: I0227 17:20:28.664072 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad3c29216c80b98f47106f0f67d9b0209aa5fa69cd6364f79536996e0d69f347" Feb 27 17:20:28 crc kubenswrapper[4700]: I0227 17:20:28.664090 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-gx4c9" Feb 27 17:20:28 crc kubenswrapper[4700]: I0227 17:20:28.665783 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-0076-account-create-update-hwxxq" Feb 27 17:20:28 crc kubenswrapper[4700]: I0227 17:20:28.665763 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-0076-account-create-update-hwxxq" event={"ID":"18a5e20e-5c10-4966-9adc-c10670ddd4e3","Type":"ContainerDied","Data":"3f9756986364e23164a6f6076bd87bafe9c8e575f52868f1f6c43dd7e4903bfd"} Feb 27 17:20:28 crc kubenswrapper[4700]: I0227 17:20:28.666081 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f9756986364e23164a6f6076bd87bafe9c8e575f52868f1f6c43dd7e4903bfd" Feb 27 17:20:28 crc kubenswrapper[4700]: I0227 17:20:28.673666 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-rtn6x" event={"ID":"a28896eb-9f45-43fa-a078-8ae9175cb25d","Type":"ContainerStarted","Data":"20b3e99076a1920df2f7dcc64116aa8eb129c6501f838c6690a19dcde220b103"} Feb 27 17:20:30 crc kubenswrapper[4700]: I0227 17:20:30.342582 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Feb 27 17:20:30 crc kubenswrapper[4700]: I0227 17:20:30.342924 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Feb 27 17:20:30 crc kubenswrapper[4700]: I0227 17:20:30.465876 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9df57c73-8d37-4f7d-b62f-209210dbba42-etc-swift\") pod \"swift-storage-0\" (UID: \"9df57c73-8d37-4f7d-b62f-209210dbba42\") " pod="openstack/swift-storage-0" Feb 27 17:20:30 crc kubenswrapper[4700]: E0227 17:20:30.466090 4700 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 27 17:20:30 crc kubenswrapper[4700]: E0227 17:20:30.466108 4700 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 27 17:20:30 crc kubenswrapper[4700]: E0227 17:20:30.466171 4700 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9df57c73-8d37-4f7d-b62f-209210dbba42-etc-swift podName:9df57c73-8d37-4f7d-b62f-209210dbba42 nodeName:}" failed. No retries permitted until 2026-02-27 17:20:46.46615152 +0000 UTC m=+1206.451464287 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9df57c73-8d37-4f7d-b62f-209210dbba42-etc-swift") pod "swift-storage-0" (UID: "9df57c73-8d37-4f7d-b62f-209210dbba42") : configmap "swift-ring-files" not found Feb 27 17:20:30 crc kubenswrapper[4700]: I0227 17:20:30.695189 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-rtn6x" event={"ID":"a28896eb-9f45-43fa-a078-8ae9175cb25d","Type":"ContainerStarted","Data":"7b07b3db778ef64ad19e650e40a5aaf03eb204e105f667cd3736c791ca8a4dad"} Feb 27 17:20:30 crc kubenswrapper[4700]: I0227 17:20:30.790453 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-rtn6x" podStartSLOduration=3.790436428 podStartE2EDuration="3.790436428s" podCreationTimestamp="2026-02-27 17:20:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:20:30.717842738 +0000 UTC m=+1190.703155475" watchObservedRunningTime="2026-02-27 17:20:30.790436428 +0000 UTC m=+1190.775749165" Feb 27 17:20:30 crc kubenswrapper[4700]: I0227 17:20:30.792357 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-2ppp5"] Feb 27 17:20:30 crc kubenswrapper[4700]: E0227 17:20:30.792687 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0612657-780a-4388-9b79-ae606ed152e3" containerName="mariadb-database-create" Feb 27 17:20:30 crc kubenswrapper[4700]: I0227 17:20:30.792705 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0612657-780a-4388-9b79-ae606ed152e3" containerName="mariadb-database-create" Feb 27 17:20:30 crc kubenswrapper[4700]: E0227 17:20:30.792735 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18a5e20e-5c10-4966-9adc-c10670ddd4e3" containerName="mariadb-account-create-update" Feb 27 17:20:30 crc kubenswrapper[4700]: I0227 17:20:30.792745 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="18a5e20e-5c10-4966-9adc-c10670ddd4e3" containerName="mariadb-account-create-update" Feb 27 17:20:30 crc kubenswrapper[4700]: I0227 17:20:30.792928 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0612657-780a-4388-9b79-ae606ed152e3" containerName="mariadb-database-create" Feb 27 17:20:30 crc kubenswrapper[4700]: I0227 17:20:30.792956 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="18a5e20e-5c10-4966-9adc-c10670ddd4e3" containerName="mariadb-account-create-update" Feb 27 17:20:30 crc kubenswrapper[4700]: I0227 17:20:30.793573 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-2ppp5" Feb 27 17:20:30 crc kubenswrapper[4700]: I0227 17:20:30.808033 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-2ppp5"] Feb 27 17:20:30 crc kubenswrapper[4700]: I0227 17:20:30.873796 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/42f288bd-13a4-4e0c-a616-b09462acd2bf-operator-scripts\") pod \"glance-db-create-2ppp5\" (UID: \"42f288bd-13a4-4e0c-a616-b09462acd2bf\") " pod="openstack/glance-db-create-2ppp5" Feb 27 17:20:30 crc kubenswrapper[4700]: I0227 17:20:30.873871 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srlx5\" (UniqueName: \"kubernetes.io/projected/42f288bd-13a4-4e0c-a616-b09462acd2bf-kube-api-access-srlx5\") pod \"glance-db-create-2ppp5\" (UID: \"42f288bd-13a4-4e0c-a616-b09462acd2bf\") " pod="openstack/glance-db-create-2ppp5" Feb 27 17:20:30 crc kubenswrapper[4700]: I0227 17:20:30.912430 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-b838-account-create-update-g6qs7"] Feb 27 17:20:30 crc kubenswrapper[4700]: I0227 17:20:30.913825 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b838-account-create-update-g6qs7" Feb 27 17:20:30 crc kubenswrapper[4700]: I0227 17:20:30.916867 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Feb 27 17:20:30 crc kubenswrapper[4700]: I0227 17:20:30.923860 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-b838-account-create-update-g6qs7"] Feb 27 17:20:30 crc kubenswrapper[4700]: I0227 17:20:30.975293 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/42f288bd-13a4-4e0c-a616-b09462acd2bf-operator-scripts\") pod \"glance-db-create-2ppp5\" (UID: \"42f288bd-13a4-4e0c-a616-b09462acd2bf\") " pod="openstack/glance-db-create-2ppp5" Feb 27 17:20:30 crc kubenswrapper[4700]: I0227 17:20:30.975338 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srlx5\" (UniqueName: \"kubernetes.io/projected/42f288bd-13a4-4e0c-a616-b09462acd2bf-kube-api-access-srlx5\") pod \"glance-db-create-2ppp5\" (UID: \"42f288bd-13a4-4e0c-a616-b09462acd2bf\") " pod="openstack/glance-db-create-2ppp5" Feb 27 17:20:30 crc kubenswrapper[4700]: I0227 17:20:30.976243 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/42f288bd-13a4-4e0c-a616-b09462acd2bf-operator-scripts\") pod \"glance-db-create-2ppp5\" (UID: \"42f288bd-13a4-4e0c-a616-b09462acd2bf\") " pod="openstack/glance-db-create-2ppp5" Feb 27 17:20:30 crc kubenswrapper[4700]: I0227 17:20:30.997603 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srlx5\" (UniqueName: \"kubernetes.io/projected/42f288bd-13a4-4e0c-a616-b09462acd2bf-kube-api-access-srlx5\") pod \"glance-db-create-2ppp5\" (UID: \"42f288bd-13a4-4e0c-a616-b09462acd2bf\") " pod="openstack/glance-db-create-2ppp5" Feb 27 17:20:31 crc kubenswrapper[4700]: I0227 17:20:31.076868 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6dhq\" (UniqueName: \"kubernetes.io/projected/a889ce63-746b-45bc-9f99-487f7416a80c-kube-api-access-t6dhq\") pod \"glance-b838-account-create-update-g6qs7\" (UID: \"a889ce63-746b-45bc-9f99-487f7416a80c\") " pod="openstack/glance-b838-account-create-update-g6qs7" Feb 27 17:20:31 crc kubenswrapper[4700]: I0227 17:20:31.077013 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a889ce63-746b-45bc-9f99-487f7416a80c-operator-scripts\") pod \"glance-b838-account-create-update-g6qs7\" (UID: \"a889ce63-746b-45bc-9f99-487f7416a80c\") " pod="openstack/glance-b838-account-create-update-g6qs7" Feb 27 17:20:31 crc kubenswrapper[4700]: I0227 17:20:31.115052 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-2ppp5" Feb 27 17:20:31 crc kubenswrapper[4700]: I0227 17:20:31.192782 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a889ce63-746b-45bc-9f99-487f7416a80c-operator-scripts\") pod \"glance-b838-account-create-update-g6qs7\" (UID: \"a889ce63-746b-45bc-9f99-487f7416a80c\") " pod="openstack/glance-b838-account-create-update-g6qs7" Feb 27 17:20:31 crc kubenswrapper[4700]: I0227 17:20:31.192953 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6dhq\" (UniqueName: \"kubernetes.io/projected/a889ce63-746b-45bc-9f99-487f7416a80c-kube-api-access-t6dhq\") pod \"glance-b838-account-create-update-g6qs7\" (UID: \"a889ce63-746b-45bc-9f99-487f7416a80c\") " pod="openstack/glance-b838-account-create-update-g6qs7" Feb 27 17:20:31 crc kubenswrapper[4700]: I0227 17:20:31.193942 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a889ce63-746b-45bc-9f99-487f7416a80c-operator-scripts\") pod \"glance-b838-account-create-update-g6qs7\" (UID: \"a889ce63-746b-45bc-9f99-487f7416a80c\") " pod="openstack/glance-b838-account-create-update-g6qs7" Feb 27 17:20:31 crc kubenswrapper[4700]: I0227 17:20:31.209574 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6dhq\" (UniqueName: \"kubernetes.io/projected/a889ce63-746b-45bc-9f99-487f7416a80c-kube-api-access-t6dhq\") pod \"glance-b838-account-create-update-g6qs7\" (UID: \"a889ce63-746b-45bc-9f99-487f7416a80c\") " pod="openstack/glance-b838-account-create-update-g6qs7" Feb 27 17:20:31 crc kubenswrapper[4700]: I0227 17:20:31.231203 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b838-account-create-update-g6qs7" Feb 27 17:20:31 crc kubenswrapper[4700]: I0227 17:20:31.294307 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Feb 27 17:20:31 crc kubenswrapper[4700]: I0227 17:20:31.446582 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Feb 27 17:20:31 crc kubenswrapper[4700]: I0227 17:20:31.703706 4700 generic.go:334] "Generic (PLEG): container finished" podID="a28896eb-9f45-43fa-a078-8ae9175cb25d" containerID="7b07b3db778ef64ad19e650e40a5aaf03eb204e105f667cd3736c791ca8a4dad" exitCode=0 Feb 27 17:20:31 crc kubenswrapper[4700]: I0227 17:20:31.704612 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-rtn6x" event={"ID":"a28896eb-9f45-43fa-a078-8ae9175cb25d","Type":"ContainerDied","Data":"7b07b3db778ef64ad19e650e40a5aaf03eb204e105f667cd3736c791ca8a4dad"} Feb 27 17:20:31 crc kubenswrapper[4700]: I0227 17:20:31.785454 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-vfjlj"] Feb 27 17:20:31 crc kubenswrapper[4700]: I0227 17:20:31.786594 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-vfjlj" Feb 27 17:20:31 crc kubenswrapper[4700]: I0227 17:20:31.791711 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-vfjlj"] Feb 27 17:20:31 crc kubenswrapper[4700]: I0227 17:20:31.874880 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-4fb2-account-create-update-h494r"] Feb 27 17:20:31 crc kubenswrapper[4700]: I0227 17:20:31.875962 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-4fb2-account-create-update-h494r" Feb 27 17:20:31 crc kubenswrapper[4700]: I0227 17:20:31.879407 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Feb 27 17:20:31 crc kubenswrapper[4700]: I0227 17:20:31.889622 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-4fb2-account-create-update-h494r"] Feb 27 17:20:31 crc kubenswrapper[4700]: I0227 17:20:31.926348 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b393dd07-e44c-499b-8a95-81c50f78efe8-operator-scripts\") pod \"keystone-db-create-vfjlj\" (UID: \"b393dd07-e44c-499b-8a95-81c50f78efe8\") " pod="openstack/keystone-db-create-vfjlj" Feb 27 17:20:31 crc kubenswrapper[4700]: I0227 17:20:31.926634 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cw5rr\" (UniqueName: \"kubernetes.io/projected/b393dd07-e44c-499b-8a95-81c50f78efe8-kube-api-access-cw5rr\") pod \"keystone-db-create-vfjlj\" (UID: \"b393dd07-e44c-499b-8a95-81c50f78efe8\") " pod="openstack/keystone-db-create-vfjlj" Feb 27 17:20:31 crc kubenswrapper[4700]: I0227 17:20:31.976161 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-glgwk"] Feb 27 17:20:31 crc kubenswrapper[4700]: I0227 17:20:31.977445 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-glgwk" Feb 27 17:20:31 crc kubenswrapper[4700]: I0227 17:20:31.998884 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-glgwk"] Feb 27 17:20:32 crc kubenswrapper[4700]: I0227 17:20:32.032446 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rbmz\" (UniqueName: \"kubernetes.io/projected/c7812092-9d78-409a-80f8-4a4db92af41e-kube-api-access-2rbmz\") pod \"keystone-4fb2-account-create-update-h494r\" (UID: \"c7812092-9d78-409a-80f8-4a4db92af41e\") " pod="openstack/keystone-4fb2-account-create-update-h494r" Feb 27 17:20:32 crc kubenswrapper[4700]: I0227 17:20:32.032511 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cw5rr\" (UniqueName: \"kubernetes.io/projected/b393dd07-e44c-499b-8a95-81c50f78efe8-kube-api-access-cw5rr\") pod \"keystone-db-create-vfjlj\" (UID: \"b393dd07-e44c-499b-8a95-81c50f78efe8\") " pod="openstack/keystone-db-create-vfjlj" Feb 27 17:20:32 crc kubenswrapper[4700]: I0227 17:20:32.032658 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7812092-9d78-409a-80f8-4a4db92af41e-operator-scripts\") pod \"keystone-4fb2-account-create-update-h494r\" (UID: \"c7812092-9d78-409a-80f8-4a4db92af41e\") " pod="openstack/keystone-4fb2-account-create-update-h494r" Feb 27 17:20:32 crc kubenswrapper[4700]: I0227 17:20:32.032881 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b393dd07-e44c-499b-8a95-81c50f78efe8-operator-scripts\") pod \"keystone-db-create-vfjlj\" (UID: \"b393dd07-e44c-499b-8a95-81c50f78efe8\") " pod="openstack/keystone-db-create-vfjlj" Feb 27 17:20:32 crc kubenswrapper[4700]: I0227 17:20:32.033758 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b393dd07-e44c-499b-8a95-81c50f78efe8-operator-scripts\") pod \"keystone-db-create-vfjlj\" (UID: \"b393dd07-e44c-499b-8a95-81c50f78efe8\") " pod="openstack/keystone-db-create-vfjlj" Feb 27 17:20:32 crc kubenswrapper[4700]: I0227 17:20:32.056380 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cw5rr\" (UniqueName: \"kubernetes.io/projected/b393dd07-e44c-499b-8a95-81c50f78efe8-kube-api-access-cw5rr\") pod \"keystone-db-create-vfjlj\" (UID: \"b393dd07-e44c-499b-8a95-81c50f78efe8\") " pod="openstack/keystone-db-create-vfjlj" Feb 27 17:20:32 crc kubenswrapper[4700]: I0227 17:20:32.095785 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-94be-account-create-update-snkzp"] Feb 27 17:20:32 crc kubenswrapper[4700]: I0227 17:20:32.096982 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-94be-account-create-update-snkzp" Feb 27 17:20:32 crc kubenswrapper[4700]: I0227 17:20:32.101315 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Feb 27 17:20:32 crc kubenswrapper[4700]: I0227 17:20:32.107837 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-94be-account-create-update-snkzp"] Feb 27 17:20:32 crc kubenswrapper[4700]: I0227 17:20:32.140548 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rbmz\" (UniqueName: \"kubernetes.io/projected/c7812092-9d78-409a-80f8-4a4db92af41e-kube-api-access-2rbmz\") pod \"keystone-4fb2-account-create-update-h494r\" (UID: \"c7812092-9d78-409a-80f8-4a4db92af41e\") " pod="openstack/keystone-4fb2-account-create-update-h494r" Feb 27 17:20:32 crc kubenswrapper[4700]: I0227 17:20:32.140610 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7812092-9d78-409a-80f8-4a4db92af41e-operator-scripts\") pod \"keystone-4fb2-account-create-update-h494r\" (UID: \"c7812092-9d78-409a-80f8-4a4db92af41e\") " pod="openstack/keystone-4fb2-account-create-update-h494r" Feb 27 17:20:32 crc kubenswrapper[4700]: I0227 17:20:32.140724 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0af77e17-d33d-4d3b-8246-e38499ea68f7-operator-scripts\") pod \"placement-db-create-glgwk\" (UID: \"0af77e17-d33d-4d3b-8246-e38499ea68f7\") " pod="openstack/placement-db-create-glgwk" Feb 27 17:20:32 crc kubenswrapper[4700]: I0227 17:20:32.140808 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgw9s\" (UniqueName: \"kubernetes.io/projected/0af77e17-d33d-4d3b-8246-e38499ea68f7-kube-api-access-mgw9s\") pod \"placement-db-create-glgwk\" (UID: \"0af77e17-d33d-4d3b-8246-e38499ea68f7\") " pod="openstack/placement-db-create-glgwk" Feb 27 17:20:32 crc kubenswrapper[4700]: I0227 17:20:32.141628 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7812092-9d78-409a-80f8-4a4db92af41e-operator-scripts\") pod \"keystone-4fb2-account-create-update-h494r\" (UID: \"c7812092-9d78-409a-80f8-4a4db92af41e\") " pod="openstack/keystone-4fb2-account-create-update-h494r" Feb 27 17:20:32 crc kubenswrapper[4700]: I0227 17:20:32.144579 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-vfjlj" Feb 27 17:20:32 crc kubenswrapper[4700]: I0227 17:20:32.188250 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rbmz\" (UniqueName: \"kubernetes.io/projected/c7812092-9d78-409a-80f8-4a4db92af41e-kube-api-access-2rbmz\") pod \"keystone-4fb2-account-create-update-h494r\" (UID: \"c7812092-9d78-409a-80f8-4a4db92af41e\") " pod="openstack/keystone-4fb2-account-create-update-h494r" Feb 27 17:20:32 crc kubenswrapper[4700]: I0227 17:20:32.193434 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-4fb2-account-create-update-h494r" Feb 27 17:20:32 crc kubenswrapper[4700]: I0227 17:20:32.241719 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0af77e17-d33d-4d3b-8246-e38499ea68f7-operator-scripts\") pod \"placement-db-create-glgwk\" (UID: \"0af77e17-d33d-4d3b-8246-e38499ea68f7\") " pod="openstack/placement-db-create-glgwk" Feb 27 17:20:32 crc kubenswrapper[4700]: I0227 17:20:32.241801 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgw9s\" (UniqueName: \"kubernetes.io/projected/0af77e17-d33d-4d3b-8246-e38499ea68f7-kube-api-access-mgw9s\") pod \"placement-db-create-glgwk\" (UID: \"0af77e17-d33d-4d3b-8246-e38499ea68f7\") " pod="openstack/placement-db-create-glgwk" Feb 27 17:20:32 crc kubenswrapper[4700]: I0227 17:20:32.241857 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c9e7190d-50d9-4287-8f66-5aa298c5cdc7-operator-scripts\") pod \"placement-94be-account-create-update-snkzp\" (UID: \"c9e7190d-50d9-4287-8f66-5aa298c5cdc7\") " pod="openstack/placement-94be-account-create-update-snkzp" Feb 27 17:20:32 crc kubenswrapper[4700]: I0227 17:20:32.241908 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnggv\" (UniqueName: \"kubernetes.io/projected/c9e7190d-50d9-4287-8f66-5aa298c5cdc7-kube-api-access-qnggv\") pod \"placement-94be-account-create-update-snkzp\" (UID: \"c9e7190d-50d9-4287-8f66-5aa298c5cdc7\") " pod="openstack/placement-94be-account-create-update-snkzp" Feb 27 17:20:32 crc kubenswrapper[4700]: I0227 17:20:32.242577 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0af77e17-d33d-4d3b-8246-e38499ea68f7-operator-scripts\") pod \"placement-db-create-glgwk\" (UID: \"0af77e17-d33d-4d3b-8246-e38499ea68f7\") " pod="openstack/placement-db-create-glgwk" Feb 27 17:20:32 crc kubenswrapper[4700]: I0227 17:20:32.270173 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgw9s\" (UniqueName: \"kubernetes.io/projected/0af77e17-d33d-4d3b-8246-e38499ea68f7-kube-api-access-mgw9s\") pod \"placement-db-create-glgwk\" (UID: \"0af77e17-d33d-4d3b-8246-e38499ea68f7\") " pod="openstack/placement-db-create-glgwk" Feb 27 17:20:32 crc kubenswrapper[4700]: I0227 17:20:32.293300 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-glgwk" Feb 27 17:20:32 crc kubenswrapper[4700]: I0227 17:20:32.343548 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnggv\" (UniqueName: \"kubernetes.io/projected/c9e7190d-50d9-4287-8f66-5aa298c5cdc7-kube-api-access-qnggv\") pod \"placement-94be-account-create-update-snkzp\" (UID: \"c9e7190d-50d9-4287-8f66-5aa298c5cdc7\") " pod="openstack/placement-94be-account-create-update-snkzp" Feb 27 17:20:32 crc kubenswrapper[4700]: I0227 17:20:32.343895 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c9e7190d-50d9-4287-8f66-5aa298c5cdc7-operator-scripts\") pod \"placement-94be-account-create-update-snkzp\" (UID: \"c9e7190d-50d9-4287-8f66-5aa298c5cdc7\") " pod="openstack/placement-94be-account-create-update-snkzp" Feb 27 17:20:32 crc kubenswrapper[4700]: I0227 17:20:32.344688 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c9e7190d-50d9-4287-8f66-5aa298c5cdc7-operator-scripts\") pod \"placement-94be-account-create-update-snkzp\" (UID: \"c9e7190d-50d9-4287-8f66-5aa298c5cdc7\") " pod="openstack/placement-94be-account-create-update-snkzp" Feb 27 17:20:32 crc kubenswrapper[4700]: I0227 17:20:32.357958 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnggv\" (UniqueName: \"kubernetes.io/projected/c9e7190d-50d9-4287-8f66-5aa298c5cdc7-kube-api-access-qnggv\") pod \"placement-94be-account-create-update-snkzp\" (UID: \"c9e7190d-50d9-4287-8f66-5aa298c5cdc7\") " pod="openstack/placement-94be-account-create-update-snkzp" Feb 27 17:20:32 crc kubenswrapper[4700]: I0227 17:20:32.428762 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-94be-account-create-update-snkzp" Feb 27 17:20:32 crc kubenswrapper[4700]: I0227 17:20:32.712852 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9721431e-046b-4b3b-861e-22a65396981b","Type":"ContainerStarted","Data":"99478540ee16a5f122a1e6e28f39329c9bd265403a641ecd3a881d8272258aee"} Feb 27 17:20:32 crc kubenswrapper[4700]: I0227 17:20:32.741715 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=11.634324697 podStartE2EDuration="59.741700824s" podCreationTimestamp="2026-02-27 17:19:33 +0000 UTC" firstStartedPulling="2026-02-27 17:19:44.309601587 +0000 UTC m=+1144.294914334" lastFinishedPulling="2026-02-27 17:20:32.416977694 +0000 UTC m=+1192.402290461" observedRunningTime="2026-02-27 17:20:32.738603962 +0000 UTC m=+1192.723916709" watchObservedRunningTime="2026-02-27 17:20:32.741700824 +0000 UTC m=+1192.727013571" Feb 27 17:20:32 crc kubenswrapper[4700]: I0227 17:20:32.874755 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-2ppp5"] Feb 27 17:20:33 crc kubenswrapper[4700]: I0227 17:20:33.127214 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-4fb2-account-create-update-h494r"] Feb 27 17:20:33 crc kubenswrapper[4700]: I0227 17:20:33.140778 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-b838-account-create-update-g6qs7"] Feb 27 17:20:33 crc kubenswrapper[4700]: I0227 17:20:33.176317 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-94be-account-create-update-snkzp"] Feb 27 17:20:33 crc kubenswrapper[4700]: I0227 17:20:33.187975 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-vfjlj"] Feb 27 17:20:33 crc kubenswrapper[4700]: I0227 17:20:33.298040 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-glgwk"] Feb 27 17:20:33 crc kubenswrapper[4700]: W0227 17:20:33.354038 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0af77e17_d33d_4d3b_8246_e38499ea68f7.slice/crio-c66482b857f128a7ab062a40ca6b2419764ca4fbb3313d1ed4cf5841b7ec3931 WatchSource:0}: Error finding container c66482b857f128a7ab062a40ca6b2419764ca4fbb3313d1ed4cf5841b7ec3931: Status 404 returned error can't find the container with id c66482b857f128a7ab062a40ca6b2419764ca4fbb3313d1ed4cf5841b7ec3931 Feb 27 17:20:33 crc kubenswrapper[4700]: I0227 17:20:33.355219 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-rtn6x" Feb 27 17:20:33 crc kubenswrapper[4700]: I0227 17:20:33.371077 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a28896eb-9f45-43fa-a078-8ae9175cb25d-operator-scripts\") pod \"a28896eb-9f45-43fa-a078-8ae9175cb25d\" (UID: \"a28896eb-9f45-43fa-a078-8ae9175cb25d\") " Feb 27 17:20:33 crc kubenswrapper[4700]: I0227 17:20:33.371129 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pnkw4\" (UniqueName: \"kubernetes.io/projected/a28896eb-9f45-43fa-a078-8ae9175cb25d-kube-api-access-pnkw4\") pod \"a28896eb-9f45-43fa-a078-8ae9175cb25d\" (UID: \"a28896eb-9f45-43fa-a078-8ae9175cb25d\") " Feb 27 17:20:33 crc kubenswrapper[4700]: I0227 17:20:33.371497 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a28896eb-9f45-43fa-a078-8ae9175cb25d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a28896eb-9f45-43fa-a078-8ae9175cb25d" (UID: "a28896eb-9f45-43fa-a078-8ae9175cb25d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:20:33 crc kubenswrapper[4700]: I0227 17:20:33.382693 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a28896eb-9f45-43fa-a078-8ae9175cb25d-kube-api-access-pnkw4" (OuterVolumeSpecName: "kube-api-access-pnkw4") pod "a28896eb-9f45-43fa-a078-8ae9175cb25d" (UID: "a28896eb-9f45-43fa-a078-8ae9175cb25d"). InnerVolumeSpecName "kube-api-access-pnkw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:20:33 crc kubenswrapper[4700]: I0227 17:20:33.478865 4700 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a28896eb-9f45-43fa-a078-8ae9175cb25d-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:33 crc kubenswrapper[4700]: I0227 17:20:33.478900 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pnkw4\" (UniqueName: \"kubernetes.io/projected/a28896eb-9f45-43fa-a078-8ae9175cb25d-kube-api-access-pnkw4\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:33 crc kubenswrapper[4700]: I0227 17:20:33.719671 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-94be-account-create-update-snkzp" event={"ID":"c9e7190d-50d9-4287-8f66-5aa298c5cdc7","Type":"ContainerStarted","Data":"7dbfbad6525b94cb92f5314afed9b77c4b093833511b704ca4881f88a6e32547"} Feb 27 17:20:33 crc kubenswrapper[4700]: I0227 17:20:33.719752 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-94be-account-create-update-snkzp" event={"ID":"c9e7190d-50d9-4287-8f66-5aa298c5cdc7","Type":"ContainerStarted","Data":"c004aa732bdd1dfcdc4c2449d24f9074f7eb249df8bab48162677d5d48bc01c2"} Feb 27 17:20:33 crc kubenswrapper[4700]: I0227 17:20:33.721125 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-4fb2-account-create-update-h494r" event={"ID":"c7812092-9d78-409a-80f8-4a4db92af41e","Type":"ContainerStarted","Data":"c513a087e1e918ce5827b7e53435b261f731d6a8770be071707b9a006ffa3bd9"} Feb 27 17:20:33 crc kubenswrapper[4700]: I0227 17:20:33.721167 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-4fb2-account-create-update-h494r" event={"ID":"c7812092-9d78-409a-80f8-4a4db92af41e","Type":"ContainerStarted","Data":"91258d57272de973fe0bd80fca9da13739e5bff01d7851b9389d62bf5231a782"} Feb 27 17:20:33 crc kubenswrapper[4700]: I0227 17:20:33.722674 4700 generic.go:334] "Generic (PLEG): container finished" podID="42f288bd-13a4-4e0c-a616-b09462acd2bf" containerID="4de383a1ce255a5abdcc76689987552575b674f8ee3b8a58361c39735e81964a" exitCode=0 Feb 27 17:20:33 crc kubenswrapper[4700]: I0227 17:20:33.722767 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-2ppp5" event={"ID":"42f288bd-13a4-4e0c-a616-b09462acd2bf","Type":"ContainerDied","Data":"4de383a1ce255a5abdcc76689987552575b674f8ee3b8a58361c39735e81964a"} Feb 27 17:20:33 crc kubenswrapper[4700]: I0227 17:20:33.722835 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-2ppp5" event={"ID":"42f288bd-13a4-4e0c-a616-b09462acd2bf","Type":"ContainerStarted","Data":"3f8888338909c2f754cdddb7f35d2f58422df2d973a3e4b5982af39ba84d059c"} Feb 27 17:20:33 crc kubenswrapper[4700]: I0227 17:20:33.724165 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-glgwk" event={"ID":"0af77e17-d33d-4d3b-8246-e38499ea68f7","Type":"ContainerStarted","Data":"3e260f40afc484c8c52c1b94079e2bccc68fd3802ca10b847bcaf2e2dcaf288d"} Feb 27 17:20:33 crc kubenswrapper[4700]: I0227 17:20:33.724498 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-glgwk" event={"ID":"0af77e17-d33d-4d3b-8246-e38499ea68f7","Type":"ContainerStarted","Data":"c66482b857f128a7ab062a40ca6b2419764ca4fbb3313d1ed4cf5841b7ec3931"} Feb 27 17:20:33 crc kubenswrapper[4700]: I0227 17:20:33.725666 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-rtn6x" event={"ID":"a28896eb-9f45-43fa-a078-8ae9175cb25d","Type":"ContainerDied","Data":"20b3e99076a1920df2f7dcc64116aa8eb129c6501f838c6690a19dcde220b103"} Feb 27 17:20:33 crc kubenswrapper[4700]: I0227 17:20:33.725705 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="20b3e99076a1920df2f7dcc64116aa8eb129c6501f838c6690a19dcde220b103" Feb 27 17:20:33 crc kubenswrapper[4700]: I0227 17:20:33.725776 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-rtn6x" Feb 27 17:20:33 crc kubenswrapper[4700]: I0227 17:20:33.732543 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b838-account-create-update-g6qs7" event={"ID":"a889ce63-746b-45bc-9f99-487f7416a80c","Type":"ContainerStarted","Data":"3c2c275c3cc206ac97d16d9dd3bc253b01082dc180beb82d9a0f98a67dae1c47"} Feb 27 17:20:33 crc kubenswrapper[4700]: I0227 17:20:33.732702 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b838-account-create-update-g6qs7" event={"ID":"a889ce63-746b-45bc-9f99-487f7416a80c","Type":"ContainerStarted","Data":"980f7ce9773359e64d0ccb38ff3584b95df7ad7a14366f4b73f55bdda33ad6c7"} Feb 27 17:20:33 crc kubenswrapper[4700]: I0227 17:20:33.734344 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-vfjlj" event={"ID":"b393dd07-e44c-499b-8a95-81c50f78efe8","Type":"ContainerStarted","Data":"aa2b25a6011304a5a553000d459adb9829a298cf83d72a61403e200eafccaa22"} Feb 27 17:20:33 crc kubenswrapper[4700]: I0227 17:20:33.734401 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-vfjlj" event={"ID":"b393dd07-e44c-499b-8a95-81c50f78efe8","Type":"ContainerStarted","Data":"4fd48e60f0e91b467a180771f5e56dc9978a31b0721f0d191a5d091c4259cc03"} Feb 27 17:20:33 crc kubenswrapper[4700]: I0227 17:20:33.743587 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-94be-account-create-update-snkzp" podStartSLOduration=1.7435661439999999 podStartE2EDuration="1.743566144s" podCreationTimestamp="2026-02-27 17:20:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:20:33.734202827 +0000 UTC m=+1193.719515614" watchObservedRunningTime="2026-02-27 17:20:33.743566144 +0000 UTC m=+1193.728878891" Feb 27 17:20:33 crc kubenswrapper[4700]: I0227 17:20:33.767537 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-4fb2-account-create-update-h494r" podStartSLOduration=2.767518338 podStartE2EDuration="2.767518338s" podCreationTimestamp="2026-02-27 17:20:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:20:33.760871542 +0000 UTC m=+1193.746184299" watchObservedRunningTime="2026-02-27 17:20:33.767518338 +0000 UTC m=+1193.752831085" Feb 27 17:20:33 crc kubenswrapper[4700]: I0227 17:20:33.778231 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-b838-account-create-update-g6qs7" podStartSLOduration=3.778216211 podStartE2EDuration="3.778216211s" podCreationTimestamp="2026-02-27 17:20:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:20:33.773308131 +0000 UTC m=+1193.758620888" watchObservedRunningTime="2026-02-27 17:20:33.778216211 +0000 UTC m=+1193.763528948" Feb 27 17:20:33 crc kubenswrapper[4700]: I0227 17:20:33.791207 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-glgwk" podStartSLOduration=2.7911935039999998 podStartE2EDuration="2.791193504s" podCreationTimestamp="2026-02-27 17:20:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:20:33.784955929 +0000 UTC m=+1193.770268676" watchObservedRunningTime="2026-02-27 17:20:33.791193504 +0000 UTC m=+1193.776506251" Feb 27 17:20:33 crc kubenswrapper[4700]: I0227 17:20:33.820058 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-vfjlj" podStartSLOduration=2.820037657 podStartE2EDuration="2.820037657s" podCreationTimestamp="2026-02-27 17:20:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:20:33.817836489 +0000 UTC m=+1193.803149236" watchObservedRunningTime="2026-02-27 17:20:33.820037657 +0000 UTC m=+1193.805350394" Feb 27 17:20:34 crc kubenswrapper[4700]: I0227 17:20:34.555176 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:34 crc kubenswrapper[4700]: I0227 17:20:34.555688 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:34 crc kubenswrapper[4700]: I0227 17:20:34.559395 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:34 crc kubenswrapper[4700]: I0227 17:20:34.745772 4700 generic.go:334] "Generic (PLEG): container finished" podID="0af77e17-d33d-4d3b-8246-e38499ea68f7" containerID="3e260f40afc484c8c52c1b94079e2bccc68fd3802ca10b847bcaf2e2dcaf288d" exitCode=0 Feb 27 17:20:34 crc kubenswrapper[4700]: I0227 17:20:34.745819 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-glgwk" event={"ID":"0af77e17-d33d-4d3b-8246-e38499ea68f7","Type":"ContainerDied","Data":"3e260f40afc484c8c52c1b94079e2bccc68fd3802ca10b847bcaf2e2dcaf288d"} Feb 27 17:20:34 crc kubenswrapper[4700]: I0227 17:20:34.747690 4700 generic.go:334] "Generic (PLEG): container finished" podID="a889ce63-746b-45bc-9f99-487f7416a80c" containerID="3c2c275c3cc206ac97d16d9dd3bc253b01082dc180beb82d9a0f98a67dae1c47" exitCode=0 Feb 27 17:20:34 crc kubenswrapper[4700]: I0227 17:20:34.747744 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b838-account-create-update-g6qs7" event={"ID":"a889ce63-746b-45bc-9f99-487f7416a80c","Type":"ContainerDied","Data":"3c2c275c3cc206ac97d16d9dd3bc253b01082dc180beb82d9a0f98a67dae1c47"} Feb 27 17:20:34 crc kubenswrapper[4700]: I0227 17:20:34.749341 4700 generic.go:334] "Generic (PLEG): container finished" podID="b393dd07-e44c-499b-8a95-81c50f78efe8" containerID="aa2b25a6011304a5a553000d459adb9829a298cf83d72a61403e200eafccaa22" exitCode=0 Feb 27 17:20:34 crc kubenswrapper[4700]: I0227 17:20:34.749424 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-vfjlj" event={"ID":"b393dd07-e44c-499b-8a95-81c50f78efe8","Type":"ContainerDied","Data":"aa2b25a6011304a5a553000d459adb9829a298cf83d72a61403e200eafccaa22"} Feb 27 17:20:34 crc kubenswrapper[4700]: I0227 17:20:34.751830 4700 generic.go:334] "Generic (PLEG): container finished" podID="89befb8e-a997-46a9-9db2-24f0dc5bf10d" containerID="efa7faa3a869c2e7843333aaff00f538d61d9b36d5342a1745b26e101297ad21" exitCode=0 Feb 27 17:20:34 crc kubenswrapper[4700]: I0227 17:20:34.751934 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-m7qsc" event={"ID":"89befb8e-a997-46a9-9db2-24f0dc5bf10d","Type":"ContainerDied","Data":"efa7faa3a869c2e7843333aaff00f538d61d9b36d5342a1745b26e101297ad21"} Feb 27 17:20:34 crc kubenswrapper[4700]: I0227 17:20:34.753250 4700 generic.go:334] "Generic (PLEG): container finished" podID="c9e7190d-50d9-4287-8f66-5aa298c5cdc7" containerID="7dbfbad6525b94cb92f5314afed9b77c4b093833511b704ca4881f88a6e32547" exitCode=0 Feb 27 17:20:34 crc kubenswrapper[4700]: I0227 17:20:34.753303 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-94be-account-create-update-snkzp" event={"ID":"c9e7190d-50d9-4287-8f66-5aa298c5cdc7","Type":"ContainerDied","Data":"7dbfbad6525b94cb92f5314afed9b77c4b093833511b704ca4881f88a6e32547"} Feb 27 17:20:34 crc kubenswrapper[4700]: I0227 17:20:34.754611 4700 generic.go:334] "Generic (PLEG): container finished" podID="c7812092-9d78-409a-80f8-4a4db92af41e" containerID="c513a087e1e918ce5827b7e53435b261f731d6a8770be071707b9a006ffa3bd9" exitCode=0 Feb 27 17:20:34 crc kubenswrapper[4700]: I0227 17:20:34.754667 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-4fb2-account-create-update-h494r" event={"ID":"c7812092-9d78-409a-80f8-4a4db92af41e","Type":"ContainerDied","Data":"c513a087e1e918ce5827b7e53435b261f731d6a8770be071707b9a006ffa3bd9"} Feb 27 17:20:34 crc kubenswrapper[4700]: I0227 17:20:34.756302 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:35 crc kubenswrapper[4700]: I0227 17:20:35.237223 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-2ppp5" Feb 27 17:20:35 crc kubenswrapper[4700]: I0227 17:20:35.407039 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/42f288bd-13a4-4e0c-a616-b09462acd2bf-operator-scripts\") pod \"42f288bd-13a4-4e0c-a616-b09462acd2bf\" (UID: \"42f288bd-13a4-4e0c-a616-b09462acd2bf\") " Feb 27 17:20:35 crc kubenswrapper[4700]: I0227 17:20:35.407087 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srlx5\" (UniqueName: \"kubernetes.io/projected/42f288bd-13a4-4e0c-a616-b09462acd2bf-kube-api-access-srlx5\") pod \"42f288bd-13a4-4e0c-a616-b09462acd2bf\" (UID: \"42f288bd-13a4-4e0c-a616-b09462acd2bf\") " Feb 27 17:20:35 crc kubenswrapper[4700]: I0227 17:20:35.407791 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/42f288bd-13a4-4e0c-a616-b09462acd2bf-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "42f288bd-13a4-4e0c-a616-b09462acd2bf" (UID: "42f288bd-13a4-4e0c-a616-b09462acd2bf"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:20:35 crc kubenswrapper[4700]: I0227 17:20:35.408274 4700 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/42f288bd-13a4-4e0c-a616-b09462acd2bf-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:35 crc kubenswrapper[4700]: I0227 17:20:35.415320 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42f288bd-13a4-4e0c-a616-b09462acd2bf-kube-api-access-srlx5" (OuterVolumeSpecName: "kube-api-access-srlx5") pod "42f288bd-13a4-4e0c-a616-b09462acd2bf" (UID: "42f288bd-13a4-4e0c-a616-b09462acd2bf"). InnerVolumeSpecName "kube-api-access-srlx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:20:35 crc kubenswrapper[4700]: I0227 17:20:35.509322 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srlx5\" (UniqueName: \"kubernetes.io/projected/42f288bd-13a4-4e0c-a616-b09462acd2bf-kube-api-access-srlx5\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:35 crc kubenswrapper[4700]: I0227 17:20:35.776707 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-2ppp5" Feb 27 17:20:35 crc kubenswrapper[4700]: I0227 17:20:35.777710 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-2ppp5" event={"ID":"42f288bd-13a4-4e0c-a616-b09462acd2bf","Type":"ContainerDied","Data":"3f8888338909c2f754cdddb7f35d2f58422df2d973a3e4b5982af39ba84d059c"} Feb 27 17:20:35 crc kubenswrapper[4700]: I0227 17:20:35.777769 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f8888338909c2f754cdddb7f35d2f58422df2d973a3e4b5982af39ba84d059c" Feb 27 17:20:35 crc kubenswrapper[4700]: I0227 17:20:35.990979 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-qqqm8" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.006267 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-qqqm8" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.040801 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-sfldb" podUID="a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a" containerName="ovn-controller" probeResult="failure" output=< Feb 27 17:20:36 crc kubenswrapper[4700]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Feb 27 17:20:36 crc kubenswrapper[4700]: > Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.189374 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b838-account-create-update-g6qs7" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.219397 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-sfldb-config-mss7g"] Feb 27 17:20:36 crc kubenswrapper[4700]: E0227 17:20:36.219833 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a889ce63-746b-45bc-9f99-487f7416a80c" containerName="mariadb-account-create-update" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.219853 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="a889ce63-746b-45bc-9f99-487f7416a80c" containerName="mariadb-account-create-update" Feb 27 17:20:36 crc kubenswrapper[4700]: E0227 17:20:36.219877 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a28896eb-9f45-43fa-a078-8ae9175cb25d" containerName="mariadb-account-create-update" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.219884 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="a28896eb-9f45-43fa-a078-8ae9175cb25d" containerName="mariadb-account-create-update" Feb 27 17:20:36 crc kubenswrapper[4700]: E0227 17:20:36.219893 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42f288bd-13a4-4e0c-a616-b09462acd2bf" containerName="mariadb-database-create" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.220162 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="42f288bd-13a4-4e0c-a616-b09462acd2bf" containerName="mariadb-database-create" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.220355 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="a889ce63-746b-45bc-9f99-487f7416a80c" containerName="mariadb-account-create-update" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.220385 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="42f288bd-13a4-4e0c-a616-b09462acd2bf" containerName="mariadb-database-create" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.220395 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="a28896eb-9f45-43fa-a078-8ae9175cb25d" containerName="mariadb-account-create-update" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.221027 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-sfldb-config-mss7g" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.224221 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.234919 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-sfldb-config-mss7g"] Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.326558 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a889ce63-746b-45bc-9f99-487f7416a80c-operator-scripts\") pod \"a889ce63-746b-45bc-9f99-487f7416a80c\" (UID: \"a889ce63-746b-45bc-9f99-487f7416a80c\") " Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.326606 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6dhq\" (UniqueName: \"kubernetes.io/projected/a889ce63-746b-45bc-9f99-487f7416a80c-kube-api-access-t6dhq\") pod \"a889ce63-746b-45bc-9f99-487f7416a80c\" (UID: \"a889ce63-746b-45bc-9f99-487f7416a80c\") " Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.326824 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b032049e-5ad9-44a8-9520-fd7c1699748e-var-log-ovn\") pod \"ovn-controller-sfldb-config-mss7g\" (UID: \"b032049e-5ad9-44a8-9520-fd7c1699748e\") " pod="openstack/ovn-controller-sfldb-config-mss7g" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.326899 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b032049e-5ad9-44a8-9520-fd7c1699748e-additional-scripts\") pod \"ovn-controller-sfldb-config-mss7g\" (UID: \"b032049e-5ad9-44a8-9520-fd7c1699748e\") " pod="openstack/ovn-controller-sfldb-config-mss7g" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.326932 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b032049e-5ad9-44a8-9520-fd7c1699748e-scripts\") pod \"ovn-controller-sfldb-config-mss7g\" (UID: \"b032049e-5ad9-44a8-9520-fd7c1699748e\") " pod="openstack/ovn-controller-sfldb-config-mss7g" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.326955 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b032049e-5ad9-44a8-9520-fd7c1699748e-var-run-ovn\") pod \"ovn-controller-sfldb-config-mss7g\" (UID: \"b032049e-5ad9-44a8-9520-fd7c1699748e\") " pod="openstack/ovn-controller-sfldb-config-mss7g" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.326987 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pdbj\" (UniqueName: \"kubernetes.io/projected/b032049e-5ad9-44a8-9520-fd7c1699748e-kube-api-access-4pdbj\") pod \"ovn-controller-sfldb-config-mss7g\" (UID: \"b032049e-5ad9-44a8-9520-fd7c1699748e\") " pod="openstack/ovn-controller-sfldb-config-mss7g" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.327041 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b032049e-5ad9-44a8-9520-fd7c1699748e-var-run\") pod \"ovn-controller-sfldb-config-mss7g\" (UID: \"b032049e-5ad9-44a8-9520-fd7c1699748e\") " pod="openstack/ovn-controller-sfldb-config-mss7g" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.327952 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a889ce63-746b-45bc-9f99-487f7416a80c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a889ce63-746b-45bc-9f99-487f7416a80c" (UID: "a889ce63-746b-45bc-9f99-487f7416a80c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.332263 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a889ce63-746b-45bc-9f99-487f7416a80c-kube-api-access-t6dhq" (OuterVolumeSpecName: "kube-api-access-t6dhq") pod "a889ce63-746b-45bc-9f99-487f7416a80c" (UID: "a889ce63-746b-45bc-9f99-487f7416a80c"). InnerVolumeSpecName "kube-api-access-t6dhq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.412084 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.412127 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.428319 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b032049e-5ad9-44a8-9520-fd7c1699748e-var-run\") pod \"ovn-controller-sfldb-config-mss7g\" (UID: \"b032049e-5ad9-44a8-9520-fd7c1699748e\") " pod="openstack/ovn-controller-sfldb-config-mss7g" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.428365 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b032049e-5ad9-44a8-9520-fd7c1699748e-var-log-ovn\") pod \"ovn-controller-sfldb-config-mss7g\" (UID: \"b032049e-5ad9-44a8-9520-fd7c1699748e\") " pod="openstack/ovn-controller-sfldb-config-mss7g" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.428432 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b032049e-5ad9-44a8-9520-fd7c1699748e-additional-scripts\") pod \"ovn-controller-sfldb-config-mss7g\" (UID: \"b032049e-5ad9-44a8-9520-fd7c1699748e\") " pod="openstack/ovn-controller-sfldb-config-mss7g" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.428467 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b032049e-5ad9-44a8-9520-fd7c1699748e-scripts\") pod \"ovn-controller-sfldb-config-mss7g\" (UID: \"b032049e-5ad9-44a8-9520-fd7c1699748e\") " pod="openstack/ovn-controller-sfldb-config-mss7g" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.428513 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b032049e-5ad9-44a8-9520-fd7c1699748e-var-run-ovn\") pod \"ovn-controller-sfldb-config-mss7g\" (UID: \"b032049e-5ad9-44a8-9520-fd7c1699748e\") " pod="openstack/ovn-controller-sfldb-config-mss7g" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.428547 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pdbj\" (UniqueName: \"kubernetes.io/projected/b032049e-5ad9-44a8-9520-fd7c1699748e-kube-api-access-4pdbj\") pod \"ovn-controller-sfldb-config-mss7g\" (UID: \"b032049e-5ad9-44a8-9520-fd7c1699748e\") " pod="openstack/ovn-controller-sfldb-config-mss7g" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.428603 4700 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a889ce63-746b-45bc-9f99-487f7416a80c-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.428615 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6dhq\" (UniqueName: \"kubernetes.io/projected/a889ce63-746b-45bc-9f99-487f7416a80c-kube-api-access-t6dhq\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.428628 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b032049e-5ad9-44a8-9520-fd7c1699748e-var-run\") pod \"ovn-controller-sfldb-config-mss7g\" (UID: \"b032049e-5ad9-44a8-9520-fd7c1699748e\") " pod="openstack/ovn-controller-sfldb-config-mss7g" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.428904 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b032049e-5ad9-44a8-9520-fd7c1699748e-var-run-ovn\") pod \"ovn-controller-sfldb-config-mss7g\" (UID: \"b032049e-5ad9-44a8-9520-fd7c1699748e\") " pod="openstack/ovn-controller-sfldb-config-mss7g" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.428952 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b032049e-5ad9-44a8-9520-fd7c1699748e-var-log-ovn\") pod \"ovn-controller-sfldb-config-mss7g\" (UID: \"b032049e-5ad9-44a8-9520-fd7c1699748e\") " pod="openstack/ovn-controller-sfldb-config-mss7g" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.429220 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b032049e-5ad9-44a8-9520-fd7c1699748e-additional-scripts\") pod \"ovn-controller-sfldb-config-mss7g\" (UID: \"b032049e-5ad9-44a8-9520-fd7c1699748e\") " pod="openstack/ovn-controller-sfldb-config-mss7g" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.431099 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b032049e-5ad9-44a8-9520-fd7c1699748e-scripts\") pod \"ovn-controller-sfldb-config-mss7g\" (UID: \"b032049e-5ad9-44a8-9520-fd7c1699748e\") " pod="openstack/ovn-controller-sfldb-config-mss7g" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.452094 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pdbj\" (UniqueName: \"kubernetes.io/projected/b032049e-5ad9-44a8-9520-fd7c1699748e-kube-api-access-4pdbj\") pod \"ovn-controller-sfldb-config-mss7g\" (UID: \"b032049e-5ad9-44a8-9520-fd7c1699748e\") " pod="openstack/ovn-controller-sfldb-config-mss7g" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.514139 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-glgwk" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.515121 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-94be-account-create-update-snkzp" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.520623 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-m7qsc" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.536753 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-sfldb-config-mss7g" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.544341 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-vfjlj" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.568825 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-4fb2-account-create-update-h494r" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.633153 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0af77e17-d33d-4d3b-8246-e38499ea68f7-operator-scripts\") pod \"0af77e17-d33d-4d3b-8246-e38499ea68f7\" (UID: \"0af77e17-d33d-4d3b-8246-e38499ea68f7\") " Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.633201 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2rbmz\" (UniqueName: \"kubernetes.io/projected/c7812092-9d78-409a-80f8-4a4db92af41e-kube-api-access-2rbmz\") pod \"c7812092-9d78-409a-80f8-4a4db92af41e\" (UID: \"c7812092-9d78-409a-80f8-4a4db92af41e\") " Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.633229 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b393dd07-e44c-499b-8a95-81c50f78efe8-operator-scripts\") pod \"b393dd07-e44c-499b-8a95-81c50f78efe8\" (UID: \"b393dd07-e44c-499b-8a95-81c50f78efe8\") " Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.633247 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgw9s\" (UniqueName: \"kubernetes.io/projected/0af77e17-d33d-4d3b-8246-e38499ea68f7-kube-api-access-mgw9s\") pod \"0af77e17-d33d-4d3b-8246-e38499ea68f7\" (UID: \"0af77e17-d33d-4d3b-8246-e38499ea68f7\") " Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.633264 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnggv\" (UniqueName: \"kubernetes.io/projected/c9e7190d-50d9-4287-8f66-5aa298c5cdc7-kube-api-access-qnggv\") pod \"c9e7190d-50d9-4287-8f66-5aa298c5cdc7\" (UID: \"c9e7190d-50d9-4287-8f66-5aa298c5cdc7\") " Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.633351 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npln6\" (UniqueName: \"kubernetes.io/projected/89befb8e-a997-46a9-9db2-24f0dc5bf10d-kube-api-access-npln6\") pod \"89befb8e-a997-46a9-9db2-24f0dc5bf10d\" (UID: \"89befb8e-a997-46a9-9db2-24f0dc5bf10d\") " Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.633367 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c9e7190d-50d9-4287-8f66-5aa298c5cdc7-operator-scripts\") pod \"c9e7190d-50d9-4287-8f66-5aa298c5cdc7\" (UID: \"c9e7190d-50d9-4287-8f66-5aa298c5cdc7\") " Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.633393 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cw5rr\" (UniqueName: \"kubernetes.io/projected/b393dd07-e44c-499b-8a95-81c50f78efe8-kube-api-access-cw5rr\") pod \"b393dd07-e44c-499b-8a95-81c50f78efe8\" (UID: \"b393dd07-e44c-499b-8a95-81c50f78efe8\") " Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.633413 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89befb8e-a997-46a9-9db2-24f0dc5bf10d-combined-ca-bundle\") pod \"89befb8e-a997-46a9-9db2-24f0dc5bf10d\" (UID: \"89befb8e-a997-46a9-9db2-24f0dc5bf10d\") " Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.633463 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/89befb8e-a997-46a9-9db2-24f0dc5bf10d-swiftconf\") pod \"89befb8e-a997-46a9-9db2-24f0dc5bf10d\" (UID: \"89befb8e-a997-46a9-9db2-24f0dc5bf10d\") " Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.633492 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89befb8e-a997-46a9-9db2-24f0dc5bf10d-scripts\") pod \"89befb8e-a997-46a9-9db2-24f0dc5bf10d\" (UID: \"89befb8e-a997-46a9-9db2-24f0dc5bf10d\") " Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.633511 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/89befb8e-a997-46a9-9db2-24f0dc5bf10d-etc-swift\") pod \"89befb8e-a997-46a9-9db2-24f0dc5bf10d\" (UID: \"89befb8e-a997-46a9-9db2-24f0dc5bf10d\") " Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.633526 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7812092-9d78-409a-80f8-4a4db92af41e-operator-scripts\") pod \"c7812092-9d78-409a-80f8-4a4db92af41e\" (UID: \"c7812092-9d78-409a-80f8-4a4db92af41e\") " Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.633553 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/89befb8e-a997-46a9-9db2-24f0dc5bf10d-ring-data-devices\") pod \"89befb8e-a997-46a9-9db2-24f0dc5bf10d\" (UID: \"89befb8e-a997-46a9-9db2-24f0dc5bf10d\") " Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.633573 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/89befb8e-a997-46a9-9db2-24f0dc5bf10d-dispersionconf\") pod \"89befb8e-a997-46a9-9db2-24f0dc5bf10d\" (UID: \"89befb8e-a997-46a9-9db2-24f0dc5bf10d\") " Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.635065 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9e7190d-50d9-4287-8f66-5aa298c5cdc7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c9e7190d-50d9-4287-8f66-5aa298c5cdc7" (UID: "c9e7190d-50d9-4287-8f66-5aa298c5cdc7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.635231 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7812092-9d78-409a-80f8-4a4db92af41e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c7812092-9d78-409a-80f8-4a4db92af41e" (UID: "c7812092-9d78-409a-80f8-4a4db92af41e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.633631 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0af77e17-d33d-4d3b-8246-e38499ea68f7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0af77e17-d33d-4d3b-8246-e38499ea68f7" (UID: "0af77e17-d33d-4d3b-8246-e38499ea68f7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.635386 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89befb8e-a997-46a9-9db2-24f0dc5bf10d-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "89befb8e-a997-46a9-9db2-24f0dc5bf10d" (UID: "89befb8e-a997-46a9-9db2-24f0dc5bf10d"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.635978 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b393dd07-e44c-499b-8a95-81c50f78efe8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b393dd07-e44c-499b-8a95-81c50f78efe8" (UID: "b393dd07-e44c-499b-8a95-81c50f78efe8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.636147 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89befb8e-a997-46a9-9db2-24f0dc5bf10d-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "89befb8e-a997-46a9-9db2-24f0dc5bf10d" (UID: "89befb8e-a997-46a9-9db2-24f0dc5bf10d"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.638404 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89befb8e-a997-46a9-9db2-24f0dc5bf10d-kube-api-access-npln6" (OuterVolumeSpecName: "kube-api-access-npln6") pod "89befb8e-a997-46a9-9db2-24f0dc5bf10d" (UID: "89befb8e-a997-46a9-9db2-24f0dc5bf10d"). InnerVolumeSpecName "kube-api-access-npln6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.642502 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7812092-9d78-409a-80f8-4a4db92af41e-kube-api-access-2rbmz" (OuterVolumeSpecName: "kube-api-access-2rbmz") pod "c7812092-9d78-409a-80f8-4a4db92af41e" (UID: "c7812092-9d78-409a-80f8-4a4db92af41e"). InnerVolumeSpecName "kube-api-access-2rbmz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.670448 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0af77e17-d33d-4d3b-8246-e38499ea68f7-kube-api-access-mgw9s" (OuterVolumeSpecName: "kube-api-access-mgw9s") pod "0af77e17-d33d-4d3b-8246-e38499ea68f7" (UID: "0af77e17-d33d-4d3b-8246-e38499ea68f7"). InnerVolumeSpecName "kube-api-access-mgw9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.671804 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9e7190d-50d9-4287-8f66-5aa298c5cdc7-kube-api-access-qnggv" (OuterVolumeSpecName: "kube-api-access-qnggv") pod "c9e7190d-50d9-4287-8f66-5aa298c5cdc7" (UID: "c9e7190d-50d9-4287-8f66-5aa298c5cdc7"). InnerVolumeSpecName "kube-api-access-qnggv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.672708 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b393dd07-e44c-499b-8a95-81c50f78efe8-kube-api-access-cw5rr" (OuterVolumeSpecName: "kube-api-access-cw5rr") pod "b393dd07-e44c-499b-8a95-81c50f78efe8" (UID: "b393dd07-e44c-499b-8a95-81c50f78efe8"). InnerVolumeSpecName "kube-api-access-cw5rr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.683944 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89befb8e-a997-46a9-9db2-24f0dc5bf10d-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "89befb8e-a997-46a9-9db2-24f0dc5bf10d" (UID: "89befb8e-a997-46a9-9db2-24f0dc5bf10d"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.695504 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89befb8e-a997-46a9-9db2-24f0dc5bf10d-scripts" (OuterVolumeSpecName: "scripts") pod "89befb8e-a997-46a9-9db2-24f0dc5bf10d" (UID: "89befb8e-a997-46a9-9db2-24f0dc5bf10d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.700089 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89befb8e-a997-46a9-9db2-24f0dc5bf10d-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "89befb8e-a997-46a9-9db2-24f0dc5bf10d" (UID: "89befb8e-a997-46a9-9db2-24f0dc5bf10d"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.704176 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89befb8e-a997-46a9-9db2-24f0dc5bf10d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "89befb8e-a997-46a9-9db2-24f0dc5bf10d" (UID: "89befb8e-a997-46a9-9db2-24f0dc5bf10d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.735339 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89befb8e-a997-46a9-9db2-24f0dc5bf10d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.735396 4700 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/89befb8e-a997-46a9-9db2-24f0dc5bf10d-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.735405 4700 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89befb8e-a997-46a9-9db2-24f0dc5bf10d-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.735419 4700 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/89befb8e-a997-46a9-9db2-24f0dc5bf10d-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.735428 4700 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7812092-9d78-409a-80f8-4a4db92af41e-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.735437 4700 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/89befb8e-a997-46a9-9db2-24f0dc5bf10d-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.735445 4700 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/89befb8e-a997-46a9-9db2-24f0dc5bf10d-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.735454 4700 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0af77e17-d33d-4d3b-8246-e38499ea68f7-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.735476 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2rbmz\" (UniqueName: \"kubernetes.io/projected/c7812092-9d78-409a-80f8-4a4db92af41e-kube-api-access-2rbmz\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.735487 4700 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b393dd07-e44c-499b-8a95-81c50f78efe8-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.735495 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgw9s\" (UniqueName: \"kubernetes.io/projected/0af77e17-d33d-4d3b-8246-e38499ea68f7-kube-api-access-mgw9s\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.735505 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnggv\" (UniqueName: \"kubernetes.io/projected/c9e7190d-50d9-4287-8f66-5aa298c5cdc7-kube-api-access-qnggv\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.735513 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npln6\" (UniqueName: \"kubernetes.io/projected/89befb8e-a997-46a9-9db2-24f0dc5bf10d-kube-api-access-npln6\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.735521 4700 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c9e7190d-50d9-4287-8f66-5aa298c5cdc7-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.735530 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cw5rr\" (UniqueName: \"kubernetes.io/projected/b393dd07-e44c-499b-8a95-81c50f78efe8-kube-api-access-cw5rr\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.788519 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-glgwk" event={"ID":"0af77e17-d33d-4d3b-8246-e38499ea68f7","Type":"ContainerDied","Data":"c66482b857f128a7ab062a40ca6b2419764ca4fbb3313d1ed4cf5841b7ec3931"} Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.788550 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c66482b857f128a7ab062a40ca6b2419764ca4fbb3313d1ed4cf5841b7ec3931" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.788578 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-glgwk" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.790954 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b838-account-create-update-g6qs7" event={"ID":"a889ce63-746b-45bc-9f99-487f7416a80c","Type":"ContainerDied","Data":"980f7ce9773359e64d0ccb38ff3584b95df7ad7a14366f4b73f55bdda33ad6c7"} Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.790978 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="980f7ce9773359e64d0ccb38ff3584b95df7ad7a14366f4b73f55bdda33ad6c7" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.791032 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b838-account-create-update-g6qs7" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.798404 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-vfjlj" event={"ID":"b393dd07-e44c-499b-8a95-81c50f78efe8","Type":"ContainerDied","Data":"4fd48e60f0e91b467a180771f5e56dc9978a31b0721f0d191a5d091c4259cc03"} Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.798435 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4fd48e60f0e91b467a180771f5e56dc9978a31b0721f0d191a5d091c4259cc03" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.798496 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-vfjlj" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.801540 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-m7qsc" event={"ID":"89befb8e-a997-46a9-9db2-24f0dc5bf10d","Type":"ContainerDied","Data":"e201d4853198997660f8cf9b9866a504737185cd637f0fc081d2984aa5794595"} Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.801575 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e201d4853198997660f8cf9b9866a504737185cd637f0fc081d2984aa5794595" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.801629 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-m7qsc" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.803763 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-94be-account-create-update-snkzp" event={"ID":"c9e7190d-50d9-4287-8f66-5aa298c5cdc7","Type":"ContainerDied","Data":"c004aa732bdd1dfcdc4c2449d24f9074f7eb249df8bab48162677d5d48bc01c2"} Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.803792 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c004aa732bdd1dfcdc4c2449d24f9074f7eb249df8bab48162677d5d48bc01c2" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.803837 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-94be-account-create-update-snkzp" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.806523 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-4fb2-account-create-update-h494r" event={"ID":"c7812092-9d78-409a-80f8-4a4db92af41e","Type":"ContainerDied","Data":"91258d57272de973fe0bd80fca9da13739e5bff01d7851b9389d62bf5231a782"} Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.806547 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="91258d57272de973fe0bd80fca9da13739e5bff01d7851b9389d62bf5231a782" Feb 27 17:20:36 crc kubenswrapper[4700]: I0227 17:20:36.806552 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-4fb2-account-create-update-h494r" Feb 27 17:20:37 crc kubenswrapper[4700]: I0227 17:20:37.033831 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-sfldb-config-mss7g"] Feb 27 17:20:37 crc kubenswrapper[4700]: I0227 17:20:37.445013 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Feb 27 17:20:37 crc kubenswrapper[4700]: I0227 17:20:37.816060 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-sfldb-config-mss7g" event={"ID":"b032049e-5ad9-44a8-9520-fd7c1699748e","Type":"ContainerStarted","Data":"546d803ac65c71659c694790207913d1286cb34b93d6768a646ed07444bf7fca"} Feb 27 17:20:37 crc kubenswrapper[4700]: I0227 17:20:37.816118 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-sfldb-config-mss7g" event={"ID":"b032049e-5ad9-44a8-9520-fd7c1699748e","Type":"ContainerStarted","Data":"8872c8e1736cd2b9c46cf4ecf1920e8b88798c3eec1c73b337a90d4580f25d09"} Feb 27 17:20:37 crc kubenswrapper[4700]: I0227 17:20:37.816238 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="9721431e-046b-4b3b-861e-22a65396981b" containerName="prometheus" containerID="cri-o://7b9277446cc5fc2593bdc3b3cb6e1b7c283d1c39909d8b7b4ad6eef57493d935" gracePeriod=600 Feb 27 17:20:37 crc kubenswrapper[4700]: I0227 17:20:37.816260 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="9721431e-046b-4b3b-861e-22a65396981b" containerName="thanos-sidecar" containerID="cri-o://99478540ee16a5f122a1e6e28f39329c9bd265403a641ecd3a881d8272258aee" gracePeriod=600 Feb 27 17:20:37 crc kubenswrapper[4700]: I0227 17:20:37.816751 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="9721431e-046b-4b3b-861e-22a65396981b" containerName="config-reloader" containerID="cri-o://4016795d07ac4553ff298150bb0cf07997d9258aee1bdce657c1c5422071402b" gracePeriod=600 Feb 27 17:20:37 crc kubenswrapper[4700]: I0227 17:20:37.865263 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-sfldb-config-mss7g" podStartSLOduration=1.8652477410000001 podStartE2EDuration="1.865247741s" podCreationTimestamp="2026-02-27 17:20:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:20:37.860313221 +0000 UTC m=+1197.845625988" watchObservedRunningTime="2026-02-27 17:20:37.865247741 +0000 UTC m=+1197.850560488" Feb 27 17:20:38 crc kubenswrapper[4700]: E0227 17:20:38.131837 4700 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb032049e_5ad9_44a8_9520_fd7c1699748e.slice/crio-conmon-546d803ac65c71659c694790207913d1286cb34b93d6768a646ed07444bf7fca.scope\": RecentStats: unable to find data in memory cache]" Feb 27 17:20:38 crc kubenswrapper[4700]: I0227 17:20:38.831960 4700 generic.go:334] "Generic (PLEG): container finished" podID="b032049e-5ad9-44a8-9520-fd7c1699748e" containerID="546d803ac65c71659c694790207913d1286cb34b93d6768a646ed07444bf7fca" exitCode=0 Feb 27 17:20:38 crc kubenswrapper[4700]: I0227 17:20:38.832028 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-sfldb-config-mss7g" event={"ID":"b032049e-5ad9-44a8-9520-fd7c1699748e","Type":"ContainerDied","Data":"546d803ac65c71659c694790207913d1286cb34b93d6768a646ed07444bf7fca"} Feb 27 17:20:38 crc kubenswrapper[4700]: I0227 17:20:38.837891 4700 generic.go:334] "Generic (PLEG): container finished" podID="9721431e-046b-4b3b-861e-22a65396981b" containerID="99478540ee16a5f122a1e6e28f39329c9bd265403a641ecd3a881d8272258aee" exitCode=0 Feb 27 17:20:38 crc kubenswrapper[4700]: I0227 17:20:38.837924 4700 generic.go:334] "Generic (PLEG): container finished" podID="9721431e-046b-4b3b-861e-22a65396981b" containerID="4016795d07ac4553ff298150bb0cf07997d9258aee1bdce657c1c5422071402b" exitCode=0 Feb 27 17:20:38 crc kubenswrapper[4700]: I0227 17:20:38.837931 4700 generic.go:334] "Generic (PLEG): container finished" podID="9721431e-046b-4b3b-861e-22a65396981b" containerID="7b9277446cc5fc2593bdc3b3cb6e1b7c283d1c39909d8b7b4ad6eef57493d935" exitCode=0 Feb 27 17:20:38 crc kubenswrapper[4700]: I0227 17:20:38.837951 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9721431e-046b-4b3b-861e-22a65396981b","Type":"ContainerDied","Data":"99478540ee16a5f122a1e6e28f39329c9bd265403a641ecd3a881d8272258aee"} Feb 27 17:20:38 crc kubenswrapper[4700]: I0227 17:20:38.837977 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9721431e-046b-4b3b-861e-22a65396981b","Type":"ContainerDied","Data":"4016795d07ac4553ff298150bb0cf07997d9258aee1bdce657c1c5422071402b"} Feb 27 17:20:38 crc kubenswrapper[4700]: I0227 17:20:38.837988 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9721431e-046b-4b3b-861e-22a65396981b","Type":"ContainerDied","Data":"7b9277446cc5fc2593bdc3b3cb6e1b7c283d1c39909d8b7b4ad6eef57493d935"} Feb 27 17:20:38 crc kubenswrapper[4700]: I0227 17:20:38.964337 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.003792 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-rtn6x"] Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.004332 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-rtn6x"] Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.062684 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-27j58"] Feb 27 17:20:39 crc kubenswrapper[4700]: E0227 17:20:39.063047 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9721431e-046b-4b3b-861e-22a65396981b" containerName="prometheus" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.063064 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="9721431e-046b-4b3b-861e-22a65396981b" containerName="prometheus" Feb 27 17:20:39 crc kubenswrapper[4700]: E0227 17:20:39.063076 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9721431e-046b-4b3b-861e-22a65396981b" containerName="thanos-sidecar" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.063083 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="9721431e-046b-4b3b-861e-22a65396981b" containerName="thanos-sidecar" Feb 27 17:20:39 crc kubenswrapper[4700]: E0227 17:20:39.063093 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7812092-9d78-409a-80f8-4a4db92af41e" containerName="mariadb-account-create-update" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.063100 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7812092-9d78-409a-80f8-4a4db92af41e" containerName="mariadb-account-create-update" Feb 27 17:20:39 crc kubenswrapper[4700]: E0227 17:20:39.063110 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0af77e17-d33d-4d3b-8246-e38499ea68f7" containerName="mariadb-database-create" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.063116 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="0af77e17-d33d-4d3b-8246-e38499ea68f7" containerName="mariadb-database-create" Feb 27 17:20:39 crc kubenswrapper[4700]: E0227 17:20:39.063129 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9721431e-046b-4b3b-861e-22a65396981b" containerName="config-reloader" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.063137 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="9721431e-046b-4b3b-861e-22a65396981b" containerName="config-reloader" Feb 27 17:20:39 crc kubenswrapper[4700]: E0227 17:20:39.063161 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89befb8e-a997-46a9-9db2-24f0dc5bf10d" containerName="swift-ring-rebalance" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.063168 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="89befb8e-a997-46a9-9db2-24f0dc5bf10d" containerName="swift-ring-rebalance" Feb 27 17:20:39 crc kubenswrapper[4700]: E0227 17:20:39.063174 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9e7190d-50d9-4287-8f66-5aa298c5cdc7" containerName="mariadb-account-create-update" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.063180 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9e7190d-50d9-4287-8f66-5aa298c5cdc7" containerName="mariadb-account-create-update" Feb 27 17:20:39 crc kubenswrapper[4700]: E0227 17:20:39.063200 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b393dd07-e44c-499b-8a95-81c50f78efe8" containerName="mariadb-database-create" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.063206 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="b393dd07-e44c-499b-8a95-81c50f78efe8" containerName="mariadb-database-create" Feb 27 17:20:39 crc kubenswrapper[4700]: E0227 17:20:39.063216 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9721431e-046b-4b3b-861e-22a65396981b" containerName="init-config-reloader" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.063222 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="9721431e-046b-4b3b-861e-22a65396981b" containerName="init-config-reloader" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.063371 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="89befb8e-a997-46a9-9db2-24f0dc5bf10d" containerName="swift-ring-rebalance" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.063383 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="9721431e-046b-4b3b-861e-22a65396981b" containerName="thanos-sidecar" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.063393 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7812092-9d78-409a-80f8-4a4db92af41e" containerName="mariadb-account-create-update" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.063404 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="9721431e-046b-4b3b-861e-22a65396981b" containerName="prometheus" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.063411 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="b393dd07-e44c-499b-8a95-81c50f78efe8" containerName="mariadb-database-create" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.063419 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="0af77e17-d33d-4d3b-8246-e38499ea68f7" containerName="mariadb-database-create" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.063430 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="9721431e-046b-4b3b-861e-22a65396981b" containerName="config-reloader" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.063440 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9e7190d-50d9-4287-8f66-5aa298c5cdc7" containerName="mariadb-account-create-update" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.063982 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-27j58" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.065981 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.070243 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-27j58"] Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.076380 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4vhth\" (UniqueName: \"kubernetes.io/projected/9721431e-046b-4b3b-861e-22a65396981b-kube-api-access-4vhth\") pod \"9721431e-046b-4b3b-861e-22a65396981b\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.076433 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/9721431e-046b-4b3b-861e-22a65396981b-prometheus-metric-storage-rulefiles-2\") pod \"9721431e-046b-4b3b-861e-22a65396981b\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.076458 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/9721431e-046b-4b3b-861e-22a65396981b-prometheus-metric-storage-rulefiles-0\") pod \"9721431e-046b-4b3b-861e-22a65396981b\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.076565 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9721431e-046b-4b3b-861e-22a65396981b-tls-assets\") pod \"9721431e-046b-4b3b-861e-22a65396981b\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.076642 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9721431e-046b-4b3b-861e-22a65396981b-config-out\") pod \"9721431e-046b-4b3b-861e-22a65396981b\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.076662 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/9721431e-046b-4b3b-861e-22a65396981b-thanos-prometheus-http-client-file\") pod \"9721431e-046b-4b3b-861e-22a65396981b\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.076679 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9721431e-046b-4b3b-861e-22a65396981b-web-config\") pod \"9721431e-046b-4b3b-861e-22a65396981b\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.076726 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/9721431e-046b-4b3b-861e-22a65396981b-prometheus-metric-storage-rulefiles-1\") pod \"9721431e-046b-4b3b-861e-22a65396981b\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.076925 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2\") pod \"9721431e-046b-4b3b-861e-22a65396981b\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.076960 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9721431e-046b-4b3b-861e-22a65396981b-config\") pod \"9721431e-046b-4b3b-861e-22a65396981b\" (UID: \"9721431e-046b-4b3b-861e-22a65396981b\") " Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.081522 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9721431e-046b-4b3b-861e-22a65396981b-prometheus-metric-storage-rulefiles-1" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-1") pod "9721431e-046b-4b3b-861e-22a65396981b" (UID: "9721431e-046b-4b3b-861e-22a65396981b"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.082404 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9721431e-046b-4b3b-861e-22a65396981b-prometheus-metric-storage-rulefiles-2" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-2") pod "9721431e-046b-4b3b-861e-22a65396981b" (UID: "9721431e-046b-4b3b-861e-22a65396981b"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-2". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.082692 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9721431e-046b-4b3b-861e-22a65396981b-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "9721431e-046b-4b3b-861e-22a65396981b" (UID: "9721431e-046b-4b3b-861e-22a65396981b"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.083874 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9721431e-046b-4b3b-861e-22a65396981b-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "9721431e-046b-4b3b-861e-22a65396981b" (UID: "9721431e-046b-4b3b-861e-22a65396981b"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.083920 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9721431e-046b-4b3b-861e-22a65396981b-config-out" (OuterVolumeSpecName: "config-out") pod "9721431e-046b-4b3b-861e-22a65396981b" (UID: "9721431e-046b-4b3b-861e-22a65396981b"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.091592 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9721431e-046b-4b3b-861e-22a65396981b-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "9721431e-046b-4b3b-861e-22a65396981b" (UID: "9721431e-046b-4b3b-861e-22a65396981b"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.091930 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9721431e-046b-4b3b-861e-22a65396981b-kube-api-access-4vhth" (OuterVolumeSpecName: "kube-api-access-4vhth") pod "9721431e-046b-4b3b-861e-22a65396981b" (UID: "9721431e-046b-4b3b-861e-22a65396981b"). InnerVolumeSpecName "kube-api-access-4vhth". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.100859 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "9721431e-046b-4b3b-861e-22a65396981b" (UID: "9721431e-046b-4b3b-861e-22a65396981b"). InnerVolumeSpecName "pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.103688 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9721431e-046b-4b3b-861e-22a65396981b-config" (OuterVolumeSpecName: "config") pod "9721431e-046b-4b3b-861e-22a65396981b" (UID: "9721431e-046b-4b3b-861e-22a65396981b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.112180 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9721431e-046b-4b3b-861e-22a65396981b-web-config" (OuterVolumeSpecName: "web-config") pod "9721431e-046b-4b3b-861e-22a65396981b" (UID: "9721431e-046b-4b3b-861e-22a65396981b"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.178514 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5145ef3e-2efa-4f76-b395-179abf1b1818-operator-scripts\") pod \"root-account-create-update-27j58\" (UID: \"5145ef3e-2efa-4f76-b395-179abf1b1818\") " pod="openstack/root-account-create-update-27j58" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.178563 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9gfz\" (UniqueName: \"kubernetes.io/projected/5145ef3e-2efa-4f76-b395-179abf1b1818-kube-api-access-m9gfz\") pod \"root-account-create-update-27j58\" (UID: \"5145ef3e-2efa-4f76-b395-179abf1b1818\") " pod="openstack/root-account-create-update-27j58" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.178613 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4vhth\" (UniqueName: \"kubernetes.io/projected/9721431e-046b-4b3b-861e-22a65396981b-kube-api-access-4vhth\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.178626 4700 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/9721431e-046b-4b3b-861e-22a65396981b-prometheus-metric-storage-rulefiles-2\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.178641 4700 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/9721431e-046b-4b3b-861e-22a65396981b-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.178650 4700 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9721431e-046b-4b3b-861e-22a65396981b-tls-assets\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.178659 4700 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9721431e-046b-4b3b-861e-22a65396981b-config-out\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.178670 4700 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/9721431e-046b-4b3b-861e-22a65396981b-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.178679 4700 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9721431e-046b-4b3b-861e-22a65396981b-web-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.178687 4700 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/9721431e-046b-4b3b-861e-22a65396981b-prometheus-metric-storage-rulefiles-1\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.178716 4700 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2\") on node \"crc\" " Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.178781 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/9721431e-046b-4b3b-861e-22a65396981b-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.195093 4700 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.195234 4700 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2") on node "crc" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.280662 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5145ef3e-2efa-4f76-b395-179abf1b1818-operator-scripts\") pod \"root-account-create-update-27j58\" (UID: \"5145ef3e-2efa-4f76-b395-179abf1b1818\") " pod="openstack/root-account-create-update-27j58" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.280706 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9gfz\" (UniqueName: \"kubernetes.io/projected/5145ef3e-2efa-4f76-b395-179abf1b1818-kube-api-access-m9gfz\") pod \"root-account-create-update-27j58\" (UID: \"5145ef3e-2efa-4f76-b395-179abf1b1818\") " pod="openstack/root-account-create-update-27j58" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.280911 4700 reconciler_common.go:293] "Volume detached for volume \"pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.281421 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5145ef3e-2efa-4f76-b395-179abf1b1818-operator-scripts\") pod \"root-account-create-update-27j58\" (UID: \"5145ef3e-2efa-4f76-b395-179abf1b1818\") " pod="openstack/root-account-create-update-27j58" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.295575 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9gfz\" (UniqueName: \"kubernetes.io/projected/5145ef3e-2efa-4f76-b395-179abf1b1818-kube-api-access-m9gfz\") pod \"root-account-create-update-27j58\" (UID: \"5145ef3e-2efa-4f76-b395-179abf1b1818\") " pod="openstack/root-account-create-update-27j58" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.380256 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-27j58" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.852239 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9721431e-046b-4b3b-861e-22a65396981b","Type":"ContainerDied","Data":"a1aeda0f76c4a3989dbd6ea69ae40b074a48598580bdf4195f0495834505d8a2"} Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.852260 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.852345 4700 scope.go:117] "RemoveContainer" containerID="99478540ee16a5f122a1e6e28f39329c9bd265403a641ecd3a881d8272258aee" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.886982 4700 scope.go:117] "RemoveContainer" containerID="4016795d07ac4553ff298150bb0cf07997d9258aee1bdce657c1c5422071402b" Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.902430 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.910573 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.943375 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Feb 27 17:20:39 crc kubenswrapper[4700]: I0227 17:20:39.977121 4700 scope.go:117] "RemoveContainer" containerID="7b9277446cc5fc2593bdc3b3cb6e1b7c283d1c39909d8b7b4ad6eef57493d935" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.003217 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-27j58"] Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.003513 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.010200 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-2" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.010350 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.012615 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.013105 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-jdvhl" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.013345 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.013321 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.015075 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-1" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.018408 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.025120 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.028633 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.063393 4700 scope.go:117] "RemoveContainer" containerID="750401c839d4786183e640b5a5874d5afec2bc8dbd5e838d8a1d75f3a9e675a8" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.144726 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/dc9ec3fe-326d-4df2-b561-318f574cfdf3-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.144783 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc9ec3fe-326d-4df2-b561-318f574cfdf3-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.144860 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.144878 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/dc9ec3fe-326d-4df2-b561-318f574cfdf3-config\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.144900 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/dc9ec3fe-326d-4df2-b561-318f574cfdf3-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.144925 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/dc9ec3fe-326d-4df2-b561-318f574cfdf3-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.144953 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/dc9ec3fe-326d-4df2-b561-318f574cfdf3-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.144975 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/dc9ec3fe-326d-4df2-b561-318f574cfdf3-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.145000 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/dc9ec3fe-326d-4df2-b561-318f574cfdf3-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.145016 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/dc9ec3fe-326d-4df2-b561-318f574cfdf3-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.145043 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/dc9ec3fe-326d-4df2-b561-318f574cfdf3-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.145061 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/dc9ec3fe-326d-4df2-b561-318f574cfdf3-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.145080 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bw6sb\" (UniqueName: \"kubernetes.io/projected/dc9ec3fe-326d-4df2-b561-318f574cfdf3-kube-api-access-bw6sb\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.246309 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.246354 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/dc9ec3fe-326d-4df2-b561-318f574cfdf3-config\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.246375 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/dc9ec3fe-326d-4df2-b561-318f574cfdf3-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.246404 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/dc9ec3fe-326d-4df2-b561-318f574cfdf3-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.246436 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/dc9ec3fe-326d-4df2-b561-318f574cfdf3-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.246463 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/dc9ec3fe-326d-4df2-b561-318f574cfdf3-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.246505 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/dc9ec3fe-326d-4df2-b561-318f574cfdf3-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.246523 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/dc9ec3fe-326d-4df2-b561-318f574cfdf3-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.246553 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/dc9ec3fe-326d-4df2-b561-318f574cfdf3-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.246573 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/dc9ec3fe-326d-4df2-b561-318f574cfdf3-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.246591 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw6sb\" (UniqueName: \"kubernetes.io/projected/dc9ec3fe-326d-4df2-b561-318f574cfdf3-kube-api-access-bw6sb\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.246611 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/dc9ec3fe-326d-4df2-b561-318f574cfdf3-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.246639 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc9ec3fe-326d-4df2-b561-318f574cfdf3-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.248247 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/dc9ec3fe-326d-4df2-b561-318f574cfdf3-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.258145 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/dc9ec3fe-326d-4df2-b561-318f574cfdf3-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.258837 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/dc9ec3fe-326d-4df2-b561-318f574cfdf3-config\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.259526 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/dc9ec3fe-326d-4df2-b561-318f574cfdf3-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.259842 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/dc9ec3fe-326d-4df2-b561-318f574cfdf3-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.261417 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/dc9ec3fe-326d-4df2-b561-318f574cfdf3-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.261808 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/dc9ec3fe-326d-4df2-b561-318f574cfdf3-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.262304 4700 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.262424 4700 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/51157a65261d84782fc18d298283232db1d99eb5d30079bdb31ebdc3e04ec75d/globalmount\"" pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.265424 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc9ec3fe-326d-4df2-b561-318f574cfdf3-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.269852 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/dc9ec3fe-326d-4df2-b561-318f574cfdf3-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.271945 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/dc9ec3fe-326d-4df2-b561-318f574cfdf3-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.279199 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bw6sb\" (UniqueName: \"kubernetes.io/projected/dc9ec3fe-326d-4df2-b561-318f574cfdf3-kube-api-access-bw6sb\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.283847 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/dc9ec3fe-326d-4df2-b561-318f574cfdf3-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.314008 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2\") pod \"prometheus-metric-storage-0\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.331279 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-sfldb-config-mss7g" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.356813 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.451084 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b032049e-5ad9-44a8-9520-fd7c1699748e-var-run-ovn\") pod \"b032049e-5ad9-44a8-9520-fd7c1699748e\" (UID: \"b032049e-5ad9-44a8-9520-fd7c1699748e\") " Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.451132 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b032049e-5ad9-44a8-9520-fd7c1699748e-var-run\") pod \"b032049e-5ad9-44a8-9520-fd7c1699748e\" (UID: \"b032049e-5ad9-44a8-9520-fd7c1699748e\") " Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.451175 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b032049e-5ad9-44a8-9520-fd7c1699748e-var-log-ovn\") pod \"b032049e-5ad9-44a8-9520-fd7c1699748e\" (UID: \"b032049e-5ad9-44a8-9520-fd7c1699748e\") " Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.451244 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b032049e-5ad9-44a8-9520-fd7c1699748e-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "b032049e-5ad9-44a8-9520-fd7c1699748e" (UID: "b032049e-5ad9-44a8-9520-fd7c1699748e"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.451285 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b032049e-5ad9-44a8-9520-fd7c1699748e-scripts\") pod \"b032049e-5ad9-44a8-9520-fd7c1699748e\" (UID: \"b032049e-5ad9-44a8-9520-fd7c1699748e\") " Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.451293 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b032049e-5ad9-44a8-9520-fd7c1699748e-var-run" (OuterVolumeSpecName: "var-run") pod "b032049e-5ad9-44a8-9520-fd7c1699748e" (UID: "b032049e-5ad9-44a8-9520-fd7c1699748e"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.451313 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b032049e-5ad9-44a8-9520-fd7c1699748e-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "b032049e-5ad9-44a8-9520-fd7c1699748e" (UID: "b032049e-5ad9-44a8-9520-fd7c1699748e"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.451355 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b032049e-5ad9-44a8-9520-fd7c1699748e-additional-scripts\") pod \"b032049e-5ad9-44a8-9520-fd7c1699748e\" (UID: \"b032049e-5ad9-44a8-9520-fd7c1699748e\") " Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.451417 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pdbj\" (UniqueName: \"kubernetes.io/projected/b032049e-5ad9-44a8-9520-fd7c1699748e-kube-api-access-4pdbj\") pod \"b032049e-5ad9-44a8-9520-fd7c1699748e\" (UID: \"b032049e-5ad9-44a8-9520-fd7c1699748e\") " Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.451821 4700 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b032049e-5ad9-44a8-9520-fd7c1699748e-var-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.451840 4700 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b032049e-5ad9-44a8-9520-fd7c1699748e-var-run\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.451849 4700 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b032049e-5ad9-44a8-9520-fd7c1699748e-var-log-ovn\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.452589 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b032049e-5ad9-44a8-9520-fd7c1699748e-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "b032049e-5ad9-44a8-9520-fd7c1699748e" (UID: "b032049e-5ad9-44a8-9520-fd7c1699748e"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.452902 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b032049e-5ad9-44a8-9520-fd7c1699748e-scripts" (OuterVolumeSpecName: "scripts") pod "b032049e-5ad9-44a8-9520-fd7c1699748e" (UID: "b032049e-5ad9-44a8-9520-fd7c1699748e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.456027 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b032049e-5ad9-44a8-9520-fd7c1699748e-kube-api-access-4pdbj" (OuterVolumeSpecName: "kube-api-access-4pdbj") pod "b032049e-5ad9-44a8-9520-fd7c1699748e" (UID: "b032049e-5ad9-44a8-9520-fd7c1699748e"). InnerVolumeSpecName "kube-api-access-4pdbj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.553321 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pdbj\" (UniqueName: \"kubernetes.io/projected/b032049e-5ad9-44a8-9520-fd7c1699748e-kube-api-access-4pdbj\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.553345 4700 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b032049e-5ad9-44a8-9520-fd7c1699748e-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.553354 4700 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b032049e-5ad9-44a8-9520-fd7c1699748e-additional-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.821388 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Feb 27 17:20:40 crc kubenswrapper[4700]: W0227 17:20:40.822450 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddc9ec3fe_326d_4df2_b561_318f574cfdf3.slice/crio-f04bb2e9aebd382befff678525f9d9408ceb5ed38fb1d6f61cd1910981782d4f WatchSource:0}: Error finding container f04bb2e9aebd382befff678525f9d9408ceb5ed38fb1d6f61cd1910981782d4f: Status 404 returned error can't find the container with id f04bb2e9aebd382befff678525f9d9408ceb5ed38fb1d6f61cd1910981782d4f Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.874145 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-sfldb-config-mss7g" event={"ID":"b032049e-5ad9-44a8-9520-fd7c1699748e","Type":"ContainerDied","Data":"8872c8e1736cd2b9c46cf4ecf1920e8b88798c3eec1c73b337a90d4580f25d09"} Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.874211 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8872c8e1736cd2b9c46cf4ecf1920e8b88798c3eec1c73b337a90d4580f25d09" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.874308 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-sfldb-config-mss7g" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.877751 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"dc9ec3fe-326d-4df2-b561-318f574cfdf3","Type":"ContainerStarted","Data":"f04bb2e9aebd382befff678525f9d9408ceb5ed38fb1d6f61cd1910981782d4f"} Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.883195 4700 generic.go:334] "Generic (PLEG): container finished" podID="5145ef3e-2efa-4f76-b395-179abf1b1818" containerID="c69cc325940dde65fa7783b1d92d27f073d11b12f2222b9e713d3414cb599515" exitCode=0 Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.883240 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-27j58" event={"ID":"5145ef3e-2efa-4f76-b395-179abf1b1818","Type":"ContainerDied","Data":"c69cc325940dde65fa7783b1d92d27f073d11b12f2222b9e713d3414cb599515"} Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.883264 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-27j58" event={"ID":"5145ef3e-2efa-4f76-b395-179abf1b1818","Type":"ContainerStarted","Data":"a9f2d286dee0fe12a227338d4f87192b597bd09397b9f1e48efc270abf1d787f"} Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.994281 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9721431e-046b-4b3b-861e-22a65396981b" path="/var/lib/kubelet/pods/9721431e-046b-4b3b-861e-22a65396981b/volumes" Feb 27 17:20:40 crc kubenswrapper[4700]: I0227 17:20:40.996338 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a28896eb-9f45-43fa-a078-8ae9175cb25d" path="/var/lib/kubelet/pods/a28896eb-9f45-43fa-a078-8ae9175cb25d/volumes" Feb 27 17:20:41 crc kubenswrapper[4700]: I0227 17:20:41.023684 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-sfldb" Feb 27 17:20:41 crc kubenswrapper[4700]: I0227 17:20:41.171266 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-9s7hk"] Feb 27 17:20:41 crc kubenswrapper[4700]: E0227 17:20:41.171958 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b032049e-5ad9-44a8-9520-fd7c1699748e" containerName="ovn-config" Feb 27 17:20:41 crc kubenswrapper[4700]: I0227 17:20:41.171976 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="b032049e-5ad9-44a8-9520-fd7c1699748e" containerName="ovn-config" Feb 27 17:20:41 crc kubenswrapper[4700]: I0227 17:20:41.172175 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="b032049e-5ad9-44a8-9520-fd7c1699748e" containerName="ovn-config" Feb 27 17:20:41 crc kubenswrapper[4700]: I0227 17:20:41.172898 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-9s7hk" Feb 27 17:20:41 crc kubenswrapper[4700]: I0227 17:20:41.176311 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Feb 27 17:20:41 crc kubenswrapper[4700]: I0227 17:20:41.178763 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-758nx" Feb 27 17:20:41 crc kubenswrapper[4700]: I0227 17:20:41.196670 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-9s7hk"] Feb 27 17:20:41 crc kubenswrapper[4700]: I0227 17:20:41.271854 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3d01029-07d5-414c-badd-0d28f0db9730-config-data\") pod \"glance-db-sync-9s7hk\" (UID: \"c3d01029-07d5-414c-badd-0d28f0db9730\") " pod="openstack/glance-db-sync-9s7hk" Feb 27 17:20:41 crc kubenswrapper[4700]: I0227 17:20:41.271968 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c3d01029-07d5-414c-badd-0d28f0db9730-db-sync-config-data\") pod \"glance-db-sync-9s7hk\" (UID: \"c3d01029-07d5-414c-badd-0d28f0db9730\") " pod="openstack/glance-db-sync-9s7hk" Feb 27 17:20:41 crc kubenswrapper[4700]: I0227 17:20:41.272056 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3d01029-07d5-414c-badd-0d28f0db9730-combined-ca-bundle\") pod \"glance-db-sync-9s7hk\" (UID: \"c3d01029-07d5-414c-badd-0d28f0db9730\") " pod="openstack/glance-db-sync-9s7hk" Feb 27 17:20:41 crc kubenswrapper[4700]: I0227 17:20:41.272190 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5hjn\" (UniqueName: \"kubernetes.io/projected/c3d01029-07d5-414c-badd-0d28f0db9730-kube-api-access-d5hjn\") pod \"glance-db-sync-9s7hk\" (UID: \"c3d01029-07d5-414c-badd-0d28f0db9730\") " pod="openstack/glance-db-sync-9s7hk" Feb 27 17:20:41 crc kubenswrapper[4700]: I0227 17:20:41.373984 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3d01029-07d5-414c-badd-0d28f0db9730-config-data\") pod \"glance-db-sync-9s7hk\" (UID: \"c3d01029-07d5-414c-badd-0d28f0db9730\") " pod="openstack/glance-db-sync-9s7hk" Feb 27 17:20:41 crc kubenswrapper[4700]: I0227 17:20:41.374046 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c3d01029-07d5-414c-badd-0d28f0db9730-db-sync-config-data\") pod \"glance-db-sync-9s7hk\" (UID: \"c3d01029-07d5-414c-badd-0d28f0db9730\") " pod="openstack/glance-db-sync-9s7hk" Feb 27 17:20:41 crc kubenswrapper[4700]: I0227 17:20:41.374104 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3d01029-07d5-414c-badd-0d28f0db9730-combined-ca-bundle\") pod \"glance-db-sync-9s7hk\" (UID: \"c3d01029-07d5-414c-badd-0d28f0db9730\") " pod="openstack/glance-db-sync-9s7hk" Feb 27 17:20:41 crc kubenswrapper[4700]: I0227 17:20:41.374190 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5hjn\" (UniqueName: \"kubernetes.io/projected/c3d01029-07d5-414c-badd-0d28f0db9730-kube-api-access-d5hjn\") pod \"glance-db-sync-9s7hk\" (UID: \"c3d01029-07d5-414c-badd-0d28f0db9730\") " pod="openstack/glance-db-sync-9s7hk" Feb 27 17:20:41 crc kubenswrapper[4700]: I0227 17:20:41.381524 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c3d01029-07d5-414c-badd-0d28f0db9730-db-sync-config-data\") pod \"glance-db-sync-9s7hk\" (UID: \"c3d01029-07d5-414c-badd-0d28f0db9730\") " pod="openstack/glance-db-sync-9s7hk" Feb 27 17:20:41 crc kubenswrapper[4700]: I0227 17:20:41.381955 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3d01029-07d5-414c-badd-0d28f0db9730-combined-ca-bundle\") pod \"glance-db-sync-9s7hk\" (UID: \"c3d01029-07d5-414c-badd-0d28f0db9730\") " pod="openstack/glance-db-sync-9s7hk" Feb 27 17:20:41 crc kubenswrapper[4700]: I0227 17:20:41.383572 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3d01029-07d5-414c-badd-0d28f0db9730-config-data\") pod \"glance-db-sync-9s7hk\" (UID: \"c3d01029-07d5-414c-badd-0d28f0db9730\") " pod="openstack/glance-db-sync-9s7hk" Feb 27 17:20:41 crc kubenswrapper[4700]: I0227 17:20:41.417644 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5hjn\" (UniqueName: \"kubernetes.io/projected/c3d01029-07d5-414c-badd-0d28f0db9730-kube-api-access-d5hjn\") pod \"glance-db-sync-9s7hk\" (UID: \"c3d01029-07d5-414c-badd-0d28f0db9730\") " pod="openstack/glance-db-sync-9s7hk" Feb 27 17:20:41 crc kubenswrapper[4700]: I0227 17:20:41.444406 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-sfldb-config-mss7g"] Feb 27 17:20:41 crc kubenswrapper[4700]: I0227 17:20:41.454933 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-sfldb-config-mss7g"] Feb 27 17:20:41 crc kubenswrapper[4700]: I0227 17:20:41.496998 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-9s7hk" Feb 27 17:20:42 crc kubenswrapper[4700]: I0227 17:20:42.474359 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-9s7hk"] Feb 27 17:20:42 crc kubenswrapper[4700]: W0227 17:20:42.482864 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc3d01029_07d5_414c_badd_0d28f0db9730.slice/crio-ba381c08a9b5a6b16ca31697476df256d350bd3f9823c42f93ec2e30ab086477 WatchSource:0}: Error finding container ba381c08a9b5a6b16ca31697476df256d350bd3f9823c42f93ec2e30ab086477: Status 404 returned error can't find the container with id ba381c08a9b5a6b16ca31697476df256d350bd3f9823c42f93ec2e30ab086477 Feb 27 17:20:42 crc kubenswrapper[4700]: I0227 17:20:42.834381 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-27j58" Feb 27 17:20:42 crc kubenswrapper[4700]: I0227 17:20:42.942209 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5145ef3e-2efa-4f76-b395-179abf1b1818-operator-scripts\") pod \"5145ef3e-2efa-4f76-b395-179abf1b1818\" (UID: \"5145ef3e-2efa-4f76-b395-179abf1b1818\") " Feb 27 17:20:42 crc kubenswrapper[4700]: I0227 17:20:42.942268 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9gfz\" (UniqueName: \"kubernetes.io/projected/5145ef3e-2efa-4f76-b395-179abf1b1818-kube-api-access-m9gfz\") pod \"5145ef3e-2efa-4f76-b395-179abf1b1818\" (UID: \"5145ef3e-2efa-4f76-b395-179abf1b1818\") " Feb 27 17:20:42 crc kubenswrapper[4700]: I0227 17:20:42.944043 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5145ef3e-2efa-4f76-b395-179abf1b1818-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5145ef3e-2efa-4f76-b395-179abf1b1818" (UID: "5145ef3e-2efa-4f76-b395-179abf1b1818"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:20:42 crc kubenswrapper[4700]: I0227 17:20:42.949881 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5145ef3e-2efa-4f76-b395-179abf1b1818-kube-api-access-m9gfz" (OuterVolumeSpecName: "kube-api-access-m9gfz") pod "5145ef3e-2efa-4f76-b395-179abf1b1818" (UID: "5145ef3e-2efa-4f76-b395-179abf1b1818"). InnerVolumeSpecName "kube-api-access-m9gfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:20:42 crc kubenswrapper[4700]: I0227 17:20:42.995021 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b032049e-5ad9-44a8-9520-fd7c1699748e" path="/var/lib/kubelet/pods/b032049e-5ad9-44a8-9520-fd7c1699748e/volumes" Feb 27 17:20:43 crc kubenswrapper[4700]: I0227 17:20:43.044779 4700 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5145ef3e-2efa-4f76-b395-179abf1b1818-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:43 crc kubenswrapper[4700]: I0227 17:20:43.044815 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9gfz\" (UniqueName: \"kubernetes.io/projected/5145ef3e-2efa-4f76-b395-179abf1b1818-kube-api-access-m9gfz\") on node \"crc\" DevicePath \"\"" Feb 27 17:20:43 crc kubenswrapper[4700]: I0227 17:20:43.362854 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-27j58" Feb 27 17:20:43 crc kubenswrapper[4700]: I0227 17:20:43.362919 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-27j58" event={"ID":"5145ef3e-2efa-4f76-b395-179abf1b1818","Type":"ContainerDied","Data":"a9f2d286dee0fe12a227338d4f87192b597bd09397b9f1e48efc270abf1d787f"} Feb 27 17:20:43 crc kubenswrapper[4700]: I0227 17:20:43.362973 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9f2d286dee0fe12a227338d4f87192b597bd09397b9f1e48efc270abf1d787f" Feb 27 17:20:43 crc kubenswrapper[4700]: I0227 17:20:43.364280 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-9s7hk" event={"ID":"c3d01029-07d5-414c-badd-0d28f0db9730","Type":"ContainerStarted","Data":"ba381c08a9b5a6b16ca31697476df256d350bd3f9823c42f93ec2e30ab086477"} Feb 27 17:20:44 crc kubenswrapper[4700]: I0227 17:20:44.374122 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"dc9ec3fe-326d-4df2-b561-318f574cfdf3","Type":"ContainerStarted","Data":"2628eb269c1943b0e5e9f37bb8c617751b464272b789e5fc05c1a2050a641c11"} Feb 27 17:20:45 crc kubenswrapper[4700]: I0227 17:20:45.349878 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Feb 27 17:20:45 crc kubenswrapper[4700]: I0227 17:20:45.385120 4700 generic.go:334] "Generic (PLEG): container finished" podID="458e6422-b856-4bcf-8308-4b6cb9ec8fef" containerID="3c345536df56b1f58f851e4bdb0af75eeb80bcf3825e3bbbaf62f032bc472793" exitCode=0 Feb 27 17:20:45 crc kubenswrapper[4700]: I0227 17:20:45.385198 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"458e6422-b856-4bcf-8308-4b6cb9ec8fef","Type":"ContainerDied","Data":"3c345536df56b1f58f851e4bdb0af75eeb80bcf3825e3bbbaf62f032bc472793"} Feb 27 17:20:45 crc kubenswrapper[4700]: I0227 17:20:45.388891 4700 generic.go:334] "Generic (PLEG): container finished" podID="efe7d637-da3a-4995-a40f-fae00257ac1d" containerID="eb50098b16c98521dac5370ccb4478720cceb5dc15fabc0e5682e5ccc113eb78" exitCode=0 Feb 27 17:20:45 crc kubenswrapper[4700]: I0227 17:20:45.388962 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"efe7d637-da3a-4995-a40f-fae00257ac1d","Type":"ContainerDied","Data":"eb50098b16c98521dac5370ccb4478720cceb5dc15fabc0e5682e5ccc113eb78"} Feb 27 17:20:46 crc kubenswrapper[4700]: I0227 17:20:46.400343 4700 generic.go:334] "Generic (PLEG): container finished" podID="fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da" containerID="2f40fb47f39b8deb36a5e78ca5f657280f1b38d8420607c6ccf0836addcac851" exitCode=0 Feb 27 17:20:46 crc kubenswrapper[4700]: I0227 17:20:46.400480 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/notifications-rabbitmq-server-0" event={"ID":"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da","Type":"ContainerDied","Data":"2f40fb47f39b8deb36a5e78ca5f657280f1b38d8420607c6ccf0836addcac851"} Feb 27 17:20:46 crc kubenswrapper[4700]: I0227 17:20:46.405837 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"458e6422-b856-4bcf-8308-4b6cb9ec8fef","Type":"ContainerStarted","Data":"520ea869bacd9c0536d61aea44f09946e35bd5b34aac34d9e9100c9c2ed55787"} Feb 27 17:20:46 crc kubenswrapper[4700]: I0227 17:20:46.406148 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:20:46 crc kubenswrapper[4700]: I0227 17:20:46.409866 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"efe7d637-da3a-4995-a40f-fae00257ac1d","Type":"ContainerStarted","Data":"3a5a3f9cb7f22f3a2b43a0279b69f515963804662da1ec627fe5bcf97b746739"} Feb 27 17:20:46 crc kubenswrapper[4700]: I0227 17:20:46.410590 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Feb 27 17:20:46 crc kubenswrapper[4700]: I0227 17:20:46.474683 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=54.140665837 podStartE2EDuration="1m21.474662199s" podCreationTimestamp="2026-02-27 17:19:25 +0000 UTC" firstStartedPulling="2026-02-27 17:19:43.723802367 +0000 UTC m=+1143.709115104" lastFinishedPulling="2026-02-27 17:20:11.057798719 +0000 UTC m=+1171.043111466" observedRunningTime="2026-02-27 17:20:46.453667524 +0000 UTC m=+1206.438980281" watchObservedRunningTime="2026-02-27 17:20:46.474662199 +0000 UTC m=+1206.459974966" Feb 27 17:20:46 crc kubenswrapper[4700]: I0227 17:20:46.477961 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=-9223371956.376825 podStartE2EDuration="1m20.477949806s" podCreationTimestamp="2026-02-27 17:19:26 +0000 UTC" firstStartedPulling="2026-02-27 17:19:43.767197078 +0000 UTC m=+1143.752509825" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:20:46.473432697 +0000 UTC m=+1206.458745444" watchObservedRunningTime="2026-02-27 17:20:46.477949806 +0000 UTC m=+1206.463262573" Feb 27 17:20:46 crc kubenswrapper[4700]: I0227 17:20:46.514528 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9df57c73-8d37-4f7d-b62f-209210dbba42-etc-swift\") pod \"swift-storage-0\" (UID: \"9df57c73-8d37-4f7d-b62f-209210dbba42\") " pod="openstack/swift-storage-0" Feb 27 17:20:46 crc kubenswrapper[4700]: I0227 17:20:46.519409 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9df57c73-8d37-4f7d-b62f-209210dbba42-etc-swift\") pod \"swift-storage-0\" (UID: \"9df57c73-8d37-4f7d-b62f-209210dbba42\") " pod="openstack/swift-storage-0" Feb 27 17:20:46 crc kubenswrapper[4700]: I0227 17:20:46.776525 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 27 17:20:47 crc kubenswrapper[4700]: I0227 17:20:47.380426 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Feb 27 17:20:47 crc kubenswrapper[4700]: I0227 17:20:47.430128 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/notifications-rabbitmq-server-0" event={"ID":"fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da","Type":"ContainerStarted","Data":"abd0c970aa5bd565a191198aa35722e10298a954de6698adac9683cdb956fb7f"} Feb 27 17:20:47 crc kubenswrapper[4700]: I0227 17:20:47.456282 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/notifications-rabbitmq-server-0" podStartSLOduration=-9223371954.398512 podStartE2EDuration="1m22.456263014s" podCreationTimestamp="2026-02-27 17:19:25 +0000 UTC" firstStartedPulling="2026-02-27 17:19:43.666448665 +0000 UTC m=+1143.651761412" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:20:47.454437015 +0000 UTC m=+1207.439749772" watchObservedRunningTime="2026-02-27 17:20:47.456263014 +0000 UTC m=+1207.441575771" Feb 27 17:20:50 crc kubenswrapper[4700]: I0227 17:20:50.454685 4700 generic.go:334] "Generic (PLEG): container finished" podID="dc9ec3fe-326d-4df2-b561-318f574cfdf3" containerID="2628eb269c1943b0e5e9f37bb8c617751b464272b789e5fc05c1a2050a641c11" exitCode=0 Feb 27 17:20:50 crc kubenswrapper[4700]: I0227 17:20:50.455052 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"dc9ec3fe-326d-4df2-b561-318f574cfdf3","Type":"ContainerDied","Data":"2628eb269c1943b0e5e9f37bb8c617751b464272b789e5fc05c1a2050a641c11"} Feb 27 17:20:54 crc kubenswrapper[4700]: I0227 17:20:54.492844 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9df57c73-8d37-4f7d-b62f-209210dbba42","Type":"ContainerStarted","Data":"6ee7e5b3ae94a29141da9dc7f1bc97b0f691def54184524024b52fb8dcbd5463"} Feb 27 17:20:55 crc kubenswrapper[4700]: I0227 17:20:55.503515 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-9s7hk" event={"ID":"c3d01029-07d5-414c-badd-0d28f0db9730","Type":"ContainerStarted","Data":"45057e85f95f295f9914750cfa486175715ffae747822c8d4025a5d9b9b95ad6"} Feb 27 17:20:55 crc kubenswrapper[4700]: I0227 17:20:55.506644 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9df57c73-8d37-4f7d-b62f-209210dbba42","Type":"ContainerStarted","Data":"860c2fa40cb86fbf50501b1839e69f9169233253eb79c70ff8369ee0ab2e5aff"} Feb 27 17:20:55 crc kubenswrapper[4700]: I0227 17:20:55.509995 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"dc9ec3fe-326d-4df2-b561-318f574cfdf3","Type":"ContainerStarted","Data":"a3ba82809ba9f29c586671d969f56e51963e184bd5c34352d55e7d387d108965"} Feb 27 17:20:55 crc kubenswrapper[4700]: I0227 17:20:55.531146 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-9s7hk" podStartSLOduration=2.502054607 podStartE2EDuration="14.531128571s" podCreationTimestamp="2026-02-27 17:20:41 +0000 UTC" firstStartedPulling="2026-02-27 17:20:42.485313142 +0000 UTC m=+1202.470625899" lastFinishedPulling="2026-02-27 17:20:54.514387076 +0000 UTC m=+1214.499699863" observedRunningTime="2026-02-27 17:20:55.524599648 +0000 UTC m=+1215.509912385" watchObservedRunningTime="2026-02-27 17:20:55.531128571 +0000 UTC m=+1215.516441318" Feb 27 17:20:56 crc kubenswrapper[4700]: I0227 17:20:56.522379 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9df57c73-8d37-4f7d-b62f-209210dbba42","Type":"ContainerStarted","Data":"35344c59c968a53b263e0171a3b35fc147c3007332f381a877b50c0918bf91c0"} Feb 27 17:20:56 crc kubenswrapper[4700]: I0227 17:20:56.523094 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9df57c73-8d37-4f7d-b62f-209210dbba42","Type":"ContainerStarted","Data":"bde7e90ff655f75aaaf559e98a1303d4859facd72f06e0e209093c47f9fd3ca5"} Feb 27 17:20:56 crc kubenswrapper[4700]: I0227 17:20:56.523119 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9df57c73-8d37-4f7d-b62f-209210dbba42","Type":"ContainerStarted","Data":"572e92276e6cc58523bfdce372b41f00f2260739610f696bdf091eb473f29280"} Feb 27 17:20:57 crc kubenswrapper[4700]: I0227 17:20:57.090395 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="efe7d637-da3a-4995-a40f-fae00257ac1d" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.109:5671: connect: connection refused" Feb 27 17:20:57 crc kubenswrapper[4700]: I0227 17:20:57.383927 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:20:57 crc kubenswrapper[4700]: I0227 17:20:57.385910 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/notifications-rabbitmq-server-0" podUID="fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.110:5671: connect: connection refused" Feb 27 17:20:57 crc kubenswrapper[4700]: I0227 17:20:57.386248 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/notifications-rabbitmq-server-0" podUID="fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.110:5671: connect: connection refused" Feb 27 17:20:57 crc kubenswrapper[4700]: I0227 17:20:57.531929 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9df57c73-8d37-4f7d-b62f-209210dbba42","Type":"ContainerStarted","Data":"17bfacde52bafb15214ced1c91da4cae37a81897208f7cdd901d37fb16d5abeb"} Feb 27 17:20:57 crc kubenswrapper[4700]: I0227 17:20:57.709652 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="458e6422-b856-4bcf-8308-4b6cb9ec8fef" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.111:5671: connect: connection refused" Feb 27 17:20:58 crc kubenswrapper[4700]: I0227 17:20:58.542161 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"dc9ec3fe-326d-4df2-b561-318f574cfdf3","Type":"ContainerStarted","Data":"f2821b24a20a1fda4814c4839d720bb09a0da9f533505a5024cf4d4cb922684c"} Feb 27 17:20:58 crc kubenswrapper[4700]: I0227 17:20:58.542481 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"dc9ec3fe-326d-4df2-b561-318f574cfdf3","Type":"ContainerStarted","Data":"97260215fdb80740d553458dd4cf1caafb5e06781782252f81abe8bd85e205b1"} Feb 27 17:20:58 crc kubenswrapper[4700]: I0227 17:20:58.547162 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9df57c73-8d37-4f7d-b62f-209210dbba42","Type":"ContainerStarted","Data":"a0703defdb29a08153d9f6274fa60a28886e15312475ffbfd361160324f813b3"} Feb 27 17:20:58 crc kubenswrapper[4700]: I0227 17:20:58.547190 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9df57c73-8d37-4f7d-b62f-209210dbba42","Type":"ContainerStarted","Data":"489830825dec8f72364b95408d3de59a11f1e129db2a745980f9eeed758fa92d"} Feb 27 17:20:58 crc kubenswrapper[4700]: I0227 17:20:58.547199 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9df57c73-8d37-4f7d-b62f-209210dbba42","Type":"ContainerStarted","Data":"231d3092b1beb609e2a77aae370ee945f2b10b1fcb5050d20be30575f4f90882"} Feb 27 17:20:58 crc kubenswrapper[4700]: I0227 17:20:58.590141 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=19.590127408 podStartE2EDuration="19.590127408s" podCreationTimestamp="2026-02-27 17:20:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:20:58.587620651 +0000 UTC m=+1218.572933398" watchObservedRunningTime="2026-02-27 17:20:58.590127408 +0000 UTC m=+1218.575440155" Feb 27 17:20:59 crc kubenswrapper[4700]: I0227 17:20:59.558883 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9df57c73-8d37-4f7d-b62f-209210dbba42","Type":"ContainerStarted","Data":"e5d5a5f5e2a0a88728a755676234b1963237cf33f205c4c2da3c5de2a17d331a"} Feb 27 17:20:59 crc kubenswrapper[4700]: I0227 17:20:59.559175 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9df57c73-8d37-4f7d-b62f-209210dbba42","Type":"ContainerStarted","Data":"869949d3b00c3be7f07cbb4b0deca28b310d97f7d83c332a643bf3fb477425f9"} Feb 27 17:20:59 crc kubenswrapper[4700]: I0227 17:20:59.559187 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9df57c73-8d37-4f7d-b62f-209210dbba42","Type":"ContainerStarted","Data":"c148ad6ffac5dc26b2621d6558baccae7bd12aa17adbd3731e4cb7cb66a61d38"} Feb 27 17:20:59 crc kubenswrapper[4700]: I0227 17:20:59.559196 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9df57c73-8d37-4f7d-b62f-209210dbba42","Type":"ContainerStarted","Data":"9b5bac3d875f76953292ef4d9e15345d4822c5198d0156223ea1d2b6abb7c41d"} Feb 27 17:20:59 crc kubenswrapper[4700]: I0227 17:20:59.559206 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9df57c73-8d37-4f7d-b62f-209210dbba42","Type":"ContainerStarted","Data":"8d17a50cb89762b7812c876c1babae47b4575887465e5b2b6f06da4cde9508e0"} Feb 27 17:20:59 crc kubenswrapper[4700]: I0227 17:20:59.561022 4700 scope.go:117] "RemoveContainer" containerID="c90b908abe9c5264dc998e9123b4919a8f267032d44c1b05b1861f6277e3fef6" Feb 27 17:21:00 crc kubenswrapper[4700]: I0227 17:21:00.357505 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Feb 27 17:21:00 crc kubenswrapper[4700]: I0227 17:21:00.598269 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9df57c73-8d37-4f7d-b62f-209210dbba42","Type":"ContainerStarted","Data":"11cd180603b2adb20c3f0d23fb03bd3e45a45953d6bbdc77fa90e6e18b1b198d"} Feb 27 17:21:00 crc kubenswrapper[4700]: I0227 17:21:00.598352 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9df57c73-8d37-4f7d-b62f-209210dbba42","Type":"ContainerStarted","Data":"5ac8d45acd634cf2b2c2170479f3dc4209c469b67c8cfaca18cbc55a59b49fd2"} Feb 27 17:21:00 crc kubenswrapper[4700]: I0227 17:21:00.658512 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=43.502923047 podStartE2EDuration="47.658494221s" podCreationTimestamp="2026-02-27 17:20:13 +0000 UTC" firstStartedPulling="2026-02-27 17:20:54.389374749 +0000 UTC m=+1214.374687526" lastFinishedPulling="2026-02-27 17:20:58.544945953 +0000 UTC m=+1218.530258700" observedRunningTime="2026-02-27 17:21:00.651594899 +0000 UTC m=+1220.636907656" watchObservedRunningTime="2026-02-27 17:21:00.658494221 +0000 UTC m=+1220.643806968" Feb 27 17:21:00 crc kubenswrapper[4700]: I0227 17:21:00.999723 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-687c5777ff-vx8jc"] Feb 27 17:21:01 crc kubenswrapper[4700]: E0227 17:21:01.000486 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5145ef3e-2efa-4f76-b395-179abf1b1818" containerName="mariadb-account-create-update" Feb 27 17:21:01 crc kubenswrapper[4700]: I0227 17:21:01.000606 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="5145ef3e-2efa-4f76-b395-179abf1b1818" containerName="mariadb-account-create-update" Feb 27 17:21:01 crc kubenswrapper[4700]: I0227 17:21:01.000916 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="5145ef3e-2efa-4f76-b395-179abf1b1818" containerName="mariadb-account-create-update" Feb 27 17:21:01 crc kubenswrapper[4700]: I0227 17:21:01.003452 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-687c5777ff-vx8jc" Feb 27 17:21:01 crc kubenswrapper[4700]: I0227 17:21:01.009053 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-687c5777ff-vx8jc"] Feb 27 17:21:01 crc kubenswrapper[4700]: I0227 17:21:01.010851 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Feb 27 17:21:01 crc kubenswrapper[4700]: I0227 17:21:01.055324 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f455d2a5-6e87-4287-a00e-83f724dd180c-dns-swift-storage-0\") pod \"dnsmasq-dns-687c5777ff-vx8jc\" (UID: \"f455d2a5-6e87-4287-a00e-83f724dd180c\") " pod="openstack/dnsmasq-dns-687c5777ff-vx8jc" Feb 27 17:21:01 crc kubenswrapper[4700]: I0227 17:21:01.055382 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f455d2a5-6e87-4287-a00e-83f724dd180c-config\") pod \"dnsmasq-dns-687c5777ff-vx8jc\" (UID: \"f455d2a5-6e87-4287-a00e-83f724dd180c\") " pod="openstack/dnsmasq-dns-687c5777ff-vx8jc" Feb 27 17:21:01 crc kubenswrapper[4700]: I0227 17:21:01.055425 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f455d2a5-6e87-4287-a00e-83f724dd180c-ovsdbserver-nb\") pod \"dnsmasq-dns-687c5777ff-vx8jc\" (UID: \"f455d2a5-6e87-4287-a00e-83f724dd180c\") " pod="openstack/dnsmasq-dns-687c5777ff-vx8jc" Feb 27 17:21:01 crc kubenswrapper[4700]: I0227 17:21:01.055533 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f455d2a5-6e87-4287-a00e-83f724dd180c-ovsdbserver-sb\") pod \"dnsmasq-dns-687c5777ff-vx8jc\" (UID: \"f455d2a5-6e87-4287-a00e-83f724dd180c\") " pod="openstack/dnsmasq-dns-687c5777ff-vx8jc" Feb 27 17:21:01 crc kubenswrapper[4700]: I0227 17:21:01.055584 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f455d2a5-6e87-4287-a00e-83f724dd180c-dns-svc\") pod \"dnsmasq-dns-687c5777ff-vx8jc\" (UID: \"f455d2a5-6e87-4287-a00e-83f724dd180c\") " pod="openstack/dnsmasq-dns-687c5777ff-vx8jc" Feb 27 17:21:01 crc kubenswrapper[4700]: I0227 17:21:01.055632 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkxr6\" (UniqueName: \"kubernetes.io/projected/f455d2a5-6e87-4287-a00e-83f724dd180c-kube-api-access-dkxr6\") pod \"dnsmasq-dns-687c5777ff-vx8jc\" (UID: \"f455d2a5-6e87-4287-a00e-83f724dd180c\") " pod="openstack/dnsmasq-dns-687c5777ff-vx8jc" Feb 27 17:21:01 crc kubenswrapper[4700]: I0227 17:21:01.157557 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f455d2a5-6e87-4287-a00e-83f724dd180c-dns-swift-storage-0\") pod \"dnsmasq-dns-687c5777ff-vx8jc\" (UID: \"f455d2a5-6e87-4287-a00e-83f724dd180c\") " pod="openstack/dnsmasq-dns-687c5777ff-vx8jc" Feb 27 17:21:01 crc kubenswrapper[4700]: I0227 17:21:01.157863 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f455d2a5-6e87-4287-a00e-83f724dd180c-config\") pod \"dnsmasq-dns-687c5777ff-vx8jc\" (UID: \"f455d2a5-6e87-4287-a00e-83f724dd180c\") " pod="openstack/dnsmasq-dns-687c5777ff-vx8jc" Feb 27 17:21:01 crc kubenswrapper[4700]: I0227 17:21:01.157923 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f455d2a5-6e87-4287-a00e-83f724dd180c-ovsdbserver-nb\") pod \"dnsmasq-dns-687c5777ff-vx8jc\" (UID: \"f455d2a5-6e87-4287-a00e-83f724dd180c\") " pod="openstack/dnsmasq-dns-687c5777ff-vx8jc" Feb 27 17:21:01 crc kubenswrapper[4700]: I0227 17:21:01.157996 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f455d2a5-6e87-4287-a00e-83f724dd180c-ovsdbserver-sb\") pod \"dnsmasq-dns-687c5777ff-vx8jc\" (UID: \"f455d2a5-6e87-4287-a00e-83f724dd180c\") " pod="openstack/dnsmasq-dns-687c5777ff-vx8jc" Feb 27 17:21:01 crc kubenswrapper[4700]: I0227 17:21:01.158049 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f455d2a5-6e87-4287-a00e-83f724dd180c-dns-svc\") pod \"dnsmasq-dns-687c5777ff-vx8jc\" (UID: \"f455d2a5-6e87-4287-a00e-83f724dd180c\") " pod="openstack/dnsmasq-dns-687c5777ff-vx8jc" Feb 27 17:21:01 crc kubenswrapper[4700]: I0227 17:21:01.158072 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkxr6\" (UniqueName: \"kubernetes.io/projected/f455d2a5-6e87-4287-a00e-83f724dd180c-kube-api-access-dkxr6\") pod \"dnsmasq-dns-687c5777ff-vx8jc\" (UID: \"f455d2a5-6e87-4287-a00e-83f724dd180c\") " pod="openstack/dnsmasq-dns-687c5777ff-vx8jc" Feb 27 17:21:01 crc kubenswrapper[4700]: I0227 17:21:01.158352 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f455d2a5-6e87-4287-a00e-83f724dd180c-dns-swift-storage-0\") pod \"dnsmasq-dns-687c5777ff-vx8jc\" (UID: \"f455d2a5-6e87-4287-a00e-83f724dd180c\") " pod="openstack/dnsmasq-dns-687c5777ff-vx8jc" Feb 27 17:21:01 crc kubenswrapper[4700]: I0227 17:21:01.158868 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f455d2a5-6e87-4287-a00e-83f724dd180c-config\") pod \"dnsmasq-dns-687c5777ff-vx8jc\" (UID: \"f455d2a5-6e87-4287-a00e-83f724dd180c\") " pod="openstack/dnsmasq-dns-687c5777ff-vx8jc" Feb 27 17:21:01 crc kubenswrapper[4700]: I0227 17:21:01.158978 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f455d2a5-6e87-4287-a00e-83f724dd180c-ovsdbserver-sb\") pod \"dnsmasq-dns-687c5777ff-vx8jc\" (UID: \"f455d2a5-6e87-4287-a00e-83f724dd180c\") " pod="openstack/dnsmasq-dns-687c5777ff-vx8jc" Feb 27 17:21:01 crc kubenswrapper[4700]: I0227 17:21:01.159373 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f455d2a5-6e87-4287-a00e-83f724dd180c-dns-svc\") pod \"dnsmasq-dns-687c5777ff-vx8jc\" (UID: \"f455d2a5-6e87-4287-a00e-83f724dd180c\") " pod="openstack/dnsmasq-dns-687c5777ff-vx8jc" Feb 27 17:21:01 crc kubenswrapper[4700]: I0227 17:21:01.159633 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f455d2a5-6e87-4287-a00e-83f724dd180c-ovsdbserver-nb\") pod \"dnsmasq-dns-687c5777ff-vx8jc\" (UID: \"f455d2a5-6e87-4287-a00e-83f724dd180c\") " pod="openstack/dnsmasq-dns-687c5777ff-vx8jc" Feb 27 17:21:01 crc kubenswrapper[4700]: I0227 17:21:01.176262 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkxr6\" (UniqueName: \"kubernetes.io/projected/f455d2a5-6e87-4287-a00e-83f724dd180c-kube-api-access-dkxr6\") pod \"dnsmasq-dns-687c5777ff-vx8jc\" (UID: \"f455d2a5-6e87-4287-a00e-83f724dd180c\") " pod="openstack/dnsmasq-dns-687c5777ff-vx8jc" Feb 27 17:21:01 crc kubenswrapper[4700]: I0227 17:21:01.338752 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-687c5777ff-vx8jc" Feb 27 17:21:02 crc kubenswrapper[4700]: I0227 17:21:02.323891 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-687c5777ff-vx8jc"] Feb 27 17:21:02 crc kubenswrapper[4700]: W0227 17:21:02.325680 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf455d2a5_6e87_4287_a00e_83f724dd180c.slice/crio-1e731afb9fba9bc466ee7e696586017bd40620ff1196b0fa0e6a4600de02eb23 WatchSource:0}: Error finding container 1e731afb9fba9bc466ee7e696586017bd40620ff1196b0fa0e6a4600de02eb23: Status 404 returned error can't find the container with id 1e731afb9fba9bc466ee7e696586017bd40620ff1196b0fa0e6a4600de02eb23 Feb 27 17:21:02 crc kubenswrapper[4700]: I0227 17:21:02.623627 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-687c5777ff-vx8jc" event={"ID":"f455d2a5-6e87-4287-a00e-83f724dd180c","Type":"ContainerStarted","Data":"2380b8d90f49e4e6906e0ad15caa0ab116899dbec6d1a1d1077a82bb95768c07"} Feb 27 17:21:02 crc kubenswrapper[4700]: I0227 17:21:02.623989 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-687c5777ff-vx8jc" event={"ID":"f455d2a5-6e87-4287-a00e-83f724dd180c","Type":"ContainerStarted","Data":"1e731afb9fba9bc466ee7e696586017bd40620ff1196b0fa0e6a4600de02eb23"} Feb 27 17:21:02 crc kubenswrapper[4700]: I0227 17:21:02.625605 4700 generic.go:334] "Generic (PLEG): container finished" podID="c3d01029-07d5-414c-badd-0d28f0db9730" containerID="45057e85f95f295f9914750cfa486175715ffae747822c8d4025a5d9b9b95ad6" exitCode=0 Feb 27 17:21:02 crc kubenswrapper[4700]: I0227 17:21:02.625656 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-9s7hk" event={"ID":"c3d01029-07d5-414c-badd-0d28f0db9730","Type":"ContainerDied","Data":"45057e85f95f295f9914750cfa486175715ffae747822c8d4025a5d9b9b95ad6"} Feb 27 17:21:03 crc kubenswrapper[4700]: I0227 17:21:03.637625 4700 generic.go:334] "Generic (PLEG): container finished" podID="f455d2a5-6e87-4287-a00e-83f724dd180c" containerID="2380b8d90f49e4e6906e0ad15caa0ab116899dbec6d1a1d1077a82bb95768c07" exitCode=0 Feb 27 17:21:03 crc kubenswrapper[4700]: I0227 17:21:03.637726 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-687c5777ff-vx8jc" event={"ID":"f455d2a5-6e87-4287-a00e-83f724dd180c","Type":"ContainerDied","Data":"2380b8d90f49e4e6906e0ad15caa0ab116899dbec6d1a1d1077a82bb95768c07"} Feb 27 17:21:04 crc kubenswrapper[4700]: I0227 17:21:04.390063 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-9s7hk" Feb 27 17:21:04 crc kubenswrapper[4700]: I0227 17:21:04.539703 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3d01029-07d5-414c-badd-0d28f0db9730-config-data\") pod \"c3d01029-07d5-414c-badd-0d28f0db9730\" (UID: \"c3d01029-07d5-414c-badd-0d28f0db9730\") " Feb 27 17:21:04 crc kubenswrapper[4700]: I0227 17:21:04.540276 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3d01029-07d5-414c-badd-0d28f0db9730-combined-ca-bundle\") pod \"c3d01029-07d5-414c-badd-0d28f0db9730\" (UID: \"c3d01029-07d5-414c-badd-0d28f0db9730\") " Feb 27 17:21:04 crc kubenswrapper[4700]: I0227 17:21:04.540632 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5hjn\" (UniqueName: \"kubernetes.io/projected/c3d01029-07d5-414c-badd-0d28f0db9730-kube-api-access-d5hjn\") pod \"c3d01029-07d5-414c-badd-0d28f0db9730\" (UID: \"c3d01029-07d5-414c-badd-0d28f0db9730\") " Feb 27 17:21:04 crc kubenswrapper[4700]: I0227 17:21:04.541009 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c3d01029-07d5-414c-badd-0d28f0db9730-db-sync-config-data\") pod \"c3d01029-07d5-414c-badd-0d28f0db9730\" (UID: \"c3d01029-07d5-414c-badd-0d28f0db9730\") " Feb 27 17:21:04 crc kubenswrapper[4700]: I0227 17:21:04.548781 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3d01029-07d5-414c-badd-0d28f0db9730-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "c3d01029-07d5-414c-badd-0d28f0db9730" (UID: "c3d01029-07d5-414c-badd-0d28f0db9730"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:21:04 crc kubenswrapper[4700]: I0227 17:21:04.549833 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3d01029-07d5-414c-badd-0d28f0db9730-kube-api-access-d5hjn" (OuterVolumeSpecName: "kube-api-access-d5hjn") pod "c3d01029-07d5-414c-badd-0d28f0db9730" (UID: "c3d01029-07d5-414c-badd-0d28f0db9730"). InnerVolumeSpecName "kube-api-access-d5hjn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:21:04 crc kubenswrapper[4700]: I0227 17:21:04.618412 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3d01029-07d5-414c-badd-0d28f0db9730-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c3d01029-07d5-414c-badd-0d28f0db9730" (UID: "c3d01029-07d5-414c-badd-0d28f0db9730"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:21:04 crc kubenswrapper[4700]: I0227 17:21:04.632769 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3d01029-07d5-414c-badd-0d28f0db9730-config-data" (OuterVolumeSpecName: "config-data") pod "c3d01029-07d5-414c-badd-0d28f0db9730" (UID: "c3d01029-07d5-414c-badd-0d28f0db9730"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:21:04 crc kubenswrapper[4700]: I0227 17:21:04.645218 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3d01029-07d5-414c-badd-0d28f0db9730-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:04 crc kubenswrapper[4700]: I0227 17:21:04.645267 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3d01029-07d5-414c-badd-0d28f0db9730-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:04 crc kubenswrapper[4700]: I0227 17:21:04.645288 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5hjn\" (UniqueName: \"kubernetes.io/projected/c3d01029-07d5-414c-badd-0d28f0db9730-kube-api-access-d5hjn\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:04 crc kubenswrapper[4700]: I0227 17:21:04.645306 4700 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c3d01029-07d5-414c-badd-0d28f0db9730-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:04 crc kubenswrapper[4700]: I0227 17:21:04.651635 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-687c5777ff-vx8jc" event={"ID":"f455d2a5-6e87-4287-a00e-83f724dd180c","Type":"ContainerStarted","Data":"fe14e67a4dd9ee316ed15ed83bfdbe18c56cf274127121742b785da0b17d0c16"} Feb 27 17:21:04 crc kubenswrapper[4700]: I0227 17:21:04.651852 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-687c5777ff-vx8jc" Feb 27 17:21:04 crc kubenswrapper[4700]: I0227 17:21:04.653840 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-9s7hk" event={"ID":"c3d01029-07d5-414c-badd-0d28f0db9730","Type":"ContainerDied","Data":"ba381c08a9b5a6b16ca31697476df256d350bd3f9823c42f93ec2e30ab086477"} Feb 27 17:21:04 crc kubenswrapper[4700]: I0227 17:21:04.654109 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba381c08a9b5a6b16ca31697476df256d350bd3f9823c42f93ec2e30ab086477" Feb 27 17:21:04 crc kubenswrapper[4700]: I0227 17:21:04.653953 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-9s7hk" Feb 27 17:21:04 crc kubenswrapper[4700]: I0227 17:21:04.708926 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-687c5777ff-vx8jc" podStartSLOduration=4.708898113 podStartE2EDuration="4.708898113s" podCreationTimestamp="2026-02-27 17:21:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:21:04.684791585 +0000 UTC m=+1224.670104362" watchObservedRunningTime="2026-02-27 17:21:04.708898113 +0000 UTC m=+1224.694210910" Feb 27 17:21:05 crc kubenswrapper[4700]: I0227 17:21:05.150772 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-687c5777ff-vx8jc"] Feb 27 17:21:05 crc kubenswrapper[4700]: I0227 17:21:05.178665 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cb78597c7-wtsvw"] Feb 27 17:21:05 crc kubenswrapper[4700]: E0227 17:21:05.178991 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3d01029-07d5-414c-badd-0d28f0db9730" containerName="glance-db-sync" Feb 27 17:21:05 crc kubenswrapper[4700]: I0227 17:21:05.179008 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3d01029-07d5-414c-badd-0d28f0db9730" containerName="glance-db-sync" Feb 27 17:21:05 crc kubenswrapper[4700]: I0227 17:21:05.179171 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3d01029-07d5-414c-badd-0d28f0db9730" containerName="glance-db-sync" Feb 27 17:21:05 crc kubenswrapper[4700]: I0227 17:21:05.179993 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb78597c7-wtsvw" Feb 27 17:21:05 crc kubenswrapper[4700]: I0227 17:21:05.256780 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a902211b-c06e-42af-a1c8-599ab09fbd23-dns-swift-storage-0\") pod \"dnsmasq-dns-7cb78597c7-wtsvw\" (UID: \"a902211b-c06e-42af-a1c8-599ab09fbd23\") " pod="openstack/dnsmasq-dns-7cb78597c7-wtsvw" Feb 27 17:21:05 crc kubenswrapper[4700]: I0227 17:21:05.257020 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a902211b-c06e-42af-a1c8-599ab09fbd23-ovsdbserver-nb\") pod \"dnsmasq-dns-7cb78597c7-wtsvw\" (UID: \"a902211b-c06e-42af-a1c8-599ab09fbd23\") " pod="openstack/dnsmasq-dns-7cb78597c7-wtsvw" Feb 27 17:21:05 crc kubenswrapper[4700]: I0227 17:21:05.257176 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a902211b-c06e-42af-a1c8-599ab09fbd23-dns-svc\") pod \"dnsmasq-dns-7cb78597c7-wtsvw\" (UID: \"a902211b-c06e-42af-a1c8-599ab09fbd23\") " pod="openstack/dnsmasq-dns-7cb78597c7-wtsvw" Feb 27 17:21:05 crc kubenswrapper[4700]: I0227 17:21:05.257278 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sntls\" (UniqueName: \"kubernetes.io/projected/a902211b-c06e-42af-a1c8-599ab09fbd23-kube-api-access-sntls\") pod \"dnsmasq-dns-7cb78597c7-wtsvw\" (UID: \"a902211b-c06e-42af-a1c8-599ab09fbd23\") " pod="openstack/dnsmasq-dns-7cb78597c7-wtsvw" Feb 27 17:21:05 crc kubenswrapper[4700]: I0227 17:21:05.257347 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a902211b-c06e-42af-a1c8-599ab09fbd23-ovsdbserver-sb\") pod \"dnsmasq-dns-7cb78597c7-wtsvw\" (UID: \"a902211b-c06e-42af-a1c8-599ab09fbd23\") " pod="openstack/dnsmasq-dns-7cb78597c7-wtsvw" Feb 27 17:21:05 crc kubenswrapper[4700]: I0227 17:21:05.257431 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a902211b-c06e-42af-a1c8-599ab09fbd23-config\") pod \"dnsmasq-dns-7cb78597c7-wtsvw\" (UID: \"a902211b-c06e-42af-a1c8-599ab09fbd23\") " pod="openstack/dnsmasq-dns-7cb78597c7-wtsvw" Feb 27 17:21:05 crc kubenswrapper[4700]: I0227 17:21:05.262060 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb78597c7-wtsvw"] Feb 27 17:21:05 crc kubenswrapper[4700]: I0227 17:21:05.358598 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sntls\" (UniqueName: \"kubernetes.io/projected/a902211b-c06e-42af-a1c8-599ab09fbd23-kube-api-access-sntls\") pod \"dnsmasq-dns-7cb78597c7-wtsvw\" (UID: \"a902211b-c06e-42af-a1c8-599ab09fbd23\") " pod="openstack/dnsmasq-dns-7cb78597c7-wtsvw" Feb 27 17:21:05 crc kubenswrapper[4700]: I0227 17:21:05.358640 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a902211b-c06e-42af-a1c8-599ab09fbd23-ovsdbserver-sb\") pod \"dnsmasq-dns-7cb78597c7-wtsvw\" (UID: \"a902211b-c06e-42af-a1c8-599ab09fbd23\") " pod="openstack/dnsmasq-dns-7cb78597c7-wtsvw" Feb 27 17:21:05 crc kubenswrapper[4700]: I0227 17:21:05.358672 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a902211b-c06e-42af-a1c8-599ab09fbd23-config\") pod \"dnsmasq-dns-7cb78597c7-wtsvw\" (UID: \"a902211b-c06e-42af-a1c8-599ab09fbd23\") " pod="openstack/dnsmasq-dns-7cb78597c7-wtsvw" Feb 27 17:21:05 crc kubenswrapper[4700]: I0227 17:21:05.358709 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a902211b-c06e-42af-a1c8-599ab09fbd23-dns-swift-storage-0\") pod \"dnsmasq-dns-7cb78597c7-wtsvw\" (UID: \"a902211b-c06e-42af-a1c8-599ab09fbd23\") " pod="openstack/dnsmasq-dns-7cb78597c7-wtsvw" Feb 27 17:21:05 crc kubenswrapper[4700]: I0227 17:21:05.358741 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a902211b-c06e-42af-a1c8-599ab09fbd23-ovsdbserver-nb\") pod \"dnsmasq-dns-7cb78597c7-wtsvw\" (UID: \"a902211b-c06e-42af-a1c8-599ab09fbd23\") " pod="openstack/dnsmasq-dns-7cb78597c7-wtsvw" Feb 27 17:21:05 crc kubenswrapper[4700]: I0227 17:21:05.358799 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a902211b-c06e-42af-a1c8-599ab09fbd23-dns-svc\") pod \"dnsmasq-dns-7cb78597c7-wtsvw\" (UID: \"a902211b-c06e-42af-a1c8-599ab09fbd23\") " pod="openstack/dnsmasq-dns-7cb78597c7-wtsvw" Feb 27 17:21:05 crc kubenswrapper[4700]: I0227 17:21:05.359554 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a902211b-c06e-42af-a1c8-599ab09fbd23-dns-svc\") pod \"dnsmasq-dns-7cb78597c7-wtsvw\" (UID: \"a902211b-c06e-42af-a1c8-599ab09fbd23\") " pod="openstack/dnsmasq-dns-7cb78597c7-wtsvw" Feb 27 17:21:05 crc kubenswrapper[4700]: I0227 17:21:05.360274 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a902211b-c06e-42af-a1c8-599ab09fbd23-ovsdbserver-sb\") pod \"dnsmasq-dns-7cb78597c7-wtsvw\" (UID: \"a902211b-c06e-42af-a1c8-599ab09fbd23\") " pod="openstack/dnsmasq-dns-7cb78597c7-wtsvw" Feb 27 17:21:05 crc kubenswrapper[4700]: I0227 17:21:05.360784 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a902211b-c06e-42af-a1c8-599ab09fbd23-config\") pod \"dnsmasq-dns-7cb78597c7-wtsvw\" (UID: \"a902211b-c06e-42af-a1c8-599ab09fbd23\") " pod="openstack/dnsmasq-dns-7cb78597c7-wtsvw" Feb 27 17:21:05 crc kubenswrapper[4700]: I0227 17:21:05.361598 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a902211b-c06e-42af-a1c8-599ab09fbd23-dns-swift-storage-0\") pod \"dnsmasq-dns-7cb78597c7-wtsvw\" (UID: \"a902211b-c06e-42af-a1c8-599ab09fbd23\") " pod="openstack/dnsmasq-dns-7cb78597c7-wtsvw" Feb 27 17:21:05 crc kubenswrapper[4700]: I0227 17:21:05.362088 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a902211b-c06e-42af-a1c8-599ab09fbd23-ovsdbserver-nb\") pod \"dnsmasq-dns-7cb78597c7-wtsvw\" (UID: \"a902211b-c06e-42af-a1c8-599ab09fbd23\") " pod="openstack/dnsmasq-dns-7cb78597c7-wtsvw" Feb 27 17:21:05 crc kubenswrapper[4700]: I0227 17:21:05.380192 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sntls\" (UniqueName: \"kubernetes.io/projected/a902211b-c06e-42af-a1c8-599ab09fbd23-kube-api-access-sntls\") pod \"dnsmasq-dns-7cb78597c7-wtsvw\" (UID: \"a902211b-c06e-42af-a1c8-599ab09fbd23\") " pod="openstack/dnsmasq-dns-7cb78597c7-wtsvw" Feb 27 17:21:05 crc kubenswrapper[4700]: I0227 17:21:05.491835 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb78597c7-wtsvw" Feb 27 17:21:05 crc kubenswrapper[4700]: W0227 17:21:05.920048 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda902211b_c06e_42af_a1c8_599ab09fbd23.slice/crio-697573a9dd246961321407de070eb79aa12d2c399c7054ca5394cc74119e2ae5 WatchSource:0}: Error finding container 697573a9dd246961321407de070eb79aa12d2c399c7054ca5394cc74119e2ae5: Status 404 returned error can't find the container with id 697573a9dd246961321407de070eb79aa12d2c399c7054ca5394cc74119e2ae5 Feb 27 17:21:05 crc kubenswrapper[4700]: I0227 17:21:05.920946 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb78597c7-wtsvw"] Feb 27 17:21:06 crc kubenswrapper[4700]: I0227 17:21:06.410742 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:21:06 crc kubenswrapper[4700]: I0227 17:21:06.410981 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:21:06 crc kubenswrapper[4700]: I0227 17:21:06.675733 4700 generic.go:334] "Generic (PLEG): container finished" podID="a902211b-c06e-42af-a1c8-599ab09fbd23" containerID="c825731a613b3749d9fee537385612f6b55244071f3b74ed1a4fab210445940c" exitCode=0 Feb 27 17:21:06 crc kubenswrapper[4700]: I0227 17:21:06.675781 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb78597c7-wtsvw" event={"ID":"a902211b-c06e-42af-a1c8-599ab09fbd23","Type":"ContainerDied","Data":"c825731a613b3749d9fee537385612f6b55244071f3b74ed1a4fab210445940c"} Feb 27 17:21:06 crc kubenswrapper[4700]: I0227 17:21:06.675830 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb78597c7-wtsvw" event={"ID":"a902211b-c06e-42af-a1c8-599ab09fbd23","Type":"ContainerStarted","Data":"697573a9dd246961321407de070eb79aa12d2c399c7054ca5394cc74119e2ae5"} Feb 27 17:21:06 crc kubenswrapper[4700]: I0227 17:21:06.675932 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-687c5777ff-vx8jc" podUID="f455d2a5-6e87-4287-a00e-83f724dd180c" containerName="dnsmasq-dns" containerID="cri-o://fe14e67a4dd9ee316ed15ed83bfdbe18c56cf274127121742b785da0b17d0c16" gracePeriod=10 Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.091698 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.128666 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-687c5777ff-vx8jc" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.189958 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f455d2a5-6e87-4287-a00e-83f724dd180c-ovsdbserver-nb\") pod \"f455d2a5-6e87-4287-a00e-83f724dd180c\" (UID: \"f455d2a5-6e87-4287-a00e-83f724dd180c\") " Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.190099 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkxr6\" (UniqueName: \"kubernetes.io/projected/f455d2a5-6e87-4287-a00e-83f724dd180c-kube-api-access-dkxr6\") pod \"f455d2a5-6e87-4287-a00e-83f724dd180c\" (UID: \"f455d2a5-6e87-4287-a00e-83f724dd180c\") " Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.190129 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f455d2a5-6e87-4287-a00e-83f724dd180c-dns-svc\") pod \"f455d2a5-6e87-4287-a00e-83f724dd180c\" (UID: \"f455d2a5-6e87-4287-a00e-83f724dd180c\") " Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.190172 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f455d2a5-6e87-4287-a00e-83f724dd180c-ovsdbserver-sb\") pod \"f455d2a5-6e87-4287-a00e-83f724dd180c\" (UID: \"f455d2a5-6e87-4287-a00e-83f724dd180c\") " Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.190234 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f455d2a5-6e87-4287-a00e-83f724dd180c-config\") pod \"f455d2a5-6e87-4287-a00e-83f724dd180c\" (UID: \"f455d2a5-6e87-4287-a00e-83f724dd180c\") " Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.190278 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f455d2a5-6e87-4287-a00e-83f724dd180c-dns-swift-storage-0\") pod \"f455d2a5-6e87-4287-a00e-83f724dd180c\" (UID: \"f455d2a5-6e87-4287-a00e-83f724dd180c\") " Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.209655 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f455d2a5-6e87-4287-a00e-83f724dd180c-kube-api-access-dkxr6" (OuterVolumeSpecName: "kube-api-access-dkxr6") pod "f455d2a5-6e87-4287-a00e-83f724dd180c" (UID: "f455d2a5-6e87-4287-a00e-83f724dd180c"). InnerVolumeSpecName "kube-api-access-dkxr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.261720 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f455d2a5-6e87-4287-a00e-83f724dd180c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f455d2a5-6e87-4287-a00e-83f724dd180c" (UID: "f455d2a5-6e87-4287-a00e-83f724dd180c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.262482 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f455d2a5-6e87-4287-a00e-83f724dd180c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f455d2a5-6e87-4287-a00e-83f724dd180c" (UID: "f455d2a5-6e87-4287-a00e-83f724dd180c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.271872 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f455d2a5-6e87-4287-a00e-83f724dd180c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f455d2a5-6e87-4287-a00e-83f724dd180c" (UID: "f455d2a5-6e87-4287-a00e-83f724dd180c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.274415 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f455d2a5-6e87-4287-a00e-83f724dd180c-config" (OuterVolumeSpecName: "config") pod "f455d2a5-6e87-4287-a00e-83f724dd180c" (UID: "f455d2a5-6e87-4287-a00e-83f724dd180c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.292168 4700 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f455d2a5-6e87-4287-a00e-83f724dd180c-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.292198 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f455d2a5-6e87-4287-a00e-83f724dd180c-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.292207 4700 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f455d2a5-6e87-4287-a00e-83f724dd180c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.292229 4700 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f455d2a5-6e87-4287-a00e-83f724dd180c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.292240 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkxr6\" (UniqueName: \"kubernetes.io/projected/f455d2a5-6e87-4287-a00e-83f724dd180c-kube-api-access-dkxr6\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.297551 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f455d2a5-6e87-4287-a00e-83f724dd180c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f455d2a5-6e87-4287-a00e-83f724dd180c" (UID: "f455d2a5-6e87-4287-a00e-83f724dd180c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.385659 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/notifications-rabbitmq-server-0" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.393295 4700 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f455d2a5-6e87-4287-a00e-83f724dd180c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.605828 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-7xvrx"] Feb 27 17:21:07 crc kubenswrapper[4700]: E0227 17:21:07.606158 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f455d2a5-6e87-4287-a00e-83f724dd180c" containerName="dnsmasq-dns" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.606174 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="f455d2a5-6e87-4287-a00e-83f724dd180c" containerName="dnsmasq-dns" Feb 27 17:21:07 crc kubenswrapper[4700]: E0227 17:21:07.606202 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f455d2a5-6e87-4287-a00e-83f724dd180c" containerName="init" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.606209 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="f455d2a5-6e87-4287-a00e-83f724dd180c" containerName="init" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.606361 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="f455d2a5-6e87-4287-a00e-83f724dd180c" containerName="dnsmasq-dns" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.606890 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-7xvrx" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.622477 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-7xvrx"] Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.688189 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb78597c7-wtsvw" event={"ID":"a902211b-c06e-42af-a1c8-599ab09fbd23","Type":"ContainerStarted","Data":"a2577f97fd441ece5be682f8b2be1537ddf7b8c6658c22448f59f2d3be4b41da"} Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.688378 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7cb78597c7-wtsvw" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.690853 4700 generic.go:334] "Generic (PLEG): container finished" podID="f455d2a5-6e87-4287-a00e-83f724dd180c" containerID="fe14e67a4dd9ee316ed15ed83bfdbe18c56cf274127121742b785da0b17d0c16" exitCode=0 Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.690900 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-687c5777ff-vx8jc" event={"ID":"f455d2a5-6e87-4287-a00e-83f724dd180c","Type":"ContainerDied","Data":"fe14e67a4dd9ee316ed15ed83bfdbe18c56cf274127121742b785da0b17d0c16"} Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.690915 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-687c5777ff-vx8jc" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.690929 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-687c5777ff-vx8jc" event={"ID":"f455d2a5-6e87-4287-a00e-83f724dd180c","Type":"ContainerDied","Data":"1e731afb9fba9bc466ee7e696586017bd40620ff1196b0fa0e6a4600de02eb23"} Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.690947 4700 scope.go:117] "RemoveContainer" containerID="fe14e67a4dd9ee316ed15ed83bfdbe18c56cf274127121742b785da0b17d0c16" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.697600 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2mwk\" (UniqueName: \"kubernetes.io/projected/f5d89c26-6133-4987-8c61-cc19c80567e0-kube-api-access-n2mwk\") pod \"barbican-db-create-7xvrx\" (UID: \"f5d89c26-6133-4987-8c61-cc19c80567e0\") " pod="openstack/barbican-db-create-7xvrx" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.697641 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5d89c26-6133-4987-8c61-cc19c80567e0-operator-scripts\") pod \"barbican-db-create-7xvrx\" (UID: \"f5d89c26-6133-4987-8c61-cc19c80567e0\") " pod="openstack/barbican-db-create-7xvrx" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.706861 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.718256 4700 scope.go:117] "RemoveContainer" containerID="2380b8d90f49e4e6906e0ad15caa0ab116899dbec6d1a1d1077a82bb95768c07" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.728696 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-1943-account-create-update-h8f9l"] Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.730219 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-1943-account-create-update-h8f9l" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.733371 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7cb78597c7-wtsvw" podStartSLOduration=2.733349576 podStartE2EDuration="2.733349576s" podCreationTimestamp="2026-02-27 17:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:21:07.720424764 +0000 UTC m=+1227.705737511" watchObservedRunningTime="2026-02-27 17:21:07.733349576 +0000 UTC m=+1227.718662323" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.735721 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.765102 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-1943-account-create-update-h8f9l"] Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.776177 4700 scope.go:117] "RemoveContainer" containerID="fe14e67a4dd9ee316ed15ed83bfdbe18c56cf274127121742b785da0b17d0c16" Feb 27 17:21:07 crc kubenswrapper[4700]: E0227 17:21:07.776584 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe14e67a4dd9ee316ed15ed83bfdbe18c56cf274127121742b785da0b17d0c16\": container with ID starting with fe14e67a4dd9ee316ed15ed83bfdbe18c56cf274127121742b785da0b17d0c16 not found: ID does not exist" containerID="fe14e67a4dd9ee316ed15ed83bfdbe18c56cf274127121742b785da0b17d0c16" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.776614 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe14e67a4dd9ee316ed15ed83bfdbe18c56cf274127121742b785da0b17d0c16"} err="failed to get container status \"fe14e67a4dd9ee316ed15ed83bfdbe18c56cf274127121742b785da0b17d0c16\": rpc error: code = NotFound desc = could not find container \"fe14e67a4dd9ee316ed15ed83bfdbe18c56cf274127121742b785da0b17d0c16\": container with ID starting with fe14e67a4dd9ee316ed15ed83bfdbe18c56cf274127121742b785da0b17d0c16 not found: ID does not exist" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.776633 4700 scope.go:117] "RemoveContainer" containerID="2380b8d90f49e4e6906e0ad15caa0ab116899dbec6d1a1d1077a82bb95768c07" Feb 27 17:21:07 crc kubenswrapper[4700]: E0227 17:21:07.776969 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2380b8d90f49e4e6906e0ad15caa0ab116899dbec6d1a1d1077a82bb95768c07\": container with ID starting with 2380b8d90f49e4e6906e0ad15caa0ab116899dbec6d1a1d1077a82bb95768c07 not found: ID does not exist" containerID="2380b8d90f49e4e6906e0ad15caa0ab116899dbec6d1a1d1077a82bb95768c07" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.776994 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2380b8d90f49e4e6906e0ad15caa0ab116899dbec6d1a1d1077a82bb95768c07"} err="failed to get container status \"2380b8d90f49e4e6906e0ad15caa0ab116899dbec6d1a1d1077a82bb95768c07\": rpc error: code = NotFound desc = could not find container \"2380b8d90f49e4e6906e0ad15caa0ab116899dbec6d1a1d1077a82bb95768c07\": container with ID starting with 2380b8d90f49e4e6906e0ad15caa0ab116899dbec6d1a1d1077a82bb95768c07 not found: ID does not exist" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.793723 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-687c5777ff-vx8jc"] Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.806352 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2mwk\" (UniqueName: \"kubernetes.io/projected/f5d89c26-6133-4987-8c61-cc19c80567e0-kube-api-access-n2mwk\") pod \"barbican-db-create-7xvrx\" (UID: \"f5d89c26-6133-4987-8c61-cc19c80567e0\") " pod="openstack/barbican-db-create-7xvrx" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.806404 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5d89c26-6133-4987-8c61-cc19c80567e0-operator-scripts\") pod \"barbican-db-create-7xvrx\" (UID: \"f5d89c26-6133-4987-8c61-cc19c80567e0\") " pod="openstack/barbican-db-create-7xvrx" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.809087 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5d89c26-6133-4987-8c61-cc19c80567e0-operator-scripts\") pod \"barbican-db-create-7xvrx\" (UID: \"f5d89c26-6133-4987-8c61-cc19c80567e0\") " pod="openstack/barbican-db-create-7xvrx" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.810267 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-687c5777ff-vx8jc"] Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.832065 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-rc9w6"] Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.833133 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-rc9w6" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.845779 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2mwk\" (UniqueName: \"kubernetes.io/projected/f5d89c26-6133-4987-8c61-cc19c80567e0-kube-api-access-n2mwk\") pod \"barbican-db-create-7xvrx\" (UID: \"f5d89c26-6133-4987-8c61-cc19c80567e0\") " pod="openstack/barbican-db-create-7xvrx" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.857258 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-2c01-account-create-update-crm9x"] Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.858359 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-2c01-account-create-update-crm9x" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.861655 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.878579 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-rc9w6"] Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.899193 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-2c01-account-create-update-crm9x"] Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.911203 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnrxk\" (UniqueName: \"kubernetes.io/projected/2062608c-6759-45bd-9879-2a96af693b31-kube-api-access-cnrxk\") pod \"cinder-1943-account-create-update-h8f9l\" (UID: \"2062608c-6759-45bd-9879-2a96af693b31\") " pod="openstack/cinder-1943-account-create-update-h8f9l" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.911318 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e67eb82-300d-46bd-b1c6-d2607bf45ccc-operator-scripts\") pod \"cinder-db-create-rc9w6\" (UID: \"2e67eb82-300d-46bd-b1c6-d2607bf45ccc\") " pod="openstack/cinder-db-create-rc9w6" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.911531 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2062608c-6759-45bd-9879-2a96af693b31-operator-scripts\") pod \"cinder-1943-account-create-update-h8f9l\" (UID: \"2062608c-6759-45bd-9879-2a96af693b31\") " pod="openstack/cinder-1943-account-create-update-h8f9l" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.911568 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtjdq\" (UniqueName: \"kubernetes.io/projected/2e67eb82-300d-46bd-b1c6-d2607bf45ccc-kube-api-access-rtjdq\") pod \"cinder-db-create-rc9w6\" (UID: \"2e67eb82-300d-46bd-b1c6-d2607bf45ccc\") " pod="openstack/cinder-db-create-rc9w6" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.923668 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-qm46k"] Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.924902 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-qm46k" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.927137 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-7xvrx" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.929422 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.929684 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-llljd" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.930869 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.931337 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 27 17:21:07 crc kubenswrapper[4700]: I0227 17:21:07.934276 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-qm46k"] Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.013334 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnrxk\" (UniqueName: \"kubernetes.io/projected/2062608c-6759-45bd-9879-2a96af693b31-kube-api-access-cnrxk\") pod \"cinder-1943-account-create-update-h8f9l\" (UID: \"2062608c-6759-45bd-9879-2a96af693b31\") " pod="openstack/cinder-1943-account-create-update-h8f9l" Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.013651 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e67eb82-300d-46bd-b1c6-d2607bf45ccc-operator-scripts\") pod \"cinder-db-create-rc9w6\" (UID: \"2e67eb82-300d-46bd-b1c6-d2607bf45ccc\") " pod="openstack/cinder-db-create-rc9w6" Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.013704 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe426647-f5e6-46bf-b187-8f8cb62b06e1-operator-scripts\") pod \"barbican-2c01-account-create-update-crm9x\" (UID: \"fe426647-f5e6-46bf-b187-8f8cb62b06e1\") " pod="openstack/barbican-2c01-account-create-update-crm9x" Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.013732 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2062608c-6759-45bd-9879-2a96af693b31-operator-scripts\") pod \"cinder-1943-account-create-update-h8f9l\" (UID: \"2062608c-6759-45bd-9879-2a96af693b31\") " pod="openstack/cinder-1943-account-create-update-h8f9l" Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.013759 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtjdq\" (UniqueName: \"kubernetes.io/projected/2e67eb82-300d-46bd-b1c6-d2607bf45ccc-kube-api-access-rtjdq\") pod \"cinder-db-create-rc9w6\" (UID: \"2e67eb82-300d-46bd-b1c6-d2607bf45ccc\") " pod="openstack/cinder-db-create-rc9w6" Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.013796 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65cmp\" (UniqueName: \"kubernetes.io/projected/fe426647-f5e6-46bf-b187-8f8cb62b06e1-kube-api-access-65cmp\") pod \"barbican-2c01-account-create-update-crm9x\" (UID: \"fe426647-f5e6-46bf-b187-8f8cb62b06e1\") " pod="openstack/barbican-2c01-account-create-update-crm9x" Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.014532 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2062608c-6759-45bd-9879-2a96af693b31-operator-scripts\") pod \"cinder-1943-account-create-update-h8f9l\" (UID: \"2062608c-6759-45bd-9879-2a96af693b31\") " pod="openstack/cinder-1943-account-create-update-h8f9l" Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.014722 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e67eb82-300d-46bd-b1c6-d2607bf45ccc-operator-scripts\") pod \"cinder-db-create-rc9w6\" (UID: \"2e67eb82-300d-46bd-b1c6-d2607bf45ccc\") " pod="openstack/cinder-db-create-rc9w6" Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.033946 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnrxk\" (UniqueName: \"kubernetes.io/projected/2062608c-6759-45bd-9879-2a96af693b31-kube-api-access-cnrxk\") pod \"cinder-1943-account-create-update-h8f9l\" (UID: \"2062608c-6759-45bd-9879-2a96af693b31\") " pod="openstack/cinder-1943-account-create-update-h8f9l" Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.041029 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtjdq\" (UniqueName: \"kubernetes.io/projected/2e67eb82-300d-46bd-b1c6-d2607bf45ccc-kube-api-access-rtjdq\") pod \"cinder-db-create-rc9w6\" (UID: \"2e67eb82-300d-46bd-b1c6-d2607bf45ccc\") " pod="openstack/cinder-db-create-rc9w6" Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.066682 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-1943-account-create-update-h8f9l" Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.115105 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g48ql\" (UniqueName: \"kubernetes.io/projected/a238beaf-451f-452e-a30f-09779ae217c1-kube-api-access-g48ql\") pod \"keystone-db-sync-qm46k\" (UID: \"a238beaf-451f-452e-a30f-09779ae217c1\") " pod="openstack/keystone-db-sync-qm46k" Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.115229 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe426647-f5e6-46bf-b187-8f8cb62b06e1-operator-scripts\") pod \"barbican-2c01-account-create-update-crm9x\" (UID: \"fe426647-f5e6-46bf-b187-8f8cb62b06e1\") " pod="openstack/barbican-2c01-account-create-update-crm9x" Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.115280 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a238beaf-451f-452e-a30f-09779ae217c1-config-data\") pod \"keystone-db-sync-qm46k\" (UID: \"a238beaf-451f-452e-a30f-09779ae217c1\") " pod="openstack/keystone-db-sync-qm46k" Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.115343 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65cmp\" (UniqueName: \"kubernetes.io/projected/fe426647-f5e6-46bf-b187-8f8cb62b06e1-kube-api-access-65cmp\") pod \"barbican-2c01-account-create-update-crm9x\" (UID: \"fe426647-f5e6-46bf-b187-8f8cb62b06e1\") " pod="openstack/barbican-2c01-account-create-update-crm9x" Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.115387 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a238beaf-451f-452e-a30f-09779ae217c1-combined-ca-bundle\") pod \"keystone-db-sync-qm46k\" (UID: \"a238beaf-451f-452e-a30f-09779ae217c1\") " pod="openstack/keystone-db-sync-qm46k" Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.116237 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe426647-f5e6-46bf-b187-8f8cb62b06e1-operator-scripts\") pod \"barbican-2c01-account-create-update-crm9x\" (UID: \"fe426647-f5e6-46bf-b187-8f8cb62b06e1\") " pod="openstack/barbican-2c01-account-create-update-crm9x" Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.133871 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65cmp\" (UniqueName: \"kubernetes.io/projected/fe426647-f5e6-46bf-b187-8f8cb62b06e1-kube-api-access-65cmp\") pod \"barbican-2c01-account-create-update-crm9x\" (UID: \"fe426647-f5e6-46bf-b187-8f8cb62b06e1\") " pod="openstack/barbican-2c01-account-create-update-crm9x" Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.151909 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-rc9w6" Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.186236 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-2c01-account-create-update-crm9x" Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.216408 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a238beaf-451f-452e-a30f-09779ae217c1-combined-ca-bundle\") pod \"keystone-db-sync-qm46k\" (UID: \"a238beaf-451f-452e-a30f-09779ae217c1\") " pod="openstack/keystone-db-sync-qm46k" Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.216487 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g48ql\" (UniqueName: \"kubernetes.io/projected/a238beaf-451f-452e-a30f-09779ae217c1-kube-api-access-g48ql\") pod \"keystone-db-sync-qm46k\" (UID: \"a238beaf-451f-452e-a30f-09779ae217c1\") " pod="openstack/keystone-db-sync-qm46k" Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.216553 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a238beaf-451f-452e-a30f-09779ae217c1-config-data\") pod \"keystone-db-sync-qm46k\" (UID: \"a238beaf-451f-452e-a30f-09779ae217c1\") " pod="openstack/keystone-db-sync-qm46k" Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.220239 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a238beaf-451f-452e-a30f-09779ae217c1-config-data\") pod \"keystone-db-sync-qm46k\" (UID: \"a238beaf-451f-452e-a30f-09779ae217c1\") " pod="openstack/keystone-db-sync-qm46k" Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.223331 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a238beaf-451f-452e-a30f-09779ae217c1-combined-ca-bundle\") pod \"keystone-db-sync-qm46k\" (UID: \"a238beaf-451f-452e-a30f-09779ae217c1\") " pod="openstack/keystone-db-sync-qm46k" Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.234940 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g48ql\" (UniqueName: \"kubernetes.io/projected/a238beaf-451f-452e-a30f-09779ae217c1-kube-api-access-g48ql\") pod \"keystone-db-sync-qm46k\" (UID: \"a238beaf-451f-452e-a30f-09779ae217c1\") " pod="openstack/keystone-db-sync-qm46k" Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.312215 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-qm46k" Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.370907 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-7xvrx"] Feb 27 17:21:08 crc kubenswrapper[4700]: W0227 17:21:08.381504 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf5d89c26_6133_4987_8c61_cc19c80567e0.slice/crio-a62625ee5f2cd1e901acd226e6f8244bf559509a78c0aa4edd3943fe7be19d75 WatchSource:0}: Error finding container a62625ee5f2cd1e901acd226e6f8244bf559509a78c0aa4edd3943fe7be19d75: Status 404 returned error can't find the container with id a62625ee5f2cd1e901acd226e6f8244bf559509a78c0aa4edd3943fe7be19d75 Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.407513 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-rc9w6"] Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.517540 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-1943-account-create-update-h8f9l"] Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.703383 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-2c01-account-create-update-crm9x"] Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.703852 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-7xvrx" event={"ID":"f5d89c26-6133-4987-8c61-cc19c80567e0","Type":"ContainerStarted","Data":"a1755da87ed07a85b745e9409058de03e937db1f3700433133f4bcecfb9f0aa4"} Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.703893 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-7xvrx" event={"ID":"f5d89c26-6133-4987-8c61-cc19c80567e0","Type":"ContainerStarted","Data":"a62625ee5f2cd1e901acd226e6f8244bf559509a78c0aa4edd3943fe7be19d75"} Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.706005 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-rc9w6" event={"ID":"2e67eb82-300d-46bd-b1c6-d2607bf45ccc","Type":"ContainerStarted","Data":"c743a30c2358f798b11a6ad8bb7d06f0869f65ded2d52c2eb9d11e6bbafa0117"} Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.706047 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-rc9w6" event={"ID":"2e67eb82-300d-46bd-b1c6-d2607bf45ccc","Type":"ContainerStarted","Data":"d9f96c9b6146e2f3f597814ee423a55d04f4cea8c25e6543d63295aae0bfc9a3"} Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.710772 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-1943-account-create-update-h8f9l" event={"ID":"2062608c-6759-45bd-9879-2a96af693b31","Type":"ContainerStarted","Data":"4b7f93b578e644781c33d2b3aa82faa1979c30b013b09b4efe7e74cb9564891a"} Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.710819 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-1943-account-create-update-h8f9l" event={"ID":"2062608c-6759-45bd-9879-2a96af693b31","Type":"ContainerStarted","Data":"b1a42e648c08ada3917bf8cba48dd60e776820d7dc1109498df2addc1840ad15"} Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.736758 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-7xvrx" podStartSLOduration=1.736738167 podStartE2EDuration="1.736738167s" podCreationTimestamp="2026-02-27 17:21:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:21:08.730633635 +0000 UTC m=+1228.715946382" watchObservedRunningTime="2026-02-27 17:21:08.736738167 +0000 UTC m=+1228.722050914" Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.766344 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-rc9w6" podStartSLOduration=1.7663264889999999 podStartE2EDuration="1.766326489s" podCreationTimestamp="2026-02-27 17:21:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:21:08.760832314 +0000 UTC m=+1228.746145061" watchObservedRunningTime="2026-02-27 17:21:08.766326489 +0000 UTC m=+1228.751639236" Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.791754 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-qm46k"] Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.804058 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-1943-account-create-update-h8f9l" podStartSLOduration=1.804036917 podStartE2EDuration="1.804036917s" podCreationTimestamp="2026-02-27 17:21:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:21:08.782577089 +0000 UTC m=+1228.767889836" watchObservedRunningTime="2026-02-27 17:21:08.804036917 +0000 UTC m=+1228.789349664" Feb 27 17:21:08 crc kubenswrapper[4700]: W0227 17:21:08.805145 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda238beaf_451f_452e_a30f_09779ae217c1.slice/crio-bc6714991554d7aa8a5b503a986c34630d41c3edb6040160ee068dc0e8888704 WatchSource:0}: Error finding container bc6714991554d7aa8a5b503a986c34630d41c3edb6040160ee068dc0e8888704: Status 404 returned error can't find the container with id bc6714991554d7aa8a5b503a986c34630d41c3edb6040160ee068dc0e8888704 Feb 27 17:21:08 crc kubenswrapper[4700]: I0227 17:21:08.990576 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f455d2a5-6e87-4287-a00e-83f724dd180c" path="/var/lib/kubelet/pods/f455d2a5-6e87-4287-a00e-83f724dd180c/volumes" Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.039274 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-db-sync-5zzh7"] Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.040786 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-5zzh7" Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.043898 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-config-data" Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.044144 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-watcher-dockercfg-qc4hm" Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.057086 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-sync-5zzh7"] Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.145476 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16189a3f-0a2b-453c-8bb1-8e8c4d79c13f-combined-ca-bundle\") pod \"watcher-db-sync-5zzh7\" (UID: \"16189a3f-0a2b-453c-8bb1-8e8c4d79c13f\") " pod="openstack/watcher-db-sync-5zzh7" Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.145535 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/16189a3f-0a2b-453c-8bb1-8e8c4d79c13f-db-sync-config-data\") pod \"watcher-db-sync-5zzh7\" (UID: \"16189a3f-0a2b-453c-8bb1-8e8c4d79c13f\") " pod="openstack/watcher-db-sync-5zzh7" Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.145626 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16189a3f-0a2b-453c-8bb1-8e8c4d79c13f-config-data\") pod \"watcher-db-sync-5zzh7\" (UID: \"16189a3f-0a2b-453c-8bb1-8e8c4d79c13f\") " pod="openstack/watcher-db-sync-5zzh7" Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.145656 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4scl\" (UniqueName: \"kubernetes.io/projected/16189a3f-0a2b-453c-8bb1-8e8c4d79c13f-kube-api-access-z4scl\") pod \"watcher-db-sync-5zzh7\" (UID: \"16189a3f-0a2b-453c-8bb1-8e8c4d79c13f\") " pod="openstack/watcher-db-sync-5zzh7" Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.191941 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-f5xkq"] Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.193138 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-f5xkq" Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.202681 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-f5xkq"] Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.247003 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16189a3f-0a2b-453c-8bb1-8e8c4d79c13f-combined-ca-bundle\") pod \"watcher-db-sync-5zzh7\" (UID: \"16189a3f-0a2b-453c-8bb1-8e8c4d79c13f\") " pod="openstack/watcher-db-sync-5zzh7" Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.247066 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/16189a3f-0a2b-453c-8bb1-8e8c4d79c13f-db-sync-config-data\") pod \"watcher-db-sync-5zzh7\" (UID: \"16189a3f-0a2b-453c-8bb1-8e8c4d79c13f\") " pod="openstack/watcher-db-sync-5zzh7" Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.247142 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16189a3f-0a2b-453c-8bb1-8e8c4d79c13f-config-data\") pod \"watcher-db-sync-5zzh7\" (UID: \"16189a3f-0a2b-453c-8bb1-8e8c4d79c13f\") " pod="openstack/watcher-db-sync-5zzh7" Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.247173 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4scl\" (UniqueName: \"kubernetes.io/projected/16189a3f-0a2b-453c-8bb1-8e8c4d79c13f-kube-api-access-z4scl\") pod \"watcher-db-sync-5zzh7\" (UID: \"16189a3f-0a2b-453c-8bb1-8e8c4d79c13f\") " pod="openstack/watcher-db-sync-5zzh7" Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.253188 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16189a3f-0a2b-453c-8bb1-8e8c4d79c13f-combined-ca-bundle\") pod \"watcher-db-sync-5zzh7\" (UID: \"16189a3f-0a2b-453c-8bb1-8e8c4d79c13f\") " pod="openstack/watcher-db-sync-5zzh7" Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.261872 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16189a3f-0a2b-453c-8bb1-8e8c4d79c13f-config-data\") pod \"watcher-db-sync-5zzh7\" (UID: \"16189a3f-0a2b-453c-8bb1-8e8c4d79c13f\") " pod="openstack/watcher-db-sync-5zzh7" Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.264991 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/16189a3f-0a2b-453c-8bb1-8e8c4d79c13f-db-sync-config-data\") pod \"watcher-db-sync-5zzh7\" (UID: \"16189a3f-0a2b-453c-8bb1-8e8c4d79c13f\") " pod="openstack/watcher-db-sync-5zzh7" Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.267181 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4scl\" (UniqueName: \"kubernetes.io/projected/16189a3f-0a2b-453c-8bb1-8e8c4d79c13f-kube-api-access-z4scl\") pod \"watcher-db-sync-5zzh7\" (UID: \"16189a3f-0a2b-453c-8bb1-8e8c4d79c13f\") " pod="openstack/watcher-db-sync-5zzh7" Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.301417 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-582c-account-create-update-jqxth"] Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.302603 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-582c-account-create-update-jqxth" Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.304544 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.319648 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-582c-account-create-update-jqxth"] Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.349688 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bb5hf\" (UniqueName: \"kubernetes.io/projected/a11aafed-1003-4121-9e6e-e5f20c40389f-kube-api-access-bb5hf\") pod \"neutron-db-create-f5xkq\" (UID: \"a11aafed-1003-4121-9e6e-e5f20c40389f\") " pod="openstack/neutron-db-create-f5xkq" Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.349839 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a11aafed-1003-4121-9e6e-e5f20c40389f-operator-scripts\") pod \"neutron-db-create-f5xkq\" (UID: \"a11aafed-1003-4121-9e6e-e5f20c40389f\") " pod="openstack/neutron-db-create-f5xkq" Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.451322 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a11aafed-1003-4121-9e6e-e5f20c40389f-operator-scripts\") pod \"neutron-db-create-f5xkq\" (UID: \"a11aafed-1003-4121-9e6e-e5f20c40389f\") " pod="openstack/neutron-db-create-f5xkq" Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.451414 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ffc04905-1f57-4888-bb88-db5187753b5f-operator-scripts\") pod \"neutron-582c-account-create-update-jqxth\" (UID: \"ffc04905-1f57-4888-bb88-db5187753b5f\") " pod="openstack/neutron-582c-account-create-update-jqxth" Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.451545 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4tlj\" (UniqueName: \"kubernetes.io/projected/ffc04905-1f57-4888-bb88-db5187753b5f-kube-api-access-c4tlj\") pod \"neutron-582c-account-create-update-jqxth\" (UID: \"ffc04905-1f57-4888-bb88-db5187753b5f\") " pod="openstack/neutron-582c-account-create-update-jqxth" Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.451597 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bb5hf\" (UniqueName: \"kubernetes.io/projected/a11aafed-1003-4121-9e6e-e5f20c40389f-kube-api-access-bb5hf\") pod \"neutron-db-create-f5xkq\" (UID: \"a11aafed-1003-4121-9e6e-e5f20c40389f\") " pod="openstack/neutron-db-create-f5xkq" Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.452323 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a11aafed-1003-4121-9e6e-e5f20c40389f-operator-scripts\") pod \"neutron-db-create-f5xkq\" (UID: \"a11aafed-1003-4121-9e6e-e5f20c40389f\") " pod="openstack/neutron-db-create-f5xkq" Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.468942 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bb5hf\" (UniqueName: \"kubernetes.io/projected/a11aafed-1003-4121-9e6e-e5f20c40389f-kube-api-access-bb5hf\") pod \"neutron-db-create-f5xkq\" (UID: \"a11aafed-1003-4121-9e6e-e5f20c40389f\") " pod="openstack/neutron-db-create-f5xkq" Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.546318 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-5zzh7" Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.553222 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ffc04905-1f57-4888-bb88-db5187753b5f-operator-scripts\") pod \"neutron-582c-account-create-update-jqxth\" (UID: \"ffc04905-1f57-4888-bb88-db5187753b5f\") " pod="openstack/neutron-582c-account-create-update-jqxth" Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.553429 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4tlj\" (UniqueName: \"kubernetes.io/projected/ffc04905-1f57-4888-bb88-db5187753b5f-kube-api-access-c4tlj\") pod \"neutron-582c-account-create-update-jqxth\" (UID: \"ffc04905-1f57-4888-bb88-db5187753b5f\") " pod="openstack/neutron-582c-account-create-update-jqxth" Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.554106 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ffc04905-1f57-4888-bb88-db5187753b5f-operator-scripts\") pod \"neutron-582c-account-create-update-jqxth\" (UID: \"ffc04905-1f57-4888-bb88-db5187753b5f\") " pod="openstack/neutron-582c-account-create-update-jqxth" Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.572100 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4tlj\" (UniqueName: \"kubernetes.io/projected/ffc04905-1f57-4888-bb88-db5187753b5f-kube-api-access-c4tlj\") pod \"neutron-582c-account-create-update-jqxth\" (UID: \"ffc04905-1f57-4888-bb88-db5187753b5f\") " pod="openstack/neutron-582c-account-create-update-jqxth" Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.646674 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-f5xkq" Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.654026 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-582c-account-create-update-jqxth" Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.735312 4700 generic.go:334] "Generic (PLEG): container finished" podID="f5d89c26-6133-4987-8c61-cc19c80567e0" containerID="a1755da87ed07a85b745e9409058de03e937db1f3700433133f4bcecfb9f0aa4" exitCode=0 Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.735383 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-7xvrx" event={"ID":"f5d89c26-6133-4987-8c61-cc19c80567e0","Type":"ContainerDied","Data":"a1755da87ed07a85b745e9409058de03e937db1f3700433133f4bcecfb9f0aa4"} Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.741695 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-qm46k" event={"ID":"a238beaf-451f-452e-a30f-09779ae217c1","Type":"ContainerStarted","Data":"bc6714991554d7aa8a5b503a986c34630d41c3edb6040160ee068dc0e8888704"} Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.744743 4700 generic.go:334] "Generic (PLEG): container finished" podID="fe426647-f5e6-46bf-b187-8f8cb62b06e1" containerID="05e0c2020c9571a38a99efb4ec606145ffe49be0926c426091baeabfdfac356b" exitCode=0 Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.744802 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-2c01-account-create-update-crm9x" event={"ID":"fe426647-f5e6-46bf-b187-8f8cb62b06e1","Type":"ContainerDied","Data":"05e0c2020c9571a38a99efb4ec606145ffe49be0926c426091baeabfdfac356b"} Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.744829 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-2c01-account-create-update-crm9x" event={"ID":"fe426647-f5e6-46bf-b187-8f8cb62b06e1","Type":"ContainerStarted","Data":"24aea33cc53b3bf1fdd61c806c1236e2308923fcebbf1a18aa39181cc51e2aa1"} Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.748428 4700 generic.go:334] "Generic (PLEG): container finished" podID="2e67eb82-300d-46bd-b1c6-d2607bf45ccc" containerID="c743a30c2358f798b11a6ad8bb7d06f0869f65ded2d52c2eb9d11e6bbafa0117" exitCode=0 Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.748490 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-rc9w6" event={"ID":"2e67eb82-300d-46bd-b1c6-d2607bf45ccc","Type":"ContainerDied","Data":"c743a30c2358f798b11a6ad8bb7d06f0869f65ded2d52c2eb9d11e6bbafa0117"} Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.763082 4700 generic.go:334] "Generic (PLEG): container finished" podID="2062608c-6759-45bd-9879-2a96af693b31" containerID="4b7f93b578e644781c33d2b3aa82faa1979c30b013b09b4efe7e74cb9564891a" exitCode=0 Feb 27 17:21:09 crc kubenswrapper[4700]: I0227 17:21:09.763126 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-1943-account-create-update-h8f9l" event={"ID":"2062608c-6759-45bd-9879-2a96af693b31","Type":"ContainerDied","Data":"4b7f93b578e644781c33d2b3aa82faa1979c30b013b09b4efe7e74cb9564891a"} Feb 27 17:21:10 crc kubenswrapper[4700]: I0227 17:21:10.048555 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-sync-5zzh7"] Feb 27 17:21:10 crc kubenswrapper[4700]: I0227 17:21:10.236601 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-f5xkq"] Feb 27 17:21:10 crc kubenswrapper[4700]: W0227 17:21:10.237005 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda11aafed_1003_4121_9e6e_e5f20c40389f.slice/crio-fa30ab39258bf4325918b3e826c365d8b0a95f82cc29387537ff88a8b75cd7f9 WatchSource:0}: Error finding container fa30ab39258bf4325918b3e826c365d8b0a95f82cc29387537ff88a8b75cd7f9: Status 404 returned error can't find the container with id fa30ab39258bf4325918b3e826c365d8b0a95f82cc29387537ff88a8b75cd7f9 Feb 27 17:21:10 crc kubenswrapper[4700]: I0227 17:21:10.344921 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-582c-account-create-update-jqxth"] Feb 27 17:21:10 crc kubenswrapper[4700]: W0227 17:21:10.353703 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podffc04905_1f57_4888_bb88_db5187753b5f.slice/crio-9ddceb4171867821155b45a7173118f22411e156fbd5fa4ee41f2d6c800d0499 WatchSource:0}: Error finding container 9ddceb4171867821155b45a7173118f22411e156fbd5fa4ee41f2d6c800d0499: Status 404 returned error can't find the container with id 9ddceb4171867821155b45a7173118f22411e156fbd5fa4ee41f2d6c800d0499 Feb 27 17:21:10 crc kubenswrapper[4700]: I0227 17:21:10.357193 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Feb 27 17:21:10 crc kubenswrapper[4700]: I0227 17:21:10.366618 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Feb 27 17:21:10 crc kubenswrapper[4700]: I0227 17:21:10.775672 4700 generic.go:334] "Generic (PLEG): container finished" podID="a11aafed-1003-4121-9e6e-e5f20c40389f" containerID="89b673903cbfb2a6d5ce565347af43128ab227341122662ec2b9e57c4cb16400" exitCode=0 Feb 27 17:21:10 crc kubenswrapper[4700]: I0227 17:21:10.775860 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-f5xkq" event={"ID":"a11aafed-1003-4121-9e6e-e5f20c40389f","Type":"ContainerDied","Data":"89b673903cbfb2a6d5ce565347af43128ab227341122662ec2b9e57c4cb16400"} Feb 27 17:21:10 crc kubenswrapper[4700]: I0227 17:21:10.775960 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-f5xkq" event={"ID":"a11aafed-1003-4121-9e6e-e5f20c40389f","Type":"ContainerStarted","Data":"fa30ab39258bf4325918b3e826c365d8b0a95f82cc29387537ff88a8b75cd7f9"} Feb 27 17:21:10 crc kubenswrapper[4700]: I0227 17:21:10.777332 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-5zzh7" event={"ID":"16189a3f-0a2b-453c-8bb1-8e8c4d79c13f","Type":"ContainerStarted","Data":"4945f0efa13d6bc2bee73b8f6c511e29beacbaa6b7ea6d50349b2364bb4a6371"} Feb 27 17:21:10 crc kubenswrapper[4700]: I0227 17:21:10.781672 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-582c-account-create-update-jqxth" event={"ID":"ffc04905-1f57-4888-bb88-db5187753b5f","Type":"ContainerStarted","Data":"66ff4804ed8e0b99ba4fee444456685d8955deb6348ffb8f52c49d90997aa8b6"} Feb 27 17:21:10 crc kubenswrapper[4700]: I0227 17:21:10.781701 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-582c-account-create-update-jqxth" event={"ID":"ffc04905-1f57-4888-bb88-db5187753b5f","Type":"ContainerStarted","Data":"9ddceb4171867821155b45a7173118f22411e156fbd5fa4ee41f2d6c800d0499"} Feb 27 17:21:10 crc kubenswrapper[4700]: I0227 17:21:10.791086 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Feb 27 17:21:10 crc kubenswrapper[4700]: I0227 17:21:10.810332 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-582c-account-create-update-jqxth" podStartSLOduration=1.810311148 podStartE2EDuration="1.810311148s" podCreationTimestamp="2026-02-27 17:21:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:21:10.807025411 +0000 UTC m=+1230.792338158" watchObservedRunningTime="2026-02-27 17:21:10.810311148 +0000 UTC m=+1230.795623895" Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.280067 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-1943-account-create-update-h8f9l" Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.393341 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2062608c-6759-45bd-9879-2a96af693b31-operator-scripts\") pod \"2062608c-6759-45bd-9879-2a96af693b31\" (UID: \"2062608c-6759-45bd-9879-2a96af693b31\") " Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.394494 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2062608c-6759-45bd-9879-2a96af693b31-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2062608c-6759-45bd-9879-2a96af693b31" (UID: "2062608c-6759-45bd-9879-2a96af693b31"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.394593 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cnrxk\" (UniqueName: \"kubernetes.io/projected/2062608c-6759-45bd-9879-2a96af693b31-kube-api-access-cnrxk\") pod \"2062608c-6759-45bd-9879-2a96af693b31\" (UID: \"2062608c-6759-45bd-9879-2a96af693b31\") " Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.395639 4700 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2062608c-6759-45bd-9879-2a96af693b31-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.400143 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2062608c-6759-45bd-9879-2a96af693b31-kube-api-access-cnrxk" (OuterVolumeSpecName: "kube-api-access-cnrxk") pod "2062608c-6759-45bd-9879-2a96af693b31" (UID: "2062608c-6759-45bd-9879-2a96af693b31"). InnerVolumeSpecName "kube-api-access-cnrxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.425388 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-rc9w6" Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.431808 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-2c01-account-create-update-crm9x" Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.444195 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-7xvrx" Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.497268 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65cmp\" (UniqueName: \"kubernetes.io/projected/fe426647-f5e6-46bf-b187-8f8cb62b06e1-kube-api-access-65cmp\") pod \"fe426647-f5e6-46bf-b187-8f8cb62b06e1\" (UID: \"fe426647-f5e6-46bf-b187-8f8cb62b06e1\") " Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.497377 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2mwk\" (UniqueName: \"kubernetes.io/projected/f5d89c26-6133-4987-8c61-cc19c80567e0-kube-api-access-n2mwk\") pod \"f5d89c26-6133-4987-8c61-cc19c80567e0\" (UID: \"f5d89c26-6133-4987-8c61-cc19c80567e0\") " Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.497425 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5d89c26-6133-4987-8c61-cc19c80567e0-operator-scripts\") pod \"f5d89c26-6133-4987-8c61-cc19c80567e0\" (UID: \"f5d89c26-6133-4987-8c61-cc19c80567e0\") " Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.497616 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtjdq\" (UniqueName: \"kubernetes.io/projected/2e67eb82-300d-46bd-b1c6-d2607bf45ccc-kube-api-access-rtjdq\") pod \"2e67eb82-300d-46bd-b1c6-d2607bf45ccc\" (UID: \"2e67eb82-300d-46bd-b1c6-d2607bf45ccc\") " Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.497703 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e67eb82-300d-46bd-b1c6-d2607bf45ccc-operator-scripts\") pod \"2e67eb82-300d-46bd-b1c6-d2607bf45ccc\" (UID: \"2e67eb82-300d-46bd-b1c6-d2607bf45ccc\") " Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.497784 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe426647-f5e6-46bf-b187-8f8cb62b06e1-operator-scripts\") pod \"fe426647-f5e6-46bf-b187-8f8cb62b06e1\" (UID: \"fe426647-f5e6-46bf-b187-8f8cb62b06e1\") " Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.498253 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe426647-f5e6-46bf-b187-8f8cb62b06e1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fe426647-f5e6-46bf-b187-8f8cb62b06e1" (UID: "fe426647-f5e6-46bf-b187-8f8cb62b06e1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.498338 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cnrxk\" (UniqueName: \"kubernetes.io/projected/2062608c-6759-45bd-9879-2a96af693b31-kube-api-access-cnrxk\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.498344 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e67eb82-300d-46bd-b1c6-d2607bf45ccc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2e67eb82-300d-46bd-b1c6-d2607bf45ccc" (UID: "2e67eb82-300d-46bd-b1c6-d2607bf45ccc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.498935 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5d89c26-6133-4987-8c61-cc19c80567e0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f5d89c26-6133-4987-8c61-cc19c80567e0" (UID: "f5d89c26-6133-4987-8c61-cc19c80567e0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.501343 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5d89c26-6133-4987-8c61-cc19c80567e0-kube-api-access-n2mwk" (OuterVolumeSpecName: "kube-api-access-n2mwk") pod "f5d89c26-6133-4987-8c61-cc19c80567e0" (UID: "f5d89c26-6133-4987-8c61-cc19c80567e0"). InnerVolumeSpecName "kube-api-access-n2mwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.501598 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe426647-f5e6-46bf-b187-8f8cb62b06e1-kube-api-access-65cmp" (OuterVolumeSpecName: "kube-api-access-65cmp") pod "fe426647-f5e6-46bf-b187-8f8cb62b06e1" (UID: "fe426647-f5e6-46bf-b187-8f8cb62b06e1"). InnerVolumeSpecName "kube-api-access-65cmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.503893 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e67eb82-300d-46bd-b1c6-d2607bf45ccc-kube-api-access-rtjdq" (OuterVolumeSpecName: "kube-api-access-rtjdq") pod "2e67eb82-300d-46bd-b1c6-d2607bf45ccc" (UID: "2e67eb82-300d-46bd-b1c6-d2607bf45ccc"). InnerVolumeSpecName "kube-api-access-rtjdq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.600171 4700 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e67eb82-300d-46bd-b1c6-d2607bf45ccc-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.600199 4700 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe426647-f5e6-46bf-b187-8f8cb62b06e1-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.600209 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65cmp\" (UniqueName: \"kubernetes.io/projected/fe426647-f5e6-46bf-b187-8f8cb62b06e1-kube-api-access-65cmp\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.600219 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2mwk\" (UniqueName: \"kubernetes.io/projected/f5d89c26-6133-4987-8c61-cc19c80567e0-kube-api-access-n2mwk\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.600228 4700 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5d89c26-6133-4987-8c61-cc19c80567e0-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.600236 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtjdq\" (UniqueName: \"kubernetes.io/projected/2e67eb82-300d-46bd-b1c6-d2607bf45ccc-kube-api-access-rtjdq\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.792534 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-2c01-account-create-update-crm9x" Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.793239 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-2c01-account-create-update-crm9x" event={"ID":"fe426647-f5e6-46bf-b187-8f8cb62b06e1","Type":"ContainerDied","Data":"24aea33cc53b3bf1fdd61c806c1236e2308923fcebbf1a18aa39181cc51e2aa1"} Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.793270 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24aea33cc53b3bf1fdd61c806c1236e2308923fcebbf1a18aa39181cc51e2aa1" Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.801920 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-rc9w6" event={"ID":"2e67eb82-300d-46bd-b1c6-d2607bf45ccc","Type":"ContainerDied","Data":"d9f96c9b6146e2f3f597814ee423a55d04f4cea8c25e6543d63295aae0bfc9a3"} Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.801958 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9f96c9b6146e2f3f597814ee423a55d04f4cea8c25e6543d63295aae0bfc9a3" Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.801939 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-rc9w6" Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.810167 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-1943-account-create-update-h8f9l" Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.811301 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-1943-account-create-update-h8f9l" event={"ID":"2062608c-6759-45bd-9879-2a96af693b31","Type":"ContainerDied","Data":"b1a42e648c08ada3917bf8cba48dd60e776820d7dc1109498df2addc1840ad15"} Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.811349 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1a42e648c08ada3917bf8cba48dd60e776820d7dc1109498df2addc1840ad15" Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.812816 4700 generic.go:334] "Generic (PLEG): container finished" podID="ffc04905-1f57-4888-bb88-db5187753b5f" containerID="66ff4804ed8e0b99ba4fee444456685d8955deb6348ffb8f52c49d90997aa8b6" exitCode=0 Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.812879 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-582c-account-create-update-jqxth" event={"ID":"ffc04905-1f57-4888-bb88-db5187753b5f","Type":"ContainerDied","Data":"66ff4804ed8e0b99ba4fee444456685d8955deb6348ffb8f52c49d90997aa8b6"} Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.814720 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-7xvrx" event={"ID":"f5d89c26-6133-4987-8c61-cc19c80567e0","Type":"ContainerDied","Data":"a62625ee5f2cd1e901acd226e6f8244bf559509a78c0aa4edd3943fe7be19d75"} Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.814749 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a62625ee5f2cd1e901acd226e6f8244bf559509a78c0aa4edd3943fe7be19d75" Feb 27 17:21:11 crc kubenswrapper[4700]: I0227 17:21:11.814860 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-7xvrx" Feb 27 17:21:15 crc kubenswrapper[4700]: I0227 17:21:15.126208 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-f5xkq" Feb 27 17:21:15 crc kubenswrapper[4700]: I0227 17:21:15.174163 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a11aafed-1003-4121-9e6e-e5f20c40389f-operator-scripts\") pod \"a11aafed-1003-4121-9e6e-e5f20c40389f\" (UID: \"a11aafed-1003-4121-9e6e-e5f20c40389f\") " Feb 27 17:21:15 crc kubenswrapper[4700]: I0227 17:21:15.174402 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bb5hf\" (UniqueName: \"kubernetes.io/projected/a11aafed-1003-4121-9e6e-e5f20c40389f-kube-api-access-bb5hf\") pod \"a11aafed-1003-4121-9e6e-e5f20c40389f\" (UID: \"a11aafed-1003-4121-9e6e-e5f20c40389f\") " Feb 27 17:21:15 crc kubenswrapper[4700]: I0227 17:21:15.176765 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a11aafed-1003-4121-9e6e-e5f20c40389f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a11aafed-1003-4121-9e6e-e5f20c40389f" (UID: "a11aafed-1003-4121-9e6e-e5f20c40389f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:21:15 crc kubenswrapper[4700]: I0227 17:21:15.184218 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a11aafed-1003-4121-9e6e-e5f20c40389f-kube-api-access-bb5hf" (OuterVolumeSpecName: "kube-api-access-bb5hf") pod "a11aafed-1003-4121-9e6e-e5f20c40389f" (UID: "a11aafed-1003-4121-9e6e-e5f20c40389f"). InnerVolumeSpecName "kube-api-access-bb5hf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:21:15 crc kubenswrapper[4700]: I0227 17:21:15.276150 4700 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a11aafed-1003-4121-9e6e-e5f20c40389f-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:15 crc kubenswrapper[4700]: I0227 17:21:15.276186 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bb5hf\" (UniqueName: \"kubernetes.io/projected/a11aafed-1003-4121-9e6e-e5f20c40389f-kube-api-access-bb5hf\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:15 crc kubenswrapper[4700]: I0227 17:21:15.494550 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7cb78597c7-wtsvw" Feb 27 17:21:15 crc kubenswrapper[4700]: I0227 17:21:15.564087 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6dc5d795ff-nbxlk"] Feb 27 17:21:15 crc kubenswrapper[4700]: I0227 17:21:15.564303 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6dc5d795ff-nbxlk" podUID="da70ae3e-1bde-47f1-ab27-70ac8726f421" containerName="dnsmasq-dns" containerID="cri-o://6fb9866278becdccf3970d53fb04f5f1456c5ecd00d4a819521b7670a61aecd3" gracePeriod=10 Feb 27 17:21:15 crc kubenswrapper[4700]: I0227 17:21:15.852669 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-f5xkq" event={"ID":"a11aafed-1003-4121-9e6e-e5f20c40389f","Type":"ContainerDied","Data":"fa30ab39258bf4325918b3e826c365d8b0a95f82cc29387537ff88a8b75cd7f9"} Feb 27 17:21:15 crc kubenswrapper[4700]: I0227 17:21:15.852720 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa30ab39258bf4325918b3e826c365d8b0a95f82cc29387537ff88a8b75cd7f9" Feb 27 17:21:15 crc kubenswrapper[4700]: I0227 17:21:15.852728 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-f5xkq" Feb 27 17:21:15 crc kubenswrapper[4700]: I0227 17:21:15.857823 4700 generic.go:334] "Generic (PLEG): container finished" podID="da70ae3e-1bde-47f1-ab27-70ac8726f421" containerID="6fb9866278becdccf3970d53fb04f5f1456c5ecd00d4a819521b7670a61aecd3" exitCode=0 Feb 27 17:21:15 crc kubenswrapper[4700]: I0227 17:21:15.857865 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6dc5d795ff-nbxlk" event={"ID":"da70ae3e-1bde-47f1-ab27-70ac8726f421","Type":"ContainerDied","Data":"6fb9866278becdccf3970d53fb04f5f1456c5ecd00d4a819521b7670a61aecd3"} Feb 27 17:21:18 crc kubenswrapper[4700]: I0227 17:21:18.865599 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6dc5d795ff-nbxlk" podUID="da70ae3e-1bde-47f1-ab27-70ac8726f421" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.125:5353: connect: connection refused" Feb 27 17:21:20 crc kubenswrapper[4700]: I0227 17:21:20.958031 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-582c-account-create-update-jqxth" event={"ID":"ffc04905-1f57-4888-bb88-db5187753b5f","Type":"ContainerDied","Data":"9ddceb4171867821155b45a7173118f22411e156fbd5fa4ee41f2d6c800d0499"} Feb 27 17:21:20 crc kubenswrapper[4700]: I0227 17:21:20.958315 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ddceb4171867821155b45a7173118f22411e156fbd5fa4ee41f2d6c800d0499" Feb 27 17:21:21 crc kubenswrapper[4700]: I0227 17:21:21.190515 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-582c-account-create-update-jqxth" Feb 27 17:21:21 crc kubenswrapper[4700]: I0227 17:21:21.227255 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6dc5d795ff-nbxlk" Feb 27 17:21:21 crc kubenswrapper[4700]: I0227 17:21:21.304865 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ffc04905-1f57-4888-bb88-db5187753b5f-operator-scripts\") pod \"ffc04905-1f57-4888-bb88-db5187753b5f\" (UID: \"ffc04905-1f57-4888-bb88-db5187753b5f\") " Feb 27 17:21:21 crc kubenswrapper[4700]: I0227 17:21:21.304942 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjq6m\" (UniqueName: \"kubernetes.io/projected/da70ae3e-1bde-47f1-ab27-70ac8726f421-kube-api-access-fjq6m\") pod \"da70ae3e-1bde-47f1-ab27-70ac8726f421\" (UID: \"da70ae3e-1bde-47f1-ab27-70ac8726f421\") " Feb 27 17:21:21 crc kubenswrapper[4700]: I0227 17:21:21.307304 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ffc04905-1f57-4888-bb88-db5187753b5f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ffc04905-1f57-4888-bb88-db5187753b5f" (UID: "ffc04905-1f57-4888-bb88-db5187753b5f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:21:21 crc kubenswrapper[4700]: I0227 17:21:21.307431 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/da70ae3e-1bde-47f1-ab27-70ac8726f421-ovsdbserver-sb\") pod \"da70ae3e-1bde-47f1-ab27-70ac8726f421\" (UID: \"da70ae3e-1bde-47f1-ab27-70ac8726f421\") " Feb 27 17:21:21 crc kubenswrapper[4700]: I0227 17:21:21.307620 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da70ae3e-1bde-47f1-ab27-70ac8726f421-config\") pod \"da70ae3e-1bde-47f1-ab27-70ac8726f421\" (UID: \"da70ae3e-1bde-47f1-ab27-70ac8726f421\") " Feb 27 17:21:21 crc kubenswrapper[4700]: I0227 17:21:21.307693 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4tlj\" (UniqueName: \"kubernetes.io/projected/ffc04905-1f57-4888-bb88-db5187753b5f-kube-api-access-c4tlj\") pod \"ffc04905-1f57-4888-bb88-db5187753b5f\" (UID: \"ffc04905-1f57-4888-bb88-db5187753b5f\") " Feb 27 17:21:21 crc kubenswrapper[4700]: I0227 17:21:21.307751 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da70ae3e-1bde-47f1-ab27-70ac8726f421-dns-svc\") pod \"da70ae3e-1bde-47f1-ab27-70ac8726f421\" (UID: \"da70ae3e-1bde-47f1-ab27-70ac8726f421\") " Feb 27 17:21:21 crc kubenswrapper[4700]: I0227 17:21:21.307798 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/da70ae3e-1bde-47f1-ab27-70ac8726f421-ovsdbserver-nb\") pod \"da70ae3e-1bde-47f1-ab27-70ac8726f421\" (UID: \"da70ae3e-1bde-47f1-ab27-70ac8726f421\") " Feb 27 17:21:21 crc kubenswrapper[4700]: I0227 17:21:21.309570 4700 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ffc04905-1f57-4888-bb88-db5187753b5f-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:21 crc kubenswrapper[4700]: I0227 17:21:21.311959 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da70ae3e-1bde-47f1-ab27-70ac8726f421-kube-api-access-fjq6m" (OuterVolumeSpecName: "kube-api-access-fjq6m") pod "da70ae3e-1bde-47f1-ab27-70ac8726f421" (UID: "da70ae3e-1bde-47f1-ab27-70ac8726f421"). InnerVolumeSpecName "kube-api-access-fjq6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:21:21 crc kubenswrapper[4700]: I0227 17:21:21.316639 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffc04905-1f57-4888-bb88-db5187753b5f-kube-api-access-c4tlj" (OuterVolumeSpecName: "kube-api-access-c4tlj") pod "ffc04905-1f57-4888-bb88-db5187753b5f" (UID: "ffc04905-1f57-4888-bb88-db5187753b5f"). InnerVolumeSpecName "kube-api-access-c4tlj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:21:21 crc kubenswrapper[4700]: I0227 17:21:21.364645 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da70ae3e-1bde-47f1-ab27-70ac8726f421-config" (OuterVolumeSpecName: "config") pod "da70ae3e-1bde-47f1-ab27-70ac8726f421" (UID: "da70ae3e-1bde-47f1-ab27-70ac8726f421"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:21:21 crc kubenswrapper[4700]: I0227 17:21:21.367936 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da70ae3e-1bde-47f1-ab27-70ac8726f421-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "da70ae3e-1bde-47f1-ab27-70ac8726f421" (UID: "da70ae3e-1bde-47f1-ab27-70ac8726f421"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:21:21 crc kubenswrapper[4700]: I0227 17:21:21.371053 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da70ae3e-1bde-47f1-ab27-70ac8726f421-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "da70ae3e-1bde-47f1-ab27-70ac8726f421" (UID: "da70ae3e-1bde-47f1-ab27-70ac8726f421"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:21:21 crc kubenswrapper[4700]: I0227 17:21:21.372380 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da70ae3e-1bde-47f1-ab27-70ac8726f421-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "da70ae3e-1bde-47f1-ab27-70ac8726f421" (UID: "da70ae3e-1bde-47f1-ab27-70ac8726f421"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:21:21 crc kubenswrapper[4700]: I0227 17:21:21.411499 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4tlj\" (UniqueName: \"kubernetes.io/projected/ffc04905-1f57-4888-bb88-db5187753b5f-kube-api-access-c4tlj\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:21 crc kubenswrapper[4700]: I0227 17:21:21.411525 4700 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da70ae3e-1bde-47f1-ab27-70ac8726f421-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:21 crc kubenswrapper[4700]: I0227 17:21:21.411536 4700 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/da70ae3e-1bde-47f1-ab27-70ac8726f421-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:21 crc kubenswrapper[4700]: I0227 17:21:21.411545 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjq6m\" (UniqueName: \"kubernetes.io/projected/da70ae3e-1bde-47f1-ab27-70ac8726f421-kube-api-access-fjq6m\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:21 crc kubenswrapper[4700]: I0227 17:21:21.411552 4700 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/da70ae3e-1bde-47f1-ab27-70ac8726f421-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:21 crc kubenswrapper[4700]: I0227 17:21:21.411587 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da70ae3e-1bde-47f1-ab27-70ac8726f421-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:21 crc kubenswrapper[4700]: I0227 17:21:21.975692 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-qm46k" event={"ID":"a238beaf-451f-452e-a30f-09779ae217c1","Type":"ContainerStarted","Data":"008e6edb1f631b4dc678019ab290f43fec55bb8f1539b3845f7fd56bbede54c2"} Feb 27 17:21:21 crc kubenswrapper[4700]: I0227 17:21:21.986988 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6dc5d795ff-nbxlk" Feb 27 17:21:21 crc kubenswrapper[4700]: I0227 17:21:21.987207 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6dc5d795ff-nbxlk" event={"ID":"da70ae3e-1bde-47f1-ab27-70ac8726f421","Type":"ContainerDied","Data":"503169e0dc123056d4a4f574b64a53114435a1eb573a85590a860a288b099253"} Feb 27 17:21:21 crc kubenswrapper[4700]: I0227 17:21:21.987278 4700 scope.go:117] "RemoveContainer" containerID="6fb9866278becdccf3970d53fb04f5f1456c5ecd00d4a819521b7670a61aecd3" Feb 27 17:21:22 crc kubenswrapper[4700]: I0227 17:21:22.009255 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-582c-account-create-update-jqxth" Feb 27 17:21:22 crc kubenswrapper[4700]: I0227 17:21:22.009310 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-5zzh7" event={"ID":"16189a3f-0a2b-453c-8bb1-8e8c4d79c13f","Type":"ContainerStarted","Data":"715c57270ee2cdfd98da501c87de458e60122b453b9ff5f05607477c5dfcdc44"} Feb 27 17:21:22 crc kubenswrapper[4700]: I0227 17:21:22.013633 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-qm46k" podStartSLOduration=2.978476103 podStartE2EDuration="15.013607468s" podCreationTimestamp="2026-02-27 17:21:07 +0000 UTC" firstStartedPulling="2026-02-27 17:21:08.809222434 +0000 UTC m=+1228.794535181" lastFinishedPulling="2026-02-27 17:21:20.844353789 +0000 UTC m=+1240.829666546" observedRunningTime="2026-02-27 17:21:22.009158341 +0000 UTC m=+1241.994471138" watchObservedRunningTime="2026-02-27 17:21:22.013607468 +0000 UTC m=+1241.998920255" Feb 27 17:21:22 crc kubenswrapper[4700]: I0227 17:21:22.040546 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-db-sync-5zzh7" podStartSLOduration=2.140396859 podStartE2EDuration="13.04052253s" podCreationTimestamp="2026-02-27 17:21:09 +0000 UTC" firstStartedPulling="2026-02-27 17:21:10.057195456 +0000 UTC m=+1230.042508203" lastFinishedPulling="2026-02-27 17:21:20.957321127 +0000 UTC m=+1240.942633874" observedRunningTime="2026-02-27 17:21:22.0295351 +0000 UTC m=+1242.014847887" watchObservedRunningTime="2026-02-27 17:21:22.04052253 +0000 UTC m=+1242.025835317" Feb 27 17:21:22 crc kubenswrapper[4700]: I0227 17:21:22.053327 4700 scope.go:117] "RemoveContainer" containerID="e638eee0c53829a6e26465fe8359d3f713d701fa523d9c136f61d6c0fd0d2444" Feb 27 17:21:22 crc kubenswrapper[4700]: I0227 17:21:22.070691 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6dc5d795ff-nbxlk"] Feb 27 17:21:22 crc kubenswrapper[4700]: I0227 17:21:22.090778 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6dc5d795ff-nbxlk"] Feb 27 17:21:22 crc kubenswrapper[4700]: I0227 17:21:22.996787 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da70ae3e-1bde-47f1-ab27-70ac8726f421" path="/var/lib/kubelet/pods/da70ae3e-1bde-47f1-ab27-70ac8726f421/volumes" Feb 27 17:21:25 crc kubenswrapper[4700]: I0227 17:21:25.048951 4700 generic.go:334] "Generic (PLEG): container finished" podID="16189a3f-0a2b-453c-8bb1-8e8c4d79c13f" containerID="715c57270ee2cdfd98da501c87de458e60122b453b9ff5f05607477c5dfcdc44" exitCode=0 Feb 27 17:21:25 crc kubenswrapper[4700]: I0227 17:21:25.049023 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-5zzh7" event={"ID":"16189a3f-0a2b-453c-8bb1-8e8c4d79c13f","Type":"ContainerDied","Data":"715c57270ee2cdfd98da501c87de458e60122b453b9ff5f05607477c5dfcdc44"} Feb 27 17:21:26 crc kubenswrapper[4700]: I0227 17:21:26.062070 4700 generic.go:334] "Generic (PLEG): container finished" podID="a238beaf-451f-452e-a30f-09779ae217c1" containerID="008e6edb1f631b4dc678019ab290f43fec55bb8f1539b3845f7fd56bbede54c2" exitCode=0 Feb 27 17:21:26 crc kubenswrapper[4700]: I0227 17:21:26.062164 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-qm46k" event={"ID":"a238beaf-451f-452e-a30f-09779ae217c1","Type":"ContainerDied","Data":"008e6edb1f631b4dc678019ab290f43fec55bb8f1539b3845f7fd56bbede54c2"} Feb 27 17:21:26 crc kubenswrapper[4700]: I0227 17:21:26.478210 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-5zzh7" Feb 27 17:21:26 crc kubenswrapper[4700]: I0227 17:21:26.645641 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16189a3f-0a2b-453c-8bb1-8e8c4d79c13f-config-data\") pod \"16189a3f-0a2b-453c-8bb1-8e8c4d79c13f\" (UID: \"16189a3f-0a2b-453c-8bb1-8e8c4d79c13f\") " Feb 27 17:21:26 crc kubenswrapper[4700]: I0227 17:21:26.645893 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4scl\" (UniqueName: \"kubernetes.io/projected/16189a3f-0a2b-453c-8bb1-8e8c4d79c13f-kube-api-access-z4scl\") pod \"16189a3f-0a2b-453c-8bb1-8e8c4d79c13f\" (UID: \"16189a3f-0a2b-453c-8bb1-8e8c4d79c13f\") " Feb 27 17:21:26 crc kubenswrapper[4700]: I0227 17:21:26.645951 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/16189a3f-0a2b-453c-8bb1-8e8c4d79c13f-db-sync-config-data\") pod \"16189a3f-0a2b-453c-8bb1-8e8c4d79c13f\" (UID: \"16189a3f-0a2b-453c-8bb1-8e8c4d79c13f\") " Feb 27 17:21:26 crc kubenswrapper[4700]: I0227 17:21:26.646083 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16189a3f-0a2b-453c-8bb1-8e8c4d79c13f-combined-ca-bundle\") pod \"16189a3f-0a2b-453c-8bb1-8e8c4d79c13f\" (UID: \"16189a3f-0a2b-453c-8bb1-8e8c4d79c13f\") " Feb 27 17:21:26 crc kubenswrapper[4700]: I0227 17:21:26.654102 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16189a3f-0a2b-453c-8bb1-8e8c4d79c13f-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "16189a3f-0a2b-453c-8bb1-8e8c4d79c13f" (UID: "16189a3f-0a2b-453c-8bb1-8e8c4d79c13f"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:21:26 crc kubenswrapper[4700]: I0227 17:21:26.665075 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16189a3f-0a2b-453c-8bb1-8e8c4d79c13f-kube-api-access-z4scl" (OuterVolumeSpecName: "kube-api-access-z4scl") pod "16189a3f-0a2b-453c-8bb1-8e8c4d79c13f" (UID: "16189a3f-0a2b-453c-8bb1-8e8c4d79c13f"). InnerVolumeSpecName "kube-api-access-z4scl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:21:26 crc kubenswrapper[4700]: I0227 17:21:26.690949 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16189a3f-0a2b-453c-8bb1-8e8c4d79c13f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "16189a3f-0a2b-453c-8bb1-8e8c4d79c13f" (UID: "16189a3f-0a2b-453c-8bb1-8e8c4d79c13f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:21:26 crc kubenswrapper[4700]: I0227 17:21:26.729027 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16189a3f-0a2b-453c-8bb1-8e8c4d79c13f-config-data" (OuterVolumeSpecName: "config-data") pod "16189a3f-0a2b-453c-8bb1-8e8c4d79c13f" (UID: "16189a3f-0a2b-453c-8bb1-8e8c4d79c13f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:21:26 crc kubenswrapper[4700]: I0227 17:21:26.749153 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16189a3f-0a2b-453c-8bb1-8e8c4d79c13f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:26 crc kubenswrapper[4700]: I0227 17:21:26.749206 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16189a3f-0a2b-453c-8bb1-8e8c4d79c13f-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:26 crc kubenswrapper[4700]: I0227 17:21:26.749227 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4scl\" (UniqueName: \"kubernetes.io/projected/16189a3f-0a2b-453c-8bb1-8e8c4d79c13f-kube-api-access-z4scl\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:26 crc kubenswrapper[4700]: I0227 17:21:26.749250 4700 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/16189a3f-0a2b-453c-8bb1-8e8c4d79c13f-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:27 crc kubenswrapper[4700]: I0227 17:21:27.077127 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-5zzh7" Feb 27 17:21:27 crc kubenswrapper[4700]: I0227 17:21:27.077303 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-5zzh7" event={"ID":"16189a3f-0a2b-453c-8bb1-8e8c4d79c13f","Type":"ContainerDied","Data":"4945f0efa13d6bc2bee73b8f6c511e29beacbaa6b7ea6d50349b2364bb4a6371"} Feb 27 17:21:27 crc kubenswrapper[4700]: I0227 17:21:27.077358 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4945f0efa13d6bc2bee73b8f6c511e29beacbaa6b7ea6d50349b2364bb4a6371" Feb 27 17:21:27 crc kubenswrapper[4700]: I0227 17:21:27.513785 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-qm46k" Feb 27 17:21:27 crc kubenswrapper[4700]: I0227 17:21:27.665846 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a238beaf-451f-452e-a30f-09779ae217c1-config-data\") pod \"a238beaf-451f-452e-a30f-09779ae217c1\" (UID: \"a238beaf-451f-452e-a30f-09779ae217c1\") " Feb 27 17:21:27 crc kubenswrapper[4700]: I0227 17:21:27.666161 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a238beaf-451f-452e-a30f-09779ae217c1-combined-ca-bundle\") pod \"a238beaf-451f-452e-a30f-09779ae217c1\" (UID: \"a238beaf-451f-452e-a30f-09779ae217c1\") " Feb 27 17:21:27 crc kubenswrapper[4700]: I0227 17:21:27.666288 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g48ql\" (UniqueName: \"kubernetes.io/projected/a238beaf-451f-452e-a30f-09779ae217c1-kube-api-access-g48ql\") pod \"a238beaf-451f-452e-a30f-09779ae217c1\" (UID: \"a238beaf-451f-452e-a30f-09779ae217c1\") " Feb 27 17:21:27 crc kubenswrapper[4700]: I0227 17:21:27.674632 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a238beaf-451f-452e-a30f-09779ae217c1-kube-api-access-g48ql" (OuterVolumeSpecName: "kube-api-access-g48ql") pod "a238beaf-451f-452e-a30f-09779ae217c1" (UID: "a238beaf-451f-452e-a30f-09779ae217c1"). InnerVolumeSpecName "kube-api-access-g48ql". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:21:27 crc kubenswrapper[4700]: I0227 17:21:27.690694 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a238beaf-451f-452e-a30f-09779ae217c1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a238beaf-451f-452e-a30f-09779ae217c1" (UID: "a238beaf-451f-452e-a30f-09779ae217c1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:21:27 crc kubenswrapper[4700]: I0227 17:21:27.715496 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a238beaf-451f-452e-a30f-09779ae217c1-config-data" (OuterVolumeSpecName: "config-data") pod "a238beaf-451f-452e-a30f-09779ae217c1" (UID: "a238beaf-451f-452e-a30f-09779ae217c1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:21:27 crc kubenswrapper[4700]: I0227 17:21:27.771837 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a238beaf-451f-452e-a30f-09779ae217c1-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:27 crc kubenswrapper[4700]: I0227 17:21:27.771884 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a238beaf-451f-452e-a30f-09779ae217c1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:27 crc kubenswrapper[4700]: I0227 17:21:27.771906 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g48ql\" (UniqueName: \"kubernetes.io/projected/a238beaf-451f-452e-a30f-09779ae217c1-kube-api-access-g48ql\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.087601 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-qm46k" event={"ID":"a238beaf-451f-452e-a30f-09779ae217c1","Type":"ContainerDied","Data":"bc6714991554d7aa8a5b503a986c34630d41c3edb6040160ee068dc0e8888704"} Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.088444 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc6714991554d7aa8a5b503a986c34630d41c3edb6040160ee068dc0e8888704" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.087682 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-qm46k" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.368535 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-9hq27"] Feb 27 17:21:28 crc kubenswrapper[4700]: E0227 17:21:28.368988 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffc04905-1f57-4888-bb88-db5187753b5f" containerName="mariadb-account-create-update" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.369008 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffc04905-1f57-4888-bb88-db5187753b5f" containerName="mariadb-account-create-update" Feb 27 17:21:28 crc kubenswrapper[4700]: E0227 17:21:28.369028 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2062608c-6759-45bd-9879-2a96af693b31" containerName="mariadb-account-create-update" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.369038 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="2062608c-6759-45bd-9879-2a96af693b31" containerName="mariadb-account-create-update" Feb 27 17:21:28 crc kubenswrapper[4700]: E0227 17:21:28.369053 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe426647-f5e6-46bf-b187-8f8cb62b06e1" containerName="mariadb-account-create-update" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.369063 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe426647-f5e6-46bf-b187-8f8cb62b06e1" containerName="mariadb-account-create-update" Feb 27 17:21:28 crc kubenswrapper[4700]: E0227 17:21:28.369082 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a11aafed-1003-4121-9e6e-e5f20c40389f" containerName="mariadb-database-create" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.369090 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="a11aafed-1003-4121-9e6e-e5f20c40389f" containerName="mariadb-database-create" Feb 27 17:21:28 crc kubenswrapper[4700]: E0227 17:21:28.369107 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e67eb82-300d-46bd-b1c6-d2607bf45ccc" containerName="mariadb-database-create" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.369115 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e67eb82-300d-46bd-b1c6-d2607bf45ccc" containerName="mariadb-database-create" Feb 27 17:21:28 crc kubenswrapper[4700]: E0227 17:21:28.369128 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5d89c26-6133-4987-8c61-cc19c80567e0" containerName="mariadb-database-create" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.369136 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5d89c26-6133-4987-8c61-cc19c80567e0" containerName="mariadb-database-create" Feb 27 17:21:28 crc kubenswrapper[4700]: E0227 17:21:28.369149 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a238beaf-451f-452e-a30f-09779ae217c1" containerName="keystone-db-sync" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.369157 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="a238beaf-451f-452e-a30f-09779ae217c1" containerName="keystone-db-sync" Feb 27 17:21:28 crc kubenswrapper[4700]: E0227 17:21:28.369170 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da70ae3e-1bde-47f1-ab27-70ac8726f421" containerName="dnsmasq-dns" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.369179 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="da70ae3e-1bde-47f1-ab27-70ac8726f421" containerName="dnsmasq-dns" Feb 27 17:21:28 crc kubenswrapper[4700]: E0227 17:21:28.369191 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16189a3f-0a2b-453c-8bb1-8e8c4d79c13f" containerName="watcher-db-sync" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.369199 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="16189a3f-0a2b-453c-8bb1-8e8c4d79c13f" containerName="watcher-db-sync" Feb 27 17:21:28 crc kubenswrapper[4700]: E0227 17:21:28.369216 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da70ae3e-1bde-47f1-ab27-70ac8726f421" containerName="init" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.369224 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="da70ae3e-1bde-47f1-ab27-70ac8726f421" containerName="init" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.369412 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e67eb82-300d-46bd-b1c6-d2607bf45ccc" containerName="mariadb-database-create" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.369429 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5d89c26-6133-4987-8c61-cc19c80567e0" containerName="mariadb-database-create" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.369447 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="2062608c-6759-45bd-9879-2a96af693b31" containerName="mariadb-account-create-update" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.369489 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffc04905-1f57-4888-bb88-db5187753b5f" containerName="mariadb-account-create-update" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.369518 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe426647-f5e6-46bf-b187-8f8cb62b06e1" containerName="mariadb-account-create-update" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.369536 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="16189a3f-0a2b-453c-8bb1-8e8c4d79c13f" containerName="watcher-db-sync" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.369555 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="a11aafed-1003-4121-9e6e-e5f20c40389f" containerName="mariadb-database-create" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.369583 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="a238beaf-451f-452e-a30f-09779ae217c1" containerName="keystone-db-sync" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.369604 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="da70ae3e-1bde-47f1-ab27-70ac8726f421" containerName="dnsmasq-dns" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.370328 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9hq27" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.375798 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.375930 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.376799 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.377091 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.377233 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-llljd" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.387093 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7687748657-9lv22"] Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.389032 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7687748657-9lv22" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.412536 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-9hq27"] Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.429541 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7687748657-9lv22"] Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.477601 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-decision-engine-0"] Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.479049 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.482777 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-decision-engine-config-data" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.483670 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-watcher-dockercfg-qc4hm" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.485632 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.488265 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cfded842-e9db-4883-8501-76c145505e47-ovsdbserver-sb\") pod \"dnsmasq-dns-7687748657-9lv22\" (UID: \"cfded842-e9db-4883-8501-76c145505e47\") " pod="openstack/dnsmasq-dns-7687748657-9lv22" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.488363 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb266c49-545c-4e58-a9f8-6dede53accdf-scripts\") pod \"keystone-bootstrap-9hq27\" (UID: \"bb266c49-545c-4e58-a9f8-6dede53accdf\") " pod="openstack/keystone-bootstrap-9hq27" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.488406 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dbd5\" (UniqueName: \"kubernetes.io/projected/bb266c49-545c-4e58-a9f8-6dede53accdf-kube-api-access-5dbd5\") pod \"keystone-bootstrap-9hq27\" (UID: \"bb266c49-545c-4e58-a9f8-6dede53accdf\") " pod="openstack/keystone-bootstrap-9hq27" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.488436 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cfded842-e9db-4883-8501-76c145505e47-ovsdbserver-nb\") pod \"dnsmasq-dns-7687748657-9lv22\" (UID: \"cfded842-e9db-4883-8501-76c145505e47\") " pod="openstack/dnsmasq-dns-7687748657-9lv22" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.488505 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nlws\" (UniqueName: \"kubernetes.io/projected/cfded842-e9db-4883-8501-76c145505e47-kube-api-access-7nlws\") pod \"dnsmasq-dns-7687748657-9lv22\" (UID: \"cfded842-e9db-4883-8501-76c145505e47\") " pod="openstack/dnsmasq-dns-7687748657-9lv22" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.488538 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/bb266c49-545c-4e58-a9f8-6dede53accdf-credential-keys\") pod \"keystone-bootstrap-9hq27\" (UID: \"bb266c49-545c-4e58-a9f8-6dede53accdf\") " pod="openstack/keystone-bootstrap-9hq27" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.488595 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb266c49-545c-4e58-a9f8-6dede53accdf-config-data\") pod \"keystone-bootstrap-9hq27\" (UID: \"bb266c49-545c-4e58-a9f8-6dede53accdf\") " pod="openstack/keystone-bootstrap-9hq27" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.488635 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfded842-e9db-4883-8501-76c145505e47-config\") pod \"dnsmasq-dns-7687748657-9lv22\" (UID: \"cfded842-e9db-4883-8501-76c145505e47\") " pod="openstack/dnsmasq-dns-7687748657-9lv22" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.488662 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cfded842-e9db-4883-8501-76c145505e47-dns-svc\") pod \"dnsmasq-dns-7687748657-9lv22\" (UID: \"cfded842-e9db-4883-8501-76c145505e47\") " pod="openstack/dnsmasq-dns-7687748657-9lv22" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.488722 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bb266c49-545c-4e58-a9f8-6dede53accdf-fernet-keys\") pod \"keystone-bootstrap-9hq27\" (UID: \"bb266c49-545c-4e58-a9f8-6dede53accdf\") " pod="openstack/keystone-bootstrap-9hq27" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.488765 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cfded842-e9db-4883-8501-76c145505e47-dns-swift-storage-0\") pod \"dnsmasq-dns-7687748657-9lv22\" (UID: \"cfded842-e9db-4883-8501-76c145505e47\") " pod="openstack/dnsmasq-dns-7687748657-9lv22" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.488792 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb266c49-545c-4e58-a9f8-6dede53accdf-combined-ca-bundle\") pod \"keystone-bootstrap-9hq27\" (UID: \"bb266c49-545c-4e58-a9f8-6dede53accdf\") " pod="openstack/keystone-bootstrap-9hq27" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.514830 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-applier-0"] Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.516011 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.520937 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-applier-config-data" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.525583 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-api-0"] Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.534801 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.543176 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-api-config-data" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.555909 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.589903 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cfded842-e9db-4883-8501-76c145505e47-ovsdbserver-sb\") pod \"dnsmasq-dns-7687748657-9lv22\" (UID: \"cfded842-e9db-4883-8501-76c145505e47\") " pod="openstack/dnsmasq-dns-7687748657-9lv22" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.590197 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8e03836-cf5d-4ac9-80fe-97556117e65e-logs\") pod \"watcher-decision-engine-0\" (UID: \"d8e03836-cf5d-4ac9-80fe-97556117e65e\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.590300 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb266c49-545c-4e58-a9f8-6dede53accdf-scripts\") pod \"keystone-bootstrap-9hq27\" (UID: \"bb266c49-545c-4e58-a9f8-6dede53accdf\") " pod="openstack/keystone-bootstrap-9hq27" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.590376 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dbd5\" (UniqueName: \"kubernetes.io/projected/bb266c49-545c-4e58-a9f8-6dede53accdf-kube-api-access-5dbd5\") pod \"keystone-bootstrap-9hq27\" (UID: \"bb266c49-545c-4e58-a9f8-6dede53accdf\") " pod="openstack/keystone-bootstrap-9hq27" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.590441 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/d8e03836-cf5d-4ac9-80fe-97556117e65e-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"d8e03836-cf5d-4ac9-80fe-97556117e65e\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.590523 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cfded842-e9db-4883-8501-76c145505e47-ovsdbserver-nb\") pod \"dnsmasq-dns-7687748657-9lv22\" (UID: \"cfded842-e9db-4883-8501-76c145505e47\") " pod="openstack/dnsmasq-dns-7687748657-9lv22" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.590605 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8e03836-cf5d-4ac9-80fe-97556117e65e-config-data\") pod \"watcher-decision-engine-0\" (UID: \"d8e03836-cf5d-4ac9-80fe-97556117e65e\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.590670 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nlws\" (UniqueName: \"kubernetes.io/projected/cfded842-e9db-4883-8501-76c145505e47-kube-api-access-7nlws\") pod \"dnsmasq-dns-7687748657-9lv22\" (UID: \"cfded842-e9db-4883-8501-76c145505e47\") " pod="openstack/dnsmasq-dns-7687748657-9lv22" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.590734 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/bb266c49-545c-4e58-a9f8-6dede53accdf-credential-keys\") pod \"keystone-bootstrap-9hq27\" (UID: \"bb266c49-545c-4e58-a9f8-6dede53accdf\") " pod="openstack/keystone-bootstrap-9hq27" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.590806 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4099c554-cf42-48b3-88cb-e672a66f1a05-logs\") pod \"watcher-applier-0\" (UID: \"4099c554-cf42-48b3-88cb-e672a66f1a05\") " pod="openstack/watcher-applier-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.590875 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb266c49-545c-4e58-a9f8-6dede53accdf-config-data\") pod \"keystone-bootstrap-9hq27\" (UID: \"bb266c49-545c-4e58-a9f8-6dede53accdf\") " pod="openstack/keystone-bootstrap-9hq27" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.590948 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfded842-e9db-4883-8501-76c145505e47-config\") pod \"dnsmasq-dns-7687748657-9lv22\" (UID: \"cfded842-e9db-4883-8501-76c145505e47\") " pod="openstack/dnsmasq-dns-7687748657-9lv22" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.591021 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cfded842-e9db-4883-8501-76c145505e47-dns-svc\") pod \"dnsmasq-dns-7687748657-9lv22\" (UID: \"cfded842-e9db-4883-8501-76c145505e47\") " pod="openstack/dnsmasq-dns-7687748657-9lv22" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.591097 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4099c554-cf42-48b3-88cb-e672a66f1a05-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"4099c554-cf42-48b3-88cb-e672a66f1a05\") " pod="openstack/watcher-applier-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.591161 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcz7f\" (UniqueName: \"kubernetes.io/projected/4099c554-cf42-48b3-88cb-e672a66f1a05-kube-api-access-fcz7f\") pod \"watcher-applier-0\" (UID: \"4099c554-cf42-48b3-88cb-e672a66f1a05\") " pod="openstack/watcher-applier-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.591234 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bb266c49-545c-4e58-a9f8-6dede53accdf-fernet-keys\") pod \"keystone-bootstrap-9hq27\" (UID: \"bb266c49-545c-4e58-a9f8-6dede53accdf\") " pod="openstack/keystone-bootstrap-9hq27" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.591303 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8e03836-cf5d-4ac9-80fe-97556117e65e-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"d8e03836-cf5d-4ac9-80fe-97556117e65e\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.591366 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gnlv\" (UniqueName: \"kubernetes.io/projected/d8e03836-cf5d-4ac9-80fe-97556117e65e-kube-api-access-6gnlv\") pod \"watcher-decision-engine-0\" (UID: \"d8e03836-cf5d-4ac9-80fe-97556117e65e\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.591430 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cfded842-e9db-4883-8501-76c145505e47-dns-swift-storage-0\") pod \"dnsmasq-dns-7687748657-9lv22\" (UID: \"cfded842-e9db-4883-8501-76c145505e47\") " pod="openstack/dnsmasq-dns-7687748657-9lv22" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.591508 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb266c49-545c-4e58-a9f8-6dede53accdf-combined-ca-bundle\") pod \"keystone-bootstrap-9hq27\" (UID: \"bb266c49-545c-4e58-a9f8-6dede53accdf\") " pod="openstack/keystone-bootstrap-9hq27" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.591588 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4099c554-cf42-48b3-88cb-e672a66f1a05-config-data\") pod \"watcher-applier-0\" (UID: \"4099c554-cf42-48b3-88cb-e672a66f1a05\") " pod="openstack/watcher-applier-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.592110 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cfded842-e9db-4883-8501-76c145505e47-ovsdbserver-nb\") pod \"dnsmasq-dns-7687748657-9lv22\" (UID: \"cfded842-e9db-4883-8501-76c145505e47\") " pod="openstack/dnsmasq-dns-7687748657-9lv22" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.594198 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfded842-e9db-4883-8501-76c145505e47-config\") pod \"dnsmasq-dns-7687748657-9lv22\" (UID: \"cfded842-e9db-4883-8501-76c145505e47\") " pod="openstack/dnsmasq-dns-7687748657-9lv22" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.590987 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cfded842-e9db-4883-8501-76c145505e47-ovsdbserver-sb\") pod \"dnsmasq-dns-7687748657-9lv22\" (UID: \"cfded842-e9db-4883-8501-76c145505e47\") " pod="openstack/dnsmasq-dns-7687748657-9lv22" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.594820 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cfded842-e9db-4883-8501-76c145505e47-dns-svc\") pod \"dnsmasq-dns-7687748657-9lv22\" (UID: \"cfded842-e9db-4883-8501-76c145505e47\") " pod="openstack/dnsmasq-dns-7687748657-9lv22" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.601190 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cfded842-e9db-4883-8501-76c145505e47-dns-swift-storage-0\") pod \"dnsmasq-dns-7687748657-9lv22\" (UID: \"cfded842-e9db-4883-8501-76c145505e47\") " pod="openstack/dnsmasq-dns-7687748657-9lv22" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.617504 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb266c49-545c-4e58-a9f8-6dede53accdf-combined-ca-bundle\") pod \"keystone-bootstrap-9hq27\" (UID: \"bb266c49-545c-4e58-a9f8-6dede53accdf\") " pod="openstack/keystone-bootstrap-9hq27" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.629525 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.643045 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb266c49-545c-4e58-a9f8-6dede53accdf-scripts\") pod \"keystone-bootstrap-9hq27\" (UID: \"bb266c49-545c-4e58-a9f8-6dede53accdf\") " pod="openstack/keystone-bootstrap-9hq27" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.644040 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bb266c49-545c-4e58-a9f8-6dede53accdf-fernet-keys\") pod \"keystone-bootstrap-9hq27\" (UID: \"bb266c49-545c-4e58-a9f8-6dede53accdf\") " pod="openstack/keystone-bootstrap-9hq27" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.645133 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/bb266c49-545c-4e58-a9f8-6dede53accdf-credential-keys\") pod \"keystone-bootstrap-9hq27\" (UID: \"bb266c49-545c-4e58-a9f8-6dede53accdf\") " pod="openstack/keystone-bootstrap-9hq27" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.648167 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb266c49-545c-4e58-a9f8-6dede53accdf-config-data\") pod \"keystone-bootstrap-9hq27\" (UID: \"bb266c49-545c-4e58-a9f8-6dede53accdf\") " pod="openstack/keystone-bootstrap-9hq27" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.662859 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dbd5\" (UniqueName: \"kubernetes.io/projected/bb266c49-545c-4e58-a9f8-6dede53accdf-kube-api-access-5dbd5\") pod \"keystone-bootstrap-9hq27\" (UID: \"bb266c49-545c-4e58-a9f8-6dede53accdf\") " pod="openstack/keystone-bootstrap-9hq27" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.665549 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6fbcf8869-6nmv8"] Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.666955 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6fbcf8869-6nmv8" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.673890 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.674073 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-69v8d" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.674299 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.679235 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nlws\" (UniqueName: \"kubernetes.io/projected/cfded842-e9db-4883-8501-76c145505e47-kube-api-access-7nlws\") pod \"dnsmasq-dns-7687748657-9lv22\" (UID: \"cfded842-e9db-4883-8501-76c145505e47\") " pod="openstack/dnsmasq-dns-7687748657-9lv22" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.692888 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08c973c7-192d-4326-9630-aa18aea6a484-config-data\") pod \"watcher-api-0\" (UID: \"08c973c7-192d-4326-9630-aa18aea6a484\") " pod="openstack/watcher-api-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.692933 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4099c554-cf42-48b3-88cb-e672a66f1a05-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"4099c554-cf42-48b3-88cb-e672a66f1a05\") " pod="openstack/watcher-applier-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.692950 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcz7f\" (UniqueName: \"kubernetes.io/projected/4099c554-cf42-48b3-88cb-e672a66f1a05-kube-api-access-fcz7f\") pod \"watcher-applier-0\" (UID: \"4099c554-cf42-48b3-88cb-e672a66f1a05\") " pod="openstack/watcher-applier-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.692980 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8e03836-cf5d-4ac9-80fe-97556117e65e-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"d8e03836-cf5d-4ac9-80fe-97556117e65e\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.693000 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gnlv\" (UniqueName: \"kubernetes.io/projected/d8e03836-cf5d-4ac9-80fe-97556117e65e-kube-api-access-6gnlv\") pod \"watcher-decision-engine-0\" (UID: \"d8e03836-cf5d-4ac9-80fe-97556117e65e\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.693022 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/08c973c7-192d-4326-9630-aa18aea6a484-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"08c973c7-192d-4326-9630-aa18aea6a484\") " pod="openstack/watcher-api-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.693039 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08c973c7-192d-4326-9630-aa18aea6a484-logs\") pod \"watcher-api-0\" (UID: \"08c973c7-192d-4326-9630-aa18aea6a484\") " pod="openstack/watcher-api-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.693063 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4099c554-cf42-48b3-88cb-e672a66f1a05-config-data\") pod \"watcher-applier-0\" (UID: \"4099c554-cf42-48b3-88cb-e672a66f1a05\") " pod="openstack/watcher-applier-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.693094 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8e03836-cf5d-4ac9-80fe-97556117e65e-logs\") pod \"watcher-decision-engine-0\" (UID: \"d8e03836-cf5d-4ac9-80fe-97556117e65e\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.693300 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08c973c7-192d-4326-9630-aa18aea6a484-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"08c973c7-192d-4326-9630-aa18aea6a484\") " pod="openstack/watcher-api-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.693323 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/d8e03836-cf5d-4ac9-80fe-97556117e65e-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"d8e03836-cf5d-4ac9-80fe-97556117e65e\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.693352 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8e03836-cf5d-4ac9-80fe-97556117e65e-config-data\") pod \"watcher-decision-engine-0\" (UID: \"d8e03836-cf5d-4ac9-80fe-97556117e65e\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.693381 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4099c554-cf42-48b3-88cb-e672a66f1a05-logs\") pod \"watcher-applier-0\" (UID: \"4099c554-cf42-48b3-88cb-e672a66f1a05\") " pod="openstack/watcher-applier-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.693402 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfll7\" (UniqueName: \"kubernetes.io/projected/08c973c7-192d-4326-9630-aa18aea6a484-kube-api-access-qfll7\") pod \"watcher-api-0\" (UID: \"08c973c7-192d-4326-9630-aa18aea6a484\") " pod="openstack/watcher-api-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.695810 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.697775 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8e03836-cf5d-4ac9-80fe-97556117e65e-logs\") pod \"watcher-decision-engine-0\" (UID: \"d8e03836-cf5d-4ac9-80fe-97556117e65e\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.697880 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4099c554-cf42-48b3-88cb-e672a66f1a05-logs\") pod \"watcher-applier-0\" (UID: \"4099c554-cf42-48b3-88cb-e672a66f1a05\") " pod="openstack/watcher-applier-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.698844 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4099c554-cf42-48b3-88cb-e672a66f1a05-config-data\") pod \"watcher-applier-0\" (UID: \"4099c554-cf42-48b3-88cb-e672a66f1a05\") " pod="openstack/watcher-applier-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.702564 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4099c554-cf42-48b3-88cb-e672a66f1a05-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"4099c554-cf42-48b3-88cb-e672a66f1a05\") " pod="openstack/watcher-applier-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.704754 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8e03836-cf5d-4ac9-80fe-97556117e65e-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"d8e03836-cf5d-4ac9-80fe-97556117e65e\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.716142 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/d8e03836-cf5d-4ac9-80fe-97556117e65e-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"d8e03836-cf5d-4ac9-80fe-97556117e65e\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.717898 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9hq27" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.726831 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8e03836-cf5d-4ac9-80fe-97556117e65e-config-data\") pod \"watcher-decision-engine-0\" (UID: \"d8e03836-cf5d-4ac9-80fe-97556117e65e\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.727094 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7687748657-9lv22" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.735978 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcz7f\" (UniqueName: \"kubernetes.io/projected/4099c554-cf42-48b3-88cb-e672a66f1a05-kube-api-access-fcz7f\") pod \"watcher-applier-0\" (UID: \"4099c554-cf42-48b3-88cb-e672a66f1a05\") " pod="openstack/watcher-applier-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.745662 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gnlv\" (UniqueName: \"kubernetes.io/projected/d8e03836-cf5d-4ac9-80fe-97556117e65e-kube-api-access-6gnlv\") pod \"watcher-decision-engine-0\" (UID: \"d8e03836-cf5d-4ac9-80fe-97556117e65e\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.751916 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-6b5ck"] Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.756736 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-6b5ck" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.761833 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.762016 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.762197 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-qxjr4" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.785708 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-6b5ck"] Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.799817 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e-logs\") pod \"horizon-6fbcf8869-6nmv8\" (UID: \"9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e\") " pod="openstack/horizon-6fbcf8869-6nmv8" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.799857 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e-scripts\") pod \"horizon-6fbcf8869-6nmv8\" (UID: \"9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e\") " pod="openstack/horizon-6fbcf8869-6nmv8" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.799877 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65vhc\" (UniqueName: \"kubernetes.io/projected/9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e-kube-api-access-65vhc\") pod \"horizon-6fbcf8869-6nmv8\" (UID: \"9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e\") " pod="openstack/horizon-6fbcf8869-6nmv8" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.799910 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08c973c7-192d-4326-9630-aa18aea6a484-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"08c973c7-192d-4326-9630-aa18aea6a484\") " pod="openstack/watcher-api-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.799967 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e-config-data\") pod \"horizon-6fbcf8869-6nmv8\" (UID: \"9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e\") " pod="openstack/horizon-6fbcf8869-6nmv8" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.799996 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfll7\" (UniqueName: \"kubernetes.io/projected/08c973c7-192d-4326-9630-aa18aea6a484-kube-api-access-qfll7\") pod \"watcher-api-0\" (UID: \"08c973c7-192d-4326-9630-aa18aea6a484\") " pod="openstack/watcher-api-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.800011 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e-horizon-secret-key\") pod \"horizon-6fbcf8869-6nmv8\" (UID: \"9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e\") " pod="openstack/horizon-6fbcf8869-6nmv8" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.800042 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08c973c7-192d-4326-9630-aa18aea6a484-config-data\") pod \"watcher-api-0\" (UID: \"08c973c7-192d-4326-9630-aa18aea6a484\") " pod="openstack/watcher-api-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.800084 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/08c973c7-192d-4326-9630-aa18aea6a484-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"08c973c7-192d-4326-9630-aa18aea6a484\") " pod="openstack/watcher-api-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.800103 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08c973c7-192d-4326-9630-aa18aea6a484-logs\") pod \"watcher-api-0\" (UID: \"08c973c7-192d-4326-9630-aa18aea6a484\") " pod="openstack/watcher-api-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.800435 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08c973c7-192d-4326-9630-aa18aea6a484-logs\") pod \"watcher-api-0\" (UID: \"08c973c7-192d-4326-9630-aa18aea6a484\") " pod="openstack/watcher-api-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.803873 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08c973c7-192d-4326-9630-aa18aea6a484-config-data\") pod \"watcher-api-0\" (UID: \"08c973c7-192d-4326-9630-aa18aea6a484\") " pod="openstack/watcher-api-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.824939 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/08c973c7-192d-4326-9630-aa18aea6a484-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"08c973c7-192d-4326-9630-aa18aea6a484\") " pod="openstack/watcher-api-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.829930 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.850593 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08c973c7-192d-4326-9630-aa18aea6a484-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"08c973c7-192d-4326-9630-aa18aea6a484\") " pod="openstack/watcher-api-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.860521 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6fbcf8869-6nmv8"] Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.874987 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.876095 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfll7\" (UniqueName: \"kubernetes.io/projected/08c973c7-192d-4326-9630-aa18aea6a484-kube-api-access-qfll7\") pod \"watcher-api-0\" (UID: \"08c973c7-192d-4326-9630-aa18aea6a484\") " pod="openstack/watcher-api-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.892972 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.906924 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-mhcf5"] Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.908011 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-mhcf5" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.913023 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e-config-data\") pod \"horizon-6fbcf8869-6nmv8\" (UID: \"9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e\") " pod="openstack/horizon-6fbcf8869-6nmv8" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.913091 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e-horizon-secret-key\") pod \"horizon-6fbcf8869-6nmv8\" (UID: \"9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e\") " pod="openstack/horizon-6fbcf8869-6nmv8" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.913172 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e-logs\") pod \"horizon-6fbcf8869-6nmv8\" (UID: \"9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e\") " pod="openstack/horizon-6fbcf8869-6nmv8" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.913197 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glksm\" (UniqueName: \"kubernetes.io/projected/3f209c04-79ea-4048-9d31-6222874de091-kube-api-access-glksm\") pod \"neutron-db-sync-6b5ck\" (UID: \"3f209c04-79ea-4048-9d31-6222874de091\") " pod="openstack/neutron-db-sync-6b5ck" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.913218 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e-scripts\") pod \"horizon-6fbcf8869-6nmv8\" (UID: \"9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e\") " pod="openstack/horizon-6fbcf8869-6nmv8" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.913235 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65vhc\" (UniqueName: \"kubernetes.io/projected/9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e-kube-api-access-65vhc\") pod \"horizon-6fbcf8869-6nmv8\" (UID: \"9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e\") " pod="openstack/horizon-6fbcf8869-6nmv8" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.913268 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f209c04-79ea-4048-9d31-6222874de091-combined-ca-bundle\") pod \"neutron-db-sync-6b5ck\" (UID: \"3f209c04-79ea-4048-9d31-6222874de091\") " pod="openstack/neutron-db-sync-6b5ck" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.913288 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3f209c04-79ea-4048-9d31-6222874de091-config\") pod \"neutron-db-sync-6b5ck\" (UID: \"3f209c04-79ea-4048-9d31-6222874de091\") " pod="openstack/neutron-db-sync-6b5ck" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.914486 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e-config-data\") pod \"horizon-6fbcf8869-6nmv8\" (UID: \"9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e\") " pod="openstack/horizon-6fbcf8869-6nmv8" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.918743 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.918907 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.919082 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-wx827" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.919659 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e-scripts\") pod \"horizon-6fbcf8869-6nmv8\" (UID: \"9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e\") " pod="openstack/horizon-6fbcf8869-6nmv8" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.919877 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e-logs\") pod \"horizon-6fbcf8869-6nmv8\" (UID: \"9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e\") " pod="openstack/horizon-6fbcf8869-6nmv8" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.951834 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65vhc\" (UniqueName: \"kubernetes.io/projected/9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e-kube-api-access-65vhc\") pod \"horizon-6fbcf8869-6nmv8\" (UID: \"9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e\") " pod="openstack/horizon-6fbcf8869-6nmv8" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.952850 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e-horizon-secret-key\") pod \"horizon-6fbcf8869-6nmv8\" (UID: \"9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e\") " pod="openstack/horizon-6fbcf8869-6nmv8" Feb 27 17:21:28 crc kubenswrapper[4700]: I0227 17:21:28.962538 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.002656 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.017312 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dd3817a7-c2ab-43f0-9c74-b477ac59fa52-db-sync-config-data\") pod \"cinder-db-sync-mhcf5\" (UID: \"dd3817a7-c2ab-43f0-9c74-b477ac59fa52\") " pod="openstack/cinder-db-sync-mhcf5" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.017373 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glksm\" (UniqueName: \"kubernetes.io/projected/3f209c04-79ea-4048-9d31-6222874de091-kube-api-access-glksm\") pod \"neutron-db-sync-6b5ck\" (UID: \"3f209c04-79ea-4048-9d31-6222874de091\") " pod="openstack/neutron-db-sync-6b5ck" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.017422 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f209c04-79ea-4048-9d31-6222874de091-combined-ca-bundle\") pod \"neutron-db-sync-6b5ck\" (UID: \"3f209c04-79ea-4048-9d31-6222874de091\") " pod="openstack/neutron-db-sync-6b5ck" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.017444 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3f209c04-79ea-4048-9d31-6222874de091-config\") pod \"neutron-db-sync-6b5ck\" (UID: \"3f209c04-79ea-4048-9d31-6222874de091\") " pod="openstack/neutron-db-sync-6b5ck" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.017486 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd3817a7-c2ab-43f0-9c74-b477ac59fa52-config-data\") pod \"cinder-db-sync-mhcf5\" (UID: \"dd3817a7-c2ab-43f0-9c74-b477ac59fa52\") " pod="openstack/cinder-db-sync-mhcf5" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.017517 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nrcq\" (UniqueName: \"kubernetes.io/projected/dd3817a7-c2ab-43f0-9c74-b477ac59fa52-kube-api-access-9nrcq\") pod \"cinder-db-sync-mhcf5\" (UID: \"dd3817a7-c2ab-43f0-9c74-b477ac59fa52\") " pod="openstack/cinder-db-sync-mhcf5" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.017587 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd3817a7-c2ab-43f0-9c74-b477ac59fa52-combined-ca-bundle\") pod \"cinder-db-sync-mhcf5\" (UID: \"dd3817a7-c2ab-43f0-9c74-b477ac59fa52\") " pod="openstack/cinder-db-sync-mhcf5" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.017628 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd3817a7-c2ab-43f0-9c74-b477ac59fa52-scripts\") pod \"cinder-db-sync-mhcf5\" (UID: \"dd3817a7-c2ab-43f0-9c74-b477ac59fa52\") " pod="openstack/cinder-db-sync-mhcf5" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.017675 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dd3817a7-c2ab-43f0-9c74-b477ac59fa52-etc-machine-id\") pod \"cinder-db-sync-mhcf5\" (UID: \"dd3817a7-c2ab-43f0-9c74-b477ac59fa52\") " pod="openstack/cinder-db-sync-mhcf5" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.018712 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.030579 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.037441 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f209c04-79ea-4048-9d31-6222874de091-combined-ca-bundle\") pod \"neutron-db-sync-6b5ck\" (UID: \"3f209c04-79ea-4048-9d31-6222874de091\") " pod="openstack/neutron-db-sync-6b5ck" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.052969 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glksm\" (UniqueName: \"kubernetes.io/projected/3f209c04-79ea-4048-9d31-6222874de091-kube-api-access-glksm\") pod \"neutron-db-sync-6b5ck\" (UID: \"3f209c04-79ea-4048-9d31-6222874de091\") " pod="openstack/neutron-db-sync-6b5ck" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.084185 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/3f209c04-79ea-4048-9d31-6222874de091-config\") pod \"neutron-db-sync-6b5ck\" (UID: \"3f209c04-79ea-4048-9d31-6222874de091\") " pod="openstack/neutron-db-sync-6b5ck" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.136231 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/61baccca-f7e6-4442-b612-8ab97b82036d-log-httpd\") pod \"ceilometer-0\" (UID: \"61baccca-f7e6-4442-b612-8ab97b82036d\") " pod="openstack/ceilometer-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.137389 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dd3817a7-c2ab-43f0-9c74-b477ac59fa52-db-sync-config-data\") pod \"cinder-db-sync-mhcf5\" (UID: \"dd3817a7-c2ab-43f0-9c74-b477ac59fa52\") " pod="openstack/cinder-db-sync-mhcf5" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.137796 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hm7np\" (UniqueName: \"kubernetes.io/projected/61baccca-f7e6-4442-b612-8ab97b82036d-kube-api-access-hm7np\") pod \"ceilometer-0\" (UID: \"61baccca-f7e6-4442-b612-8ab97b82036d\") " pod="openstack/ceilometer-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.137938 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/61baccca-f7e6-4442-b612-8ab97b82036d-run-httpd\") pod \"ceilometer-0\" (UID: \"61baccca-f7e6-4442-b612-8ab97b82036d\") " pod="openstack/ceilometer-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.138122 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61baccca-f7e6-4442-b612-8ab97b82036d-scripts\") pod \"ceilometer-0\" (UID: \"61baccca-f7e6-4442-b612-8ab97b82036d\") " pod="openstack/ceilometer-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.138404 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/61baccca-f7e6-4442-b612-8ab97b82036d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"61baccca-f7e6-4442-b612-8ab97b82036d\") " pod="openstack/ceilometer-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.138445 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd3817a7-c2ab-43f0-9c74-b477ac59fa52-config-data\") pod \"cinder-db-sync-mhcf5\" (UID: \"dd3817a7-c2ab-43f0-9c74-b477ac59fa52\") " pod="openstack/cinder-db-sync-mhcf5" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.138517 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nrcq\" (UniqueName: \"kubernetes.io/projected/dd3817a7-c2ab-43f0-9c74-b477ac59fa52-kube-api-access-9nrcq\") pod \"cinder-db-sync-mhcf5\" (UID: \"dd3817a7-c2ab-43f0-9c74-b477ac59fa52\") " pod="openstack/cinder-db-sync-mhcf5" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.138569 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd3817a7-c2ab-43f0-9c74-b477ac59fa52-combined-ca-bundle\") pod \"cinder-db-sync-mhcf5\" (UID: \"dd3817a7-c2ab-43f0-9c74-b477ac59fa52\") " pod="openstack/cinder-db-sync-mhcf5" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.138602 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61baccca-f7e6-4442-b612-8ab97b82036d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"61baccca-f7e6-4442-b612-8ab97b82036d\") " pod="openstack/ceilometer-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.139065 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd3817a7-c2ab-43f0-9c74-b477ac59fa52-scripts\") pod \"cinder-db-sync-mhcf5\" (UID: \"dd3817a7-c2ab-43f0-9c74-b477ac59fa52\") " pod="openstack/cinder-db-sync-mhcf5" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.139201 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61baccca-f7e6-4442-b612-8ab97b82036d-config-data\") pod \"ceilometer-0\" (UID: \"61baccca-f7e6-4442-b612-8ab97b82036d\") " pod="openstack/ceilometer-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.143175 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd3817a7-c2ab-43f0-9c74-b477ac59fa52-combined-ca-bundle\") pod \"cinder-db-sync-mhcf5\" (UID: \"dd3817a7-c2ab-43f0-9c74-b477ac59fa52\") " pod="openstack/cinder-db-sync-mhcf5" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.145985 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dd3817a7-c2ab-43f0-9c74-b477ac59fa52-etc-machine-id\") pod \"cinder-db-sync-mhcf5\" (UID: \"dd3817a7-c2ab-43f0-9c74-b477ac59fa52\") " pod="openstack/cinder-db-sync-mhcf5" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.146299 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dd3817a7-c2ab-43f0-9c74-b477ac59fa52-etc-machine-id\") pod \"cinder-db-sync-mhcf5\" (UID: \"dd3817a7-c2ab-43f0-9c74-b477ac59fa52\") " pod="openstack/cinder-db-sync-mhcf5" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.152107 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6fbcf8869-6nmv8" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.188931 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dd3817a7-c2ab-43f0-9c74-b477ac59fa52-db-sync-config-data\") pod \"cinder-db-sync-mhcf5\" (UID: \"dd3817a7-c2ab-43f0-9c74-b477ac59fa52\") " pod="openstack/cinder-db-sync-mhcf5" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.189633 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-6b5ck" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.194214 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd3817a7-c2ab-43f0-9c74-b477ac59fa52-scripts\") pod \"cinder-db-sync-mhcf5\" (UID: \"dd3817a7-c2ab-43f0-9c74-b477ac59fa52\") " pod="openstack/cinder-db-sync-mhcf5" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.229673 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd3817a7-c2ab-43f0-9c74-b477ac59fa52-config-data\") pod \"cinder-db-sync-mhcf5\" (UID: \"dd3817a7-c2ab-43f0-9c74-b477ac59fa52\") " pod="openstack/cinder-db-sync-mhcf5" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.235253 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-mhcf5"] Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.235284 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.235300 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-vrb9h"] Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.237482 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-dmf4m"] Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.239727 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dmf4m" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.241197 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-vrb9h" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.247210 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.247516 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-psmcn" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.247821 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-jhr2l" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.248057 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.248306 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.248687 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nrcq\" (UniqueName: \"kubernetes.io/projected/dd3817a7-c2ab-43f0-9c74-b477ac59fa52-kube-api-access-9nrcq\") pod \"cinder-db-sync-mhcf5\" (UID: \"dd3817a7-c2ab-43f0-9c74-b477ac59fa52\") " pod="openstack/cinder-db-sync-mhcf5" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.252231 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/61baccca-f7e6-4442-b612-8ab97b82036d-run-httpd\") pod \"ceilometer-0\" (UID: \"61baccca-f7e6-4442-b612-8ab97b82036d\") " pod="openstack/ceilometer-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.252705 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61baccca-f7e6-4442-b612-8ab97b82036d-scripts\") pod \"ceilometer-0\" (UID: \"61baccca-f7e6-4442-b612-8ab97b82036d\") " pod="openstack/ceilometer-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.252844 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/61baccca-f7e6-4442-b612-8ab97b82036d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"61baccca-f7e6-4442-b612-8ab97b82036d\") " pod="openstack/ceilometer-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.252978 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/61baccca-f7e6-4442-b612-8ab97b82036d-run-httpd\") pod \"ceilometer-0\" (UID: \"61baccca-f7e6-4442-b612-8ab97b82036d\") " pod="openstack/ceilometer-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.252986 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61baccca-f7e6-4442-b612-8ab97b82036d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"61baccca-f7e6-4442-b612-8ab97b82036d\") " pod="openstack/ceilometer-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.253119 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61baccca-f7e6-4442-b612-8ab97b82036d-config-data\") pod \"ceilometer-0\" (UID: \"61baccca-f7e6-4442-b612-8ab97b82036d\") " pod="openstack/ceilometer-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.253211 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/61baccca-f7e6-4442-b612-8ab97b82036d-log-httpd\") pod \"ceilometer-0\" (UID: \"61baccca-f7e6-4442-b612-8ab97b82036d\") " pod="openstack/ceilometer-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.253294 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hm7np\" (UniqueName: \"kubernetes.io/projected/61baccca-f7e6-4442-b612-8ab97b82036d-kube-api-access-hm7np\") pod \"ceilometer-0\" (UID: \"61baccca-f7e6-4442-b612-8ab97b82036d\") " pod="openstack/ceilometer-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.254608 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/61baccca-f7e6-4442-b612-8ab97b82036d-log-httpd\") pod \"ceilometer-0\" (UID: \"61baccca-f7e6-4442-b612-8ab97b82036d\") " pod="openstack/ceilometer-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.257727 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-d7b5ff8c7-zbshw"] Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.259148 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d7b5ff8c7-zbshw" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.261800 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61baccca-f7e6-4442-b612-8ab97b82036d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"61baccca-f7e6-4442-b612-8ab97b82036d\") " pod="openstack/ceilometer-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.264913 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-vrb9h"] Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.274546 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-dmf4m"] Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.279089 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hm7np\" (UniqueName: \"kubernetes.io/projected/61baccca-f7e6-4442-b612-8ab97b82036d-kube-api-access-hm7np\") pod \"ceilometer-0\" (UID: \"61baccca-f7e6-4442-b612-8ab97b82036d\") " pod="openstack/ceilometer-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.280832 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61baccca-f7e6-4442-b612-8ab97b82036d-config-data\") pod \"ceilometer-0\" (UID: \"61baccca-f7e6-4442-b612-8ab97b82036d\") " pod="openstack/ceilometer-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.281277 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/61baccca-f7e6-4442-b612-8ab97b82036d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"61baccca-f7e6-4442-b612-8ab97b82036d\") " pod="openstack/ceilometer-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.283190 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-d7b5ff8c7-zbshw"] Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.286935 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-mhcf5" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.291996 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7687748657-9lv22"] Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.293954 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61baccca-f7e6-4442-b612-8ab97b82036d-scripts\") pod \"ceilometer-0\" (UID: \"61baccca-f7e6-4442-b612-8ab97b82036d\") " pod="openstack/ceilometer-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.295798 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.302425 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.304107 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.306263 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.306320 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-758nx" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.308742 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.308830 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.320926 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-787484f569-fhzcg"] Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.322591 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-787484f569-fhzcg" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.328955 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.340838 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-787484f569-fhzcg"] Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.354739 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/94005464-5a50-42ba-bb5f-b69b66cdec58-scripts\") pod \"horizon-d7b5ff8c7-zbshw\" (UID: \"94005464-5a50-42ba-bb5f-b69b66cdec58\") " pod="openstack/horizon-d7b5ff8c7-zbshw" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.354953 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/989eacab-9c48-4375-b7a0-f2c7f40d17ca-scripts\") pod \"placement-db-sync-dmf4m\" (UID: \"989eacab-9c48-4375-b7a0-f2c7f40d17ca\") " pod="openstack/placement-db-sync-dmf4m" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.355052 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f72833f0-cf5f-4bf8-bb5f-1f65477c4b76-combined-ca-bundle\") pod \"barbican-db-sync-vrb9h\" (UID: \"f72833f0-cf5f-4bf8-bb5f-1f65477c4b76\") " pod="openstack/barbican-db-sync-vrb9h" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.355173 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkqvc\" (UniqueName: \"kubernetes.io/projected/989eacab-9c48-4375-b7a0-f2c7f40d17ca-kube-api-access-nkqvc\") pod \"placement-db-sync-dmf4m\" (UID: \"989eacab-9c48-4375-b7a0-f2c7f40d17ca\") " pod="openstack/placement-db-sync-dmf4m" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.355264 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94005464-5a50-42ba-bb5f-b69b66cdec58-logs\") pod \"horizon-d7b5ff8c7-zbshw\" (UID: \"94005464-5a50-42ba-bb5f-b69b66cdec58\") " pod="openstack/horizon-d7b5ff8c7-zbshw" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.355326 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/989eacab-9c48-4375-b7a0-f2c7f40d17ca-combined-ca-bundle\") pod \"placement-db-sync-dmf4m\" (UID: \"989eacab-9c48-4375-b7a0-f2c7f40d17ca\") " pod="openstack/placement-db-sync-dmf4m" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.355392 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/989eacab-9c48-4375-b7a0-f2c7f40d17ca-logs\") pod \"placement-db-sync-dmf4m\" (UID: \"989eacab-9c48-4375-b7a0-f2c7f40d17ca\") " pod="openstack/placement-db-sync-dmf4m" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.355481 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f72833f0-cf5f-4bf8-bb5f-1f65477c4b76-db-sync-config-data\") pod \"barbican-db-sync-vrb9h\" (UID: \"f72833f0-cf5f-4bf8-bb5f-1f65477c4b76\") " pod="openstack/barbican-db-sync-vrb9h" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.355610 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7xnh\" (UniqueName: \"kubernetes.io/projected/94005464-5a50-42ba-bb5f-b69b66cdec58-kube-api-access-d7xnh\") pod \"horizon-d7b5ff8c7-zbshw\" (UID: \"94005464-5a50-42ba-bb5f-b69b66cdec58\") " pod="openstack/horizon-d7b5ff8c7-zbshw" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.355680 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sssn7\" (UniqueName: \"kubernetes.io/projected/f72833f0-cf5f-4bf8-bb5f-1f65477c4b76-kube-api-access-sssn7\") pod \"barbican-db-sync-vrb9h\" (UID: \"f72833f0-cf5f-4bf8-bb5f-1f65477c4b76\") " pod="openstack/barbican-db-sync-vrb9h" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.355748 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/94005464-5a50-42ba-bb5f-b69b66cdec58-config-data\") pod \"horizon-d7b5ff8c7-zbshw\" (UID: \"94005464-5a50-42ba-bb5f-b69b66cdec58\") " pod="openstack/horizon-d7b5ff8c7-zbshw" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.355838 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/94005464-5a50-42ba-bb5f-b69b66cdec58-horizon-secret-key\") pod \"horizon-d7b5ff8c7-zbshw\" (UID: \"94005464-5a50-42ba-bb5f-b69b66cdec58\") " pod="openstack/horizon-d7b5ff8c7-zbshw" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.355909 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/989eacab-9c48-4375-b7a0-f2c7f40d17ca-config-data\") pod \"placement-db-sync-dmf4m\" (UID: \"989eacab-9c48-4375-b7a0-f2c7f40d17ca\") " pod="openstack/placement-db-sync-dmf4m" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.458617 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f72833f0-cf5f-4bf8-bb5f-1f65477c4b76-db-sync-config-data\") pod \"barbican-db-sync-vrb9h\" (UID: \"f72833f0-cf5f-4bf8-bb5f-1f65477c4b76\") " pod="openstack/barbican-db-sync-vrb9h" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.458903 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5810478f-8400-4b6d-b647-e7294aa087ba-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5810478f-8400-4b6d-b647-e7294aa087ba\") " pod="openstack/glance-default-external-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.458938 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7xnh\" (UniqueName: \"kubernetes.io/projected/94005464-5a50-42ba-bb5f-b69b66cdec58-kube-api-access-d7xnh\") pod \"horizon-d7b5ff8c7-zbshw\" (UID: \"94005464-5a50-42ba-bb5f-b69b66cdec58\") " pod="openstack/horizon-d7b5ff8c7-zbshw" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.458968 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5810478f-8400-4b6d-b647-e7294aa087ba-config-data\") pod \"glance-default-external-api-0\" (UID: \"5810478f-8400-4b6d-b647-e7294aa087ba\") " pod="openstack/glance-default-external-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.458989 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sssn7\" (UniqueName: \"kubernetes.io/projected/f72833f0-cf5f-4bf8-bb5f-1f65477c4b76-kube-api-access-sssn7\") pod \"barbican-db-sync-vrb9h\" (UID: \"f72833f0-cf5f-4bf8-bb5f-1f65477c4b76\") " pod="openstack/barbican-db-sync-vrb9h" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.459018 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5810478f-8400-4b6d-b647-e7294aa087ba-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5810478f-8400-4b6d-b647-e7294aa087ba\") " pod="openstack/glance-default-external-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.459038 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/94005464-5a50-42ba-bb5f-b69b66cdec58-config-data\") pod \"horizon-d7b5ff8c7-zbshw\" (UID: \"94005464-5a50-42ba-bb5f-b69b66cdec58\") " pod="openstack/horizon-d7b5ff8c7-zbshw" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.459080 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"5810478f-8400-4b6d-b647-e7294aa087ba\") " pod="openstack/glance-default-external-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.459109 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/94005464-5a50-42ba-bb5f-b69b66cdec58-horizon-secret-key\") pod \"horizon-d7b5ff8c7-zbshw\" (UID: \"94005464-5a50-42ba-bb5f-b69b66cdec58\") " pod="openstack/horizon-d7b5ff8c7-zbshw" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.459127 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mllb4\" (UniqueName: \"kubernetes.io/projected/7d4f9f19-62e5-4bcd-957e-345c18380eac-kube-api-access-mllb4\") pod \"dnsmasq-dns-787484f569-fhzcg\" (UID: \"7d4f9f19-62e5-4bcd-957e-345c18380eac\") " pod="openstack/dnsmasq-dns-787484f569-fhzcg" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.459147 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/989eacab-9c48-4375-b7a0-f2c7f40d17ca-config-data\") pod \"placement-db-sync-dmf4m\" (UID: \"989eacab-9c48-4375-b7a0-f2c7f40d17ca\") " pod="openstack/placement-db-sync-dmf4m" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.459167 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d4f9f19-62e5-4bcd-957e-345c18380eac-ovsdbserver-nb\") pod \"dnsmasq-dns-787484f569-fhzcg\" (UID: \"7d4f9f19-62e5-4bcd-957e-345c18380eac\") " pod="openstack/dnsmasq-dns-787484f569-fhzcg" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.459185 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5810478f-8400-4b6d-b647-e7294aa087ba-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5810478f-8400-4b6d-b647-e7294aa087ba\") " pod="openstack/glance-default-external-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.459216 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/94005464-5a50-42ba-bb5f-b69b66cdec58-scripts\") pod \"horizon-d7b5ff8c7-zbshw\" (UID: \"94005464-5a50-42ba-bb5f-b69b66cdec58\") " pod="openstack/horizon-d7b5ff8c7-zbshw" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.459237 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/989eacab-9c48-4375-b7a0-f2c7f40d17ca-scripts\") pod \"placement-db-sync-dmf4m\" (UID: \"989eacab-9c48-4375-b7a0-f2c7f40d17ca\") " pod="openstack/placement-db-sync-dmf4m" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.459266 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5810478f-8400-4b6d-b647-e7294aa087ba-scripts\") pod \"glance-default-external-api-0\" (UID: \"5810478f-8400-4b6d-b647-e7294aa087ba\") " pod="openstack/glance-default-external-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.459288 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7d4f9f19-62e5-4bcd-957e-345c18380eac-ovsdbserver-sb\") pod \"dnsmasq-dns-787484f569-fhzcg\" (UID: \"7d4f9f19-62e5-4bcd-957e-345c18380eac\") " pod="openstack/dnsmasq-dns-787484f569-fhzcg" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.459310 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5rmn\" (UniqueName: \"kubernetes.io/projected/5810478f-8400-4b6d-b647-e7294aa087ba-kube-api-access-q5rmn\") pod \"glance-default-external-api-0\" (UID: \"5810478f-8400-4b6d-b647-e7294aa087ba\") " pod="openstack/glance-default-external-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.459339 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f72833f0-cf5f-4bf8-bb5f-1f65477c4b76-combined-ca-bundle\") pod \"barbican-db-sync-vrb9h\" (UID: \"f72833f0-cf5f-4bf8-bb5f-1f65477c4b76\") " pod="openstack/barbican-db-sync-vrb9h" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.459361 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5810478f-8400-4b6d-b647-e7294aa087ba-logs\") pod \"glance-default-external-api-0\" (UID: \"5810478f-8400-4b6d-b647-e7294aa087ba\") " pod="openstack/glance-default-external-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.459390 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d4f9f19-62e5-4bcd-957e-345c18380eac-config\") pod \"dnsmasq-dns-787484f569-fhzcg\" (UID: \"7d4f9f19-62e5-4bcd-957e-345c18380eac\") " pod="openstack/dnsmasq-dns-787484f569-fhzcg" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.459409 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkqvc\" (UniqueName: \"kubernetes.io/projected/989eacab-9c48-4375-b7a0-f2c7f40d17ca-kube-api-access-nkqvc\") pod \"placement-db-sync-dmf4m\" (UID: \"989eacab-9c48-4375-b7a0-f2c7f40d17ca\") " pod="openstack/placement-db-sync-dmf4m" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.459433 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94005464-5a50-42ba-bb5f-b69b66cdec58-logs\") pod \"horizon-d7b5ff8c7-zbshw\" (UID: \"94005464-5a50-42ba-bb5f-b69b66cdec58\") " pod="openstack/horizon-d7b5ff8c7-zbshw" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.459449 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/989eacab-9c48-4375-b7a0-f2c7f40d17ca-combined-ca-bundle\") pod \"placement-db-sync-dmf4m\" (UID: \"989eacab-9c48-4375-b7a0-f2c7f40d17ca\") " pod="openstack/placement-db-sync-dmf4m" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.459476 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d4f9f19-62e5-4bcd-957e-345c18380eac-dns-svc\") pod \"dnsmasq-dns-787484f569-fhzcg\" (UID: \"7d4f9f19-62e5-4bcd-957e-345c18380eac\") " pod="openstack/dnsmasq-dns-787484f569-fhzcg" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.459493 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/989eacab-9c48-4375-b7a0-f2c7f40d17ca-logs\") pod \"placement-db-sync-dmf4m\" (UID: \"989eacab-9c48-4375-b7a0-f2c7f40d17ca\") " pod="openstack/placement-db-sync-dmf4m" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.459515 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7d4f9f19-62e5-4bcd-957e-345c18380eac-dns-swift-storage-0\") pod \"dnsmasq-dns-787484f569-fhzcg\" (UID: \"7d4f9f19-62e5-4bcd-957e-345c18380eac\") " pod="openstack/dnsmasq-dns-787484f569-fhzcg" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.462309 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/94005464-5a50-42ba-bb5f-b69b66cdec58-config-data\") pod \"horizon-d7b5ff8c7-zbshw\" (UID: \"94005464-5a50-42ba-bb5f-b69b66cdec58\") " pod="openstack/horizon-d7b5ff8c7-zbshw" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.467329 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/94005464-5a50-42ba-bb5f-b69b66cdec58-scripts\") pod \"horizon-d7b5ff8c7-zbshw\" (UID: \"94005464-5a50-42ba-bb5f-b69b66cdec58\") " pod="openstack/horizon-d7b5ff8c7-zbshw" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.468095 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94005464-5a50-42ba-bb5f-b69b66cdec58-logs\") pod \"horizon-d7b5ff8c7-zbshw\" (UID: \"94005464-5a50-42ba-bb5f-b69b66cdec58\") " pod="openstack/horizon-d7b5ff8c7-zbshw" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.468152 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/989eacab-9c48-4375-b7a0-f2c7f40d17ca-logs\") pod \"placement-db-sync-dmf4m\" (UID: \"989eacab-9c48-4375-b7a0-f2c7f40d17ca\") " pod="openstack/placement-db-sync-dmf4m" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.472718 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f72833f0-cf5f-4bf8-bb5f-1f65477c4b76-combined-ca-bundle\") pod \"barbican-db-sync-vrb9h\" (UID: \"f72833f0-cf5f-4bf8-bb5f-1f65477c4b76\") " pod="openstack/barbican-db-sync-vrb9h" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.482629 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/989eacab-9c48-4375-b7a0-f2c7f40d17ca-scripts\") pod \"placement-db-sync-dmf4m\" (UID: \"989eacab-9c48-4375-b7a0-f2c7f40d17ca\") " pod="openstack/placement-db-sync-dmf4m" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.482790 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/94005464-5a50-42ba-bb5f-b69b66cdec58-horizon-secret-key\") pod \"horizon-d7b5ff8c7-zbshw\" (UID: \"94005464-5a50-42ba-bb5f-b69b66cdec58\") " pod="openstack/horizon-d7b5ff8c7-zbshw" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.483104 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f72833f0-cf5f-4bf8-bb5f-1f65477c4b76-db-sync-config-data\") pod \"barbican-db-sync-vrb9h\" (UID: \"f72833f0-cf5f-4bf8-bb5f-1f65477c4b76\") " pod="openstack/barbican-db-sync-vrb9h" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.487703 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/989eacab-9c48-4375-b7a0-f2c7f40d17ca-config-data\") pod \"placement-db-sync-dmf4m\" (UID: \"989eacab-9c48-4375-b7a0-f2c7f40d17ca\") " pod="openstack/placement-db-sync-dmf4m" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.488194 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/989eacab-9c48-4375-b7a0-f2c7f40d17ca-combined-ca-bundle\") pod \"placement-db-sync-dmf4m\" (UID: \"989eacab-9c48-4375-b7a0-f2c7f40d17ca\") " pod="openstack/placement-db-sync-dmf4m" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.489202 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7xnh\" (UniqueName: \"kubernetes.io/projected/94005464-5a50-42ba-bb5f-b69b66cdec58-kube-api-access-d7xnh\") pod \"horizon-d7b5ff8c7-zbshw\" (UID: \"94005464-5a50-42ba-bb5f-b69b66cdec58\") " pod="openstack/horizon-d7b5ff8c7-zbshw" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.489931 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sssn7\" (UniqueName: \"kubernetes.io/projected/f72833f0-cf5f-4bf8-bb5f-1f65477c4b76-kube-api-access-sssn7\") pod \"barbican-db-sync-vrb9h\" (UID: \"f72833f0-cf5f-4bf8-bb5f-1f65477c4b76\") " pod="openstack/barbican-db-sync-vrb9h" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.490965 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkqvc\" (UniqueName: \"kubernetes.io/projected/989eacab-9c48-4375-b7a0-f2c7f40d17ca-kube-api-access-nkqvc\") pod \"placement-db-sync-dmf4m\" (UID: \"989eacab-9c48-4375-b7a0-f2c7f40d17ca\") " pod="openstack/placement-db-sync-dmf4m" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.563107 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d4f9f19-62e5-4bcd-957e-345c18380eac-ovsdbserver-nb\") pod \"dnsmasq-dns-787484f569-fhzcg\" (UID: \"7d4f9f19-62e5-4bcd-957e-345c18380eac\") " pod="openstack/dnsmasq-dns-787484f569-fhzcg" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.563163 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5810478f-8400-4b6d-b647-e7294aa087ba-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5810478f-8400-4b6d-b647-e7294aa087ba\") " pod="openstack/glance-default-external-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.563259 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5810478f-8400-4b6d-b647-e7294aa087ba-scripts\") pod \"glance-default-external-api-0\" (UID: \"5810478f-8400-4b6d-b647-e7294aa087ba\") " pod="openstack/glance-default-external-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.563292 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7d4f9f19-62e5-4bcd-957e-345c18380eac-ovsdbserver-sb\") pod \"dnsmasq-dns-787484f569-fhzcg\" (UID: \"7d4f9f19-62e5-4bcd-957e-345c18380eac\") " pod="openstack/dnsmasq-dns-787484f569-fhzcg" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.563326 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5rmn\" (UniqueName: \"kubernetes.io/projected/5810478f-8400-4b6d-b647-e7294aa087ba-kube-api-access-q5rmn\") pod \"glance-default-external-api-0\" (UID: \"5810478f-8400-4b6d-b647-e7294aa087ba\") " pod="openstack/glance-default-external-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.563369 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5810478f-8400-4b6d-b647-e7294aa087ba-logs\") pod \"glance-default-external-api-0\" (UID: \"5810478f-8400-4b6d-b647-e7294aa087ba\") " pod="openstack/glance-default-external-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.563397 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d4f9f19-62e5-4bcd-957e-345c18380eac-config\") pod \"dnsmasq-dns-787484f569-fhzcg\" (UID: \"7d4f9f19-62e5-4bcd-957e-345c18380eac\") " pod="openstack/dnsmasq-dns-787484f569-fhzcg" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.563495 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d4f9f19-62e5-4bcd-957e-345c18380eac-dns-svc\") pod \"dnsmasq-dns-787484f569-fhzcg\" (UID: \"7d4f9f19-62e5-4bcd-957e-345c18380eac\") " pod="openstack/dnsmasq-dns-787484f569-fhzcg" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.563530 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7d4f9f19-62e5-4bcd-957e-345c18380eac-dns-swift-storage-0\") pod \"dnsmasq-dns-787484f569-fhzcg\" (UID: \"7d4f9f19-62e5-4bcd-957e-345c18380eac\") " pod="openstack/dnsmasq-dns-787484f569-fhzcg" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.563558 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5810478f-8400-4b6d-b647-e7294aa087ba-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5810478f-8400-4b6d-b647-e7294aa087ba\") " pod="openstack/glance-default-external-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.563613 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5810478f-8400-4b6d-b647-e7294aa087ba-config-data\") pod \"glance-default-external-api-0\" (UID: \"5810478f-8400-4b6d-b647-e7294aa087ba\") " pod="openstack/glance-default-external-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.563647 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5810478f-8400-4b6d-b647-e7294aa087ba-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5810478f-8400-4b6d-b647-e7294aa087ba\") " pod="openstack/glance-default-external-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.563698 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"5810478f-8400-4b6d-b647-e7294aa087ba\") " pod="openstack/glance-default-external-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.563741 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mllb4\" (UniqueName: \"kubernetes.io/projected/7d4f9f19-62e5-4bcd-957e-345c18380eac-kube-api-access-mllb4\") pod \"dnsmasq-dns-787484f569-fhzcg\" (UID: \"7d4f9f19-62e5-4bcd-957e-345c18380eac\") " pod="openstack/dnsmasq-dns-787484f569-fhzcg" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.565144 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d4f9f19-62e5-4bcd-957e-345c18380eac-ovsdbserver-nb\") pod \"dnsmasq-dns-787484f569-fhzcg\" (UID: \"7d4f9f19-62e5-4bcd-957e-345c18380eac\") " pod="openstack/dnsmasq-dns-787484f569-fhzcg" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.572773 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d4f9f19-62e5-4bcd-957e-345c18380eac-dns-svc\") pod \"dnsmasq-dns-787484f569-fhzcg\" (UID: \"7d4f9f19-62e5-4bcd-957e-345c18380eac\") " pod="openstack/dnsmasq-dns-787484f569-fhzcg" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.574558 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7d4f9f19-62e5-4bcd-957e-345c18380eac-ovsdbserver-sb\") pod \"dnsmasq-dns-787484f569-fhzcg\" (UID: \"7d4f9f19-62e5-4bcd-957e-345c18380eac\") " pod="openstack/dnsmasq-dns-787484f569-fhzcg" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.576556 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7d4f9f19-62e5-4bcd-957e-345c18380eac-dns-swift-storage-0\") pod \"dnsmasq-dns-787484f569-fhzcg\" (UID: \"7d4f9f19-62e5-4bcd-957e-345c18380eac\") " pod="openstack/dnsmasq-dns-787484f569-fhzcg" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.577903 4700 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"5810478f-8400-4b6d-b647-e7294aa087ba\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.578210 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d4f9f19-62e5-4bcd-957e-345c18380eac-config\") pod \"dnsmasq-dns-787484f569-fhzcg\" (UID: \"7d4f9f19-62e5-4bcd-957e-345c18380eac\") " pod="openstack/dnsmasq-dns-787484f569-fhzcg" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.578259 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5810478f-8400-4b6d-b647-e7294aa087ba-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5810478f-8400-4b6d-b647-e7294aa087ba\") " pod="openstack/glance-default-external-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.578506 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5810478f-8400-4b6d-b647-e7294aa087ba-logs\") pod \"glance-default-external-api-0\" (UID: \"5810478f-8400-4b6d-b647-e7294aa087ba\") " pod="openstack/glance-default-external-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.585664 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5810478f-8400-4b6d-b647-e7294aa087ba-config-data\") pod \"glance-default-external-api-0\" (UID: \"5810478f-8400-4b6d-b647-e7294aa087ba\") " pod="openstack/glance-default-external-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.586146 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5810478f-8400-4b6d-b647-e7294aa087ba-scripts\") pod \"glance-default-external-api-0\" (UID: \"5810478f-8400-4b6d-b647-e7294aa087ba\") " pod="openstack/glance-default-external-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.593173 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5810478f-8400-4b6d-b647-e7294aa087ba-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5810478f-8400-4b6d-b647-e7294aa087ba\") " pod="openstack/glance-default-external-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.600530 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5810478f-8400-4b6d-b647-e7294aa087ba-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5810478f-8400-4b6d-b647-e7294aa087ba\") " pod="openstack/glance-default-external-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.604104 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5rmn\" (UniqueName: \"kubernetes.io/projected/5810478f-8400-4b6d-b647-e7294aa087ba-kube-api-access-q5rmn\") pod \"glance-default-external-api-0\" (UID: \"5810478f-8400-4b6d-b647-e7294aa087ba\") " pod="openstack/glance-default-external-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.604595 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mllb4\" (UniqueName: \"kubernetes.io/projected/7d4f9f19-62e5-4bcd-957e-345c18380eac-kube-api-access-mllb4\") pod \"dnsmasq-dns-787484f569-fhzcg\" (UID: \"7d4f9f19-62e5-4bcd-957e-345c18380eac\") " pod="openstack/dnsmasq-dns-787484f569-fhzcg" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.629410 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"5810478f-8400-4b6d-b647-e7294aa087ba\") " pod="openstack/glance-default-external-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.690506 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-vrb9h" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.713845 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dmf4m" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.742421 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.744679 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.746990 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.747649 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.751339 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d7b5ff8c7-zbshw" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.784199 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.784749 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.842158 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-787484f569-fhzcg" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.888521 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"5aae014e-6d34-47fb-9af7-3b1901025e51\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.888666 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aae014e-6d34-47fb-9af7-3b1901025e51-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5aae014e-6d34-47fb-9af7-3b1901025e51\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.888746 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5aae014e-6d34-47fb-9af7-3b1901025e51-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5aae014e-6d34-47fb-9af7-3b1901025e51\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.888776 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5aae014e-6d34-47fb-9af7-3b1901025e51-logs\") pod \"glance-default-internal-api-0\" (UID: \"5aae014e-6d34-47fb-9af7-3b1901025e51\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.888818 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzfhv\" (UniqueName: \"kubernetes.io/projected/5aae014e-6d34-47fb-9af7-3b1901025e51-kube-api-access-bzfhv\") pod \"glance-default-internal-api-0\" (UID: \"5aae014e-6d34-47fb-9af7-3b1901025e51\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.888881 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5aae014e-6d34-47fb-9af7-3b1901025e51-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5aae014e-6d34-47fb-9af7-3b1901025e51\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.888928 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5aae014e-6d34-47fb-9af7-3b1901025e51-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5aae014e-6d34-47fb-9af7-3b1901025e51\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.888958 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5aae014e-6d34-47fb-9af7-3b1901025e51-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5aae014e-6d34-47fb-9af7-3b1901025e51\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.991445 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5aae014e-6d34-47fb-9af7-3b1901025e51-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5aae014e-6d34-47fb-9af7-3b1901025e51\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.992159 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5aae014e-6d34-47fb-9af7-3b1901025e51-logs\") pod \"glance-default-internal-api-0\" (UID: \"5aae014e-6d34-47fb-9af7-3b1901025e51\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.992255 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzfhv\" (UniqueName: \"kubernetes.io/projected/5aae014e-6d34-47fb-9af7-3b1901025e51-kube-api-access-bzfhv\") pod \"glance-default-internal-api-0\" (UID: \"5aae014e-6d34-47fb-9af7-3b1901025e51\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.992339 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5aae014e-6d34-47fb-9af7-3b1901025e51-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5aae014e-6d34-47fb-9af7-3b1901025e51\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.992433 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5aae014e-6d34-47fb-9af7-3b1901025e51-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5aae014e-6d34-47fb-9af7-3b1901025e51\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.992522 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5aae014e-6d34-47fb-9af7-3b1901025e51-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5aae014e-6d34-47fb-9af7-3b1901025e51\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.992603 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"5aae014e-6d34-47fb-9af7-3b1901025e51\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.992701 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aae014e-6d34-47fb-9af7-3b1901025e51-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5aae014e-6d34-47fb-9af7-3b1901025e51\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.993690 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5aae014e-6d34-47fb-9af7-3b1901025e51-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5aae014e-6d34-47fb-9af7-3b1901025e51\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.993982 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5aae014e-6d34-47fb-9af7-3b1901025e51-logs\") pod \"glance-default-internal-api-0\" (UID: \"5aae014e-6d34-47fb-9af7-3b1901025e51\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:21:29 crc kubenswrapper[4700]: I0227 17:21:29.994947 4700 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"5aae014e-6d34-47fb-9af7-3b1901025e51\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Feb 27 17:21:30 crc kubenswrapper[4700]: I0227 17:21:30.009141 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5aae014e-6d34-47fb-9af7-3b1901025e51-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5aae014e-6d34-47fb-9af7-3b1901025e51\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:21:30 crc kubenswrapper[4700]: I0227 17:21:30.011426 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5aae014e-6d34-47fb-9af7-3b1901025e51-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5aae014e-6d34-47fb-9af7-3b1901025e51\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:21:30 crc kubenswrapper[4700]: I0227 17:21:30.011832 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzfhv\" (UniqueName: \"kubernetes.io/projected/5aae014e-6d34-47fb-9af7-3b1901025e51-kube-api-access-bzfhv\") pod \"glance-default-internal-api-0\" (UID: \"5aae014e-6d34-47fb-9af7-3b1901025e51\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:21:30 crc kubenswrapper[4700]: I0227 17:21:30.012938 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5aae014e-6d34-47fb-9af7-3b1901025e51-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5aae014e-6d34-47fb-9af7-3b1901025e51\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:21:30 crc kubenswrapper[4700]: I0227 17:21:30.014503 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aae014e-6d34-47fb-9af7-3b1901025e51-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5aae014e-6d34-47fb-9af7-3b1901025e51\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:21:30 crc kubenswrapper[4700]: I0227 17:21:30.041243 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"5aae014e-6d34-47fb-9af7-3b1901025e51\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:21:30 crc kubenswrapper[4700]: W0227 17:21:30.132686 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcfded842_e9db_4883_8501_76c145505e47.slice/crio-122a0de0d9795b5873701c2b72ec5b204ba9b5c2482b76825bf36637e76b68bd WatchSource:0}: Error finding container 122a0de0d9795b5873701c2b72ec5b204ba9b5c2482b76825bf36637e76b68bd: Status 404 returned error can't find the container with id 122a0de0d9795b5873701c2b72ec5b204ba9b5c2482b76825bf36637e76b68bd Feb 27 17:21:30 crc kubenswrapper[4700]: I0227 17:21:30.133216 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7687748657-9lv22"] Feb 27 17:21:30 crc kubenswrapper[4700]: I0227 17:21:30.147323 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-9hq27"] Feb 27 17:21:30 crc kubenswrapper[4700]: I0227 17:21:30.244751 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9hq27" event={"ID":"bb266c49-545c-4e58-a9f8-6dede53accdf","Type":"ContainerStarted","Data":"b279799e1dae229de7d1b0cbba974b8e92040a99b241f0836a1b27c9a37daf90"} Feb 27 17:21:30 crc kubenswrapper[4700]: I0227 17:21:30.246079 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7687748657-9lv22" event={"ID":"cfded842-e9db-4883-8501-76c145505e47","Type":"ContainerStarted","Data":"122a0de0d9795b5873701c2b72ec5b204ba9b5c2482b76825bf36637e76b68bd"} Feb 27 17:21:30 crc kubenswrapper[4700]: I0227 17:21:30.248632 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 27 17:21:30 crc kubenswrapper[4700]: I0227 17:21:30.463640 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-mhcf5"] Feb 27 17:21:30 crc kubenswrapper[4700]: I0227 17:21:30.477345 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:21:30 crc kubenswrapper[4700]: I0227 17:21:30.649536 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Feb 27 17:21:30 crc kubenswrapper[4700]: I0227 17:21:30.661227 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Feb 27 17:21:30 crc kubenswrapper[4700]: I0227 17:21:30.672573 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6fbcf8869-6nmv8"] Feb 27 17:21:30 crc kubenswrapper[4700]: I0227 17:21:30.679491 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Feb 27 17:21:30 crc kubenswrapper[4700]: W0227 17:21:30.691053 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e0da06c_4410_4fe1_bc2e_1c08e25a1a7e.slice/crio-8335e89a460f97fa470a541a43c201d64acc7a6bf00853fec92ad673d370cb50 WatchSource:0}: Error finding container 8335e89a460f97fa470a541a43c201d64acc7a6bf00853fec92ad673d370cb50: Status 404 returned error can't find the container with id 8335e89a460f97fa470a541a43c201d64acc7a6bf00853fec92ad673d370cb50 Feb 27 17:21:30 crc kubenswrapper[4700]: W0227 17:21:30.700584 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08c973c7_192d_4326_9630_aa18aea6a484.slice/crio-b08e4a607ca6eeb709af3df47cd5219301ea2dba85282c6bd9478f4ed577e16f WatchSource:0}: Error finding container b08e4a607ca6eeb709af3df47cd5219301ea2dba85282c6bd9478f4ed577e16f: Status 404 returned error can't find the container with id b08e4a607ca6eeb709af3df47cd5219301ea2dba85282c6bd9478f4ed577e16f Feb 27 17:21:30 crc kubenswrapper[4700]: I0227 17:21:30.727515 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-d7b5ff8c7-zbshw"] Feb 27 17:21:30 crc kubenswrapper[4700]: I0227 17:21:30.742415 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-vrb9h"] Feb 27 17:21:30 crc kubenswrapper[4700]: I0227 17:21:30.750567 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-6b5ck"] Feb 27 17:21:30 crc kubenswrapper[4700]: W0227 17:21:30.761248 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf72833f0_cf5f_4bf8_bb5f_1f65477c4b76.slice/crio-6c6f06c549b0df8963eef790078a533d0845705b5c6a418be97923315f7fd804 WatchSource:0}: Error finding container 6c6f06c549b0df8963eef790078a533d0845705b5c6a418be97923315f7fd804: Status 404 returned error can't find the container with id 6c6f06c549b0df8963eef790078a533d0845705b5c6a418be97923315f7fd804 Feb 27 17:21:30 crc kubenswrapper[4700]: I0227 17:21:30.890902 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-dmf4m"] Feb 27 17:21:30 crc kubenswrapper[4700]: W0227 17:21:30.916828 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod989eacab_9c48_4375_b7a0_f2c7f40d17ca.slice/crio-fffa2885ffb0a6f235b8106c289c562b951a3d3d8c451756699d73c0ea0b6cdb WatchSource:0}: Error finding container fffa2885ffb0a6f235b8106c289c562b951a3d3d8c451756699d73c0ea0b6cdb: Status 404 returned error can't find the container with id fffa2885ffb0a6f235b8106c289c562b951a3d3d8c451756699d73c0ea0b6cdb Feb 27 17:21:30 crc kubenswrapper[4700]: I0227 17:21:30.928490 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-787484f569-fhzcg"] Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.030086 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.233550 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.282631 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.302018 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6fbcf8869-6nmv8"] Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.355312 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.366086 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.383803 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"08c973c7-192d-4326-9630-aa18aea6a484","Type":"ContainerStarted","Data":"c9fac31959703f02774d5d27d913f34715a9cc958dba2d9feb8cabb37e8fa868"} Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.384298 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"08c973c7-192d-4326-9630-aa18aea6a484","Type":"ContainerStarted","Data":"b08e4a607ca6eeb709af3df47cd5219301ea2dba85282c6bd9478f4ed577e16f"} Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.387060 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5bc46db9f7-zczkg"] Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.389274 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bc46db9f7-zczkg" Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.395207 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-vrb9h" event={"ID":"f72833f0-cf5f-4bf8-bb5f-1f65477c4b76","Type":"ContainerStarted","Data":"6c6f06c549b0df8963eef790078a533d0845705b5c6a418be97923315f7fd804"} Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.406255 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5bc46db9f7-zczkg"] Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.406304 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dmf4m" event={"ID":"989eacab-9c48-4375-b7a0-f2c7f40d17ca","Type":"ContainerStarted","Data":"fffa2885ffb0a6f235b8106c289c562b951a3d3d8c451756699d73c0ea0b6cdb"} Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.423278 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5810478f-8400-4b6d-b647-e7294aa087ba","Type":"ContainerStarted","Data":"7349ff6f87ac9b272915a050d3c4968356affd17eed934469ccf251f668fb25d"} Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.433510 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.438519 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"4099c554-cf42-48b3-88cb-e672a66f1a05","Type":"ContainerStarted","Data":"133172d8e994f23225b6e8b5515cb0f7bbadf05464ecf9b4304224beae2992bd"} Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.448398 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-787484f569-fhzcg" event={"ID":"7d4f9f19-62e5-4bcd-957e-345c18380eac","Type":"ContainerStarted","Data":"b56d1bcb9f55a4eaa8ccc45b893ff7095168b42465be0664087f40633102c8db"} Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.460115 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9hq27" event={"ID":"bb266c49-545c-4e58-a9f8-6dede53accdf","Type":"ContainerStarted","Data":"9d2bf4686af7c8e42b169470fe39fac12f905746ec23838a4b6615396a50674a"} Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.470156 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"d8e03836-cf5d-4ac9-80fe-97556117e65e","Type":"ContainerStarted","Data":"5e543ce39643617a2d1093e251750a3f81e9cf2abd59ae1d43c1c647ea768d2a"} Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.480893 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-6b5ck" event={"ID":"3f209c04-79ea-4048-9d31-6222874de091","Type":"ContainerStarted","Data":"e717d818bfbb46c1a1c7498bbf3e72872eae0252d5af74b4ba64523538038f72"} Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.484430 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6fbcf8869-6nmv8" event={"ID":"9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e","Type":"ContainerStarted","Data":"8335e89a460f97fa470a541a43c201d64acc7a6bf00853fec92ad673d370cb50"} Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.523973 4700 generic.go:334] "Generic (PLEG): container finished" podID="cfded842-e9db-4883-8501-76c145505e47" containerID="fe51e1cbb9718ac7a93764c13b6f356b7118f22ee59970f15cc979e91ec69bb8" exitCode=0 Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.524044 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7687748657-9lv22" event={"ID":"cfded842-e9db-4883-8501-76c145505e47","Type":"ContainerDied","Data":"fe51e1cbb9718ac7a93764c13b6f356b7118f22ee59970f15cc979e91ec69bb8"} Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.536551 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-9hq27" podStartSLOduration=3.53653308 podStartE2EDuration="3.53653308s" podCreationTimestamp="2026-02-27 17:21:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:21:31.512809392 +0000 UTC m=+1251.498122139" watchObservedRunningTime="2026-02-27 17:21:31.53653308 +0000 UTC m=+1251.521845817" Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.542983 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/aaf44cea-4a4a-4103-a7bd-a7c7a6d97342-config-data\") pod \"horizon-5bc46db9f7-zczkg\" (UID: \"aaf44cea-4a4a-4103-a7bd-a7c7a6d97342\") " pod="openstack/horizon-5bc46db9f7-zczkg" Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.543065 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjgr4\" (UniqueName: \"kubernetes.io/projected/aaf44cea-4a4a-4103-a7bd-a7c7a6d97342-kube-api-access-fjgr4\") pod \"horizon-5bc46db9f7-zczkg\" (UID: \"aaf44cea-4a4a-4103-a7bd-a7c7a6d97342\") " pod="openstack/horizon-5bc46db9f7-zczkg" Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.543091 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/aaf44cea-4a4a-4103-a7bd-a7c7a6d97342-horizon-secret-key\") pod \"horizon-5bc46db9f7-zczkg\" (UID: \"aaf44cea-4a4a-4103-a7bd-a7c7a6d97342\") " pod="openstack/horizon-5bc46db9f7-zczkg" Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.543107 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aaf44cea-4a4a-4103-a7bd-a7c7a6d97342-logs\") pod \"horizon-5bc46db9f7-zczkg\" (UID: \"aaf44cea-4a4a-4103-a7bd-a7c7a6d97342\") " pod="openstack/horizon-5bc46db9f7-zczkg" Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.543150 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aaf44cea-4a4a-4103-a7bd-a7c7a6d97342-scripts\") pod \"horizon-5bc46db9f7-zczkg\" (UID: \"aaf44cea-4a4a-4103-a7bd-a7c7a6d97342\") " pod="openstack/horizon-5bc46db9f7-zczkg" Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.544989 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d7b5ff8c7-zbshw" event={"ID":"94005464-5a50-42ba-bb5f-b69b66cdec58","Type":"ContainerStarted","Data":"fb75180431b1242f03100293dddb74fedf5d27f5d6c2c5a0b72f0c9fed00201a"} Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.547565 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-6b5ck" podStartSLOduration=3.547550341 podStartE2EDuration="3.547550341s" podCreationTimestamp="2026-02-27 17:21:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:21:31.529144004 +0000 UTC m=+1251.514456741" watchObservedRunningTime="2026-02-27 17:21:31.547550341 +0000 UTC m=+1251.532863088" Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.549694 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-mhcf5" event={"ID":"dd3817a7-c2ab-43f0-9c74-b477ac59fa52","Type":"ContainerStarted","Data":"123d2839568692618dc1e50207e5dd1af0d3f189f8dbd11fc2e0f38e7c894202"} Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.553475 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"61baccca-f7e6-4442-b612-8ab97b82036d","Type":"ContainerStarted","Data":"e039ff159043fe8c4a36f636e7fd3aede836587ad39537e9de0102908621e161"} Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.645239 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/aaf44cea-4a4a-4103-a7bd-a7c7a6d97342-config-data\") pod \"horizon-5bc46db9f7-zczkg\" (UID: \"aaf44cea-4a4a-4103-a7bd-a7c7a6d97342\") " pod="openstack/horizon-5bc46db9f7-zczkg" Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.645311 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjgr4\" (UniqueName: \"kubernetes.io/projected/aaf44cea-4a4a-4103-a7bd-a7c7a6d97342-kube-api-access-fjgr4\") pod \"horizon-5bc46db9f7-zczkg\" (UID: \"aaf44cea-4a4a-4103-a7bd-a7c7a6d97342\") " pod="openstack/horizon-5bc46db9f7-zczkg" Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.645335 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/aaf44cea-4a4a-4103-a7bd-a7c7a6d97342-horizon-secret-key\") pod \"horizon-5bc46db9f7-zczkg\" (UID: \"aaf44cea-4a4a-4103-a7bd-a7c7a6d97342\") " pod="openstack/horizon-5bc46db9f7-zczkg" Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.645353 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aaf44cea-4a4a-4103-a7bd-a7c7a6d97342-logs\") pod \"horizon-5bc46db9f7-zczkg\" (UID: \"aaf44cea-4a4a-4103-a7bd-a7c7a6d97342\") " pod="openstack/horizon-5bc46db9f7-zczkg" Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.645384 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aaf44cea-4a4a-4103-a7bd-a7c7a6d97342-scripts\") pod \"horizon-5bc46db9f7-zczkg\" (UID: \"aaf44cea-4a4a-4103-a7bd-a7c7a6d97342\") " pod="openstack/horizon-5bc46db9f7-zczkg" Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.646172 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aaf44cea-4a4a-4103-a7bd-a7c7a6d97342-scripts\") pod \"horizon-5bc46db9f7-zczkg\" (UID: \"aaf44cea-4a4a-4103-a7bd-a7c7a6d97342\") " pod="openstack/horizon-5bc46db9f7-zczkg" Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.646167 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aaf44cea-4a4a-4103-a7bd-a7c7a6d97342-logs\") pod \"horizon-5bc46db9f7-zczkg\" (UID: \"aaf44cea-4a4a-4103-a7bd-a7c7a6d97342\") " pod="openstack/horizon-5bc46db9f7-zczkg" Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.647411 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/aaf44cea-4a4a-4103-a7bd-a7c7a6d97342-config-data\") pod \"horizon-5bc46db9f7-zczkg\" (UID: \"aaf44cea-4a4a-4103-a7bd-a7c7a6d97342\") " pod="openstack/horizon-5bc46db9f7-zczkg" Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.667737 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/aaf44cea-4a4a-4103-a7bd-a7c7a6d97342-horizon-secret-key\") pod \"horizon-5bc46db9f7-zczkg\" (UID: \"aaf44cea-4a4a-4103-a7bd-a7c7a6d97342\") " pod="openstack/horizon-5bc46db9f7-zczkg" Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.668003 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjgr4\" (UniqueName: \"kubernetes.io/projected/aaf44cea-4a4a-4103-a7bd-a7c7a6d97342-kube-api-access-fjgr4\") pod \"horizon-5bc46db9f7-zczkg\" (UID: \"aaf44cea-4a4a-4103-a7bd-a7c7a6d97342\") " pod="openstack/horizon-5bc46db9f7-zczkg" Feb 27 17:21:31 crc kubenswrapper[4700]: I0227 17:21:31.760024 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bc46db9f7-zczkg" Feb 27 17:21:32 crc kubenswrapper[4700]: I0227 17:21:32.088319 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7687748657-9lv22" Feb 27 17:21:32 crc kubenswrapper[4700]: I0227 17:21:32.184319 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cfded842-e9db-4883-8501-76c145505e47-ovsdbserver-sb\") pod \"cfded842-e9db-4883-8501-76c145505e47\" (UID: \"cfded842-e9db-4883-8501-76c145505e47\") " Feb 27 17:21:32 crc kubenswrapper[4700]: I0227 17:21:32.184469 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfded842-e9db-4883-8501-76c145505e47-config\") pod \"cfded842-e9db-4883-8501-76c145505e47\" (UID: \"cfded842-e9db-4883-8501-76c145505e47\") " Feb 27 17:21:32 crc kubenswrapper[4700]: I0227 17:21:32.184494 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cfded842-e9db-4883-8501-76c145505e47-ovsdbserver-nb\") pod \"cfded842-e9db-4883-8501-76c145505e47\" (UID: \"cfded842-e9db-4883-8501-76c145505e47\") " Feb 27 17:21:32 crc kubenswrapper[4700]: I0227 17:21:32.184524 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cfded842-e9db-4883-8501-76c145505e47-dns-swift-storage-0\") pod \"cfded842-e9db-4883-8501-76c145505e47\" (UID: \"cfded842-e9db-4883-8501-76c145505e47\") " Feb 27 17:21:32 crc kubenswrapper[4700]: I0227 17:21:32.184647 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7nlws\" (UniqueName: \"kubernetes.io/projected/cfded842-e9db-4883-8501-76c145505e47-kube-api-access-7nlws\") pod \"cfded842-e9db-4883-8501-76c145505e47\" (UID: \"cfded842-e9db-4883-8501-76c145505e47\") " Feb 27 17:21:32 crc kubenswrapper[4700]: I0227 17:21:32.184671 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cfded842-e9db-4883-8501-76c145505e47-dns-svc\") pod \"cfded842-e9db-4883-8501-76c145505e47\" (UID: \"cfded842-e9db-4883-8501-76c145505e47\") " Feb 27 17:21:32 crc kubenswrapper[4700]: I0227 17:21:32.212431 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfded842-e9db-4883-8501-76c145505e47-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "cfded842-e9db-4883-8501-76c145505e47" (UID: "cfded842-e9db-4883-8501-76c145505e47"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:21:32 crc kubenswrapper[4700]: I0227 17:21:32.218726 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfded842-e9db-4883-8501-76c145505e47-kube-api-access-7nlws" (OuterVolumeSpecName: "kube-api-access-7nlws") pod "cfded842-e9db-4883-8501-76c145505e47" (UID: "cfded842-e9db-4883-8501-76c145505e47"). InnerVolumeSpecName "kube-api-access-7nlws". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:21:32 crc kubenswrapper[4700]: I0227 17:21:32.233095 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfded842-e9db-4883-8501-76c145505e47-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cfded842-e9db-4883-8501-76c145505e47" (UID: "cfded842-e9db-4883-8501-76c145505e47"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:21:32 crc kubenswrapper[4700]: I0227 17:21:32.280118 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfded842-e9db-4883-8501-76c145505e47-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "cfded842-e9db-4883-8501-76c145505e47" (UID: "cfded842-e9db-4883-8501-76c145505e47"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:21:32 crc kubenswrapper[4700]: I0227 17:21:32.289578 4700 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cfded842-e9db-4883-8501-76c145505e47-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:32 crc kubenswrapper[4700]: I0227 17:21:32.289608 4700 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cfded842-e9db-4883-8501-76c145505e47-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:32 crc kubenswrapper[4700]: I0227 17:21:32.289618 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7nlws\" (UniqueName: \"kubernetes.io/projected/cfded842-e9db-4883-8501-76c145505e47-kube-api-access-7nlws\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:32 crc kubenswrapper[4700]: I0227 17:21:32.289627 4700 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cfded842-e9db-4883-8501-76c145505e47-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:32 crc kubenswrapper[4700]: I0227 17:21:32.303027 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfded842-e9db-4883-8501-76c145505e47-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cfded842-e9db-4883-8501-76c145505e47" (UID: "cfded842-e9db-4883-8501-76c145505e47"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:21:32 crc kubenswrapper[4700]: I0227 17:21:32.307041 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfded842-e9db-4883-8501-76c145505e47-config" (OuterVolumeSpecName: "config") pod "cfded842-e9db-4883-8501-76c145505e47" (UID: "cfded842-e9db-4883-8501-76c145505e47"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:21:32 crc kubenswrapper[4700]: I0227 17:21:32.319170 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5bc46db9f7-zczkg"] Feb 27 17:21:32 crc kubenswrapper[4700]: W0227 17:21:32.366594 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaaf44cea_4a4a_4103_a7bd_a7c7a6d97342.slice/crio-6e77efac0053700f53620ec14cb309c5b7ab3a05d32f5660b466ab8225b88276 WatchSource:0}: Error finding container 6e77efac0053700f53620ec14cb309c5b7ab3a05d32f5660b466ab8225b88276: Status 404 returned error can't find the container with id 6e77efac0053700f53620ec14cb309c5b7ab3a05d32f5660b466ab8225b88276 Feb 27 17:21:32 crc kubenswrapper[4700]: I0227 17:21:32.392626 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfded842-e9db-4883-8501-76c145505e47-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:32 crc kubenswrapper[4700]: I0227 17:21:32.392669 4700 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cfded842-e9db-4883-8501-76c145505e47-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:32 crc kubenswrapper[4700]: I0227 17:21:32.574257 4700 generic.go:334] "Generic (PLEG): container finished" podID="7d4f9f19-62e5-4bcd-957e-345c18380eac" containerID="93575a19a3234cea9226b813f67f83075188b0e9f1af0c913dc2b560d7450b9a" exitCode=0 Feb 27 17:21:32 crc kubenswrapper[4700]: I0227 17:21:32.576852 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-787484f569-fhzcg" event={"ID":"7d4f9f19-62e5-4bcd-957e-345c18380eac","Type":"ContainerDied","Data":"93575a19a3234cea9226b813f67f83075188b0e9f1af0c913dc2b560d7450b9a"} Feb 27 17:21:32 crc kubenswrapper[4700]: I0227 17:21:32.578913 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7687748657-9lv22" Feb 27 17:21:32 crc kubenswrapper[4700]: I0227 17:21:32.579749 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7687748657-9lv22" event={"ID":"cfded842-e9db-4883-8501-76c145505e47","Type":"ContainerDied","Data":"122a0de0d9795b5873701c2b72ec5b204ba9b5c2482b76825bf36637e76b68bd"} Feb 27 17:21:32 crc kubenswrapper[4700]: I0227 17:21:32.579782 4700 scope.go:117] "RemoveContainer" containerID="fe51e1cbb9718ac7a93764c13b6f356b7118f22ee59970f15cc979e91ec69bb8" Feb 27 17:21:32 crc kubenswrapper[4700]: I0227 17:21:32.600640 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bc46db9f7-zczkg" event={"ID":"aaf44cea-4a4a-4103-a7bd-a7c7a6d97342","Type":"ContainerStarted","Data":"6e77efac0053700f53620ec14cb309c5b7ab3a05d32f5660b466ab8225b88276"} Feb 27 17:21:32 crc kubenswrapper[4700]: I0227 17:21:32.628343 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"08c973c7-192d-4326-9630-aa18aea6a484","Type":"ContainerStarted","Data":"0c16b57672e1a5f873e00382c5b2e11ad43caaad41b66d7f5c14965ad07952c4"} Feb 27 17:21:32 crc kubenswrapper[4700]: I0227 17:21:32.628648 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Feb 27 17:21:32 crc kubenswrapper[4700]: I0227 17:21:32.628536 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="08c973c7-192d-4326-9630-aa18aea6a484" containerName="watcher-api-log" containerID="cri-o://c9fac31959703f02774d5d27d913f34715a9cc958dba2d9feb8cabb37e8fa868" gracePeriod=30 Feb 27 17:21:32 crc kubenswrapper[4700]: I0227 17:21:32.628800 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="08c973c7-192d-4326-9630-aa18aea6a484" containerName="watcher-api" containerID="cri-o://0c16b57672e1a5f873e00382c5b2e11ad43caaad41b66d7f5c14965ad07952c4" gracePeriod=30 Feb 27 17:21:32 crc kubenswrapper[4700]: I0227 17:21:32.651393 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5aae014e-6d34-47fb-9af7-3b1901025e51","Type":"ContainerStarted","Data":"6ae74025b872ed814351ea8c739b32a31304a07fdc82802f42b24034bdccc514"} Feb 27 17:21:32 crc kubenswrapper[4700]: I0227 17:21:32.659996 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-6b5ck" event={"ID":"3f209c04-79ea-4048-9d31-6222874de091","Type":"ContainerStarted","Data":"1995dc0741ace72e33bf0776b3003b5630b6888aacbfc518cefa2974898a9358"} Feb 27 17:21:32 crc kubenswrapper[4700]: I0227 17:21:32.662037 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-api-0" podStartSLOduration=4.662014751 podStartE2EDuration="4.662014751s" podCreationTimestamp="2026-02-27 17:21:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:21:32.650506607 +0000 UTC m=+1252.635819354" watchObservedRunningTime="2026-02-27 17:21:32.662014751 +0000 UTC m=+1252.647327498" Feb 27 17:21:32 crc kubenswrapper[4700]: I0227 17:21:32.681771 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="08c973c7-192d-4326-9630-aa18aea6a484" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.155:9322/\": EOF" Feb 27 17:21:32 crc kubenswrapper[4700]: I0227 17:21:32.870045 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7687748657-9lv22"] Feb 27 17:21:32 crc kubenswrapper[4700]: I0227 17:21:32.881000 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7687748657-9lv22"] Feb 27 17:21:33 crc kubenswrapper[4700]: I0227 17:21:33.007137 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfded842-e9db-4883-8501-76c145505e47" path="/var/lib/kubelet/pods/cfded842-e9db-4883-8501-76c145505e47/volumes" Feb 27 17:21:33 crc kubenswrapper[4700]: I0227 17:21:33.689172 4700 generic.go:334] "Generic (PLEG): container finished" podID="08c973c7-192d-4326-9630-aa18aea6a484" containerID="c9fac31959703f02774d5d27d913f34715a9cc958dba2d9feb8cabb37e8fa868" exitCode=143 Feb 27 17:21:33 crc kubenswrapper[4700]: I0227 17:21:33.689275 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"08c973c7-192d-4326-9630-aa18aea6a484","Type":"ContainerDied","Data":"c9fac31959703f02774d5d27d913f34715a9cc958dba2d9feb8cabb37e8fa868"} Feb 27 17:21:33 crc kubenswrapper[4700]: I0227 17:21:33.703557 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5aae014e-6d34-47fb-9af7-3b1901025e51","Type":"ContainerStarted","Data":"481cd96f6890e8291a0d0587ef75e75c787464a33a4382b61f8a8f3c9dd0148f"} Feb 27 17:21:33 crc kubenswrapper[4700]: I0227 17:21:33.707065 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5810478f-8400-4b6d-b647-e7294aa087ba","Type":"ContainerStarted","Data":"9560399dcda8d30788a5744b22b16ffe16678cbfd42a8ef1cf0f18a207bafcdf"} Feb 27 17:21:33 crc kubenswrapper[4700]: I0227 17:21:33.710187 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-787484f569-fhzcg" event={"ID":"7d4f9f19-62e5-4bcd-957e-345c18380eac","Type":"ContainerStarted","Data":"7c8c08b9bd26fb61dd0d15098608d526d6a4041994ea5684b4d03ddc15abe370"} Feb 27 17:21:33 crc kubenswrapper[4700]: I0227 17:21:33.734874 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-787484f569-fhzcg" podStartSLOduration=5.734855659 podStartE2EDuration="5.734855659s" podCreationTimestamp="2026-02-27 17:21:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:21:33.731152411 +0000 UTC m=+1253.716465158" watchObservedRunningTime="2026-02-27 17:21:33.734855659 +0000 UTC m=+1253.720168406" Feb 27 17:21:33 crc kubenswrapper[4700]: I0227 17:21:33.893843 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Feb 27 17:21:34 crc kubenswrapper[4700]: I0227 17:21:34.718850 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-787484f569-fhzcg" Feb 27 17:21:36 crc kubenswrapper[4700]: I0227 17:21:36.410131 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:21:36 crc kubenswrapper[4700]: I0227 17:21:36.410501 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:21:36 crc kubenswrapper[4700]: I0227 17:21:36.410552 4700 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" Feb 27 17:21:36 crc kubenswrapper[4700]: I0227 17:21:36.411324 4700 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c99a71cb46c45356ba1bf62a0adf7979247710d0e51e5eac769c9c248f05463d"} pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 17:21:36 crc kubenswrapper[4700]: I0227 17:21:36.411371 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" containerID="cri-o://c99a71cb46c45356ba1bf62a0adf7979247710d0e51e5eac769c9c248f05463d" gracePeriod=600 Feb 27 17:21:36 crc kubenswrapper[4700]: I0227 17:21:36.426616 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="08c973c7-192d-4326-9630-aa18aea6a484" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.155:9322/\": read tcp 10.217.0.2:42616->10.217.0.155:9322: read: connection reset by peer" Feb 27 17:21:36 crc kubenswrapper[4700]: I0227 17:21:36.762761 4700 generic.go:334] "Generic (PLEG): container finished" podID="08c973c7-192d-4326-9630-aa18aea6a484" containerID="0c16b57672e1a5f873e00382c5b2e11ad43caaad41b66d7f5c14965ad07952c4" exitCode=0 Feb 27 17:21:36 crc kubenswrapper[4700]: I0227 17:21:36.762836 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"08c973c7-192d-4326-9630-aa18aea6a484","Type":"ContainerDied","Data":"0c16b57672e1a5f873e00382c5b2e11ad43caaad41b66d7f5c14965ad07952c4"} Feb 27 17:21:36 crc kubenswrapper[4700]: I0227 17:21:36.766869 4700 generic.go:334] "Generic (PLEG): container finished" podID="bb266c49-545c-4e58-a9f8-6dede53accdf" containerID="9d2bf4686af7c8e42b169470fe39fac12f905746ec23838a4b6615396a50674a" exitCode=0 Feb 27 17:21:36 crc kubenswrapper[4700]: I0227 17:21:36.766945 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9hq27" event={"ID":"bb266c49-545c-4e58-a9f8-6dede53accdf","Type":"ContainerDied","Data":"9d2bf4686af7c8e42b169470fe39fac12f905746ec23838a4b6615396a50674a"} Feb 27 17:21:36 crc kubenswrapper[4700]: I0227 17:21:36.770327 4700 generic.go:334] "Generic (PLEG): container finished" podID="84b45511-b94c-479f-98db-fd2c4eceec46" containerID="c99a71cb46c45356ba1bf62a0adf7979247710d0e51e5eac769c9c248f05463d" exitCode=0 Feb 27 17:21:36 crc kubenswrapper[4700]: I0227 17:21:36.770365 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerDied","Data":"c99a71cb46c45356ba1bf62a0adf7979247710d0e51e5eac769c9c248f05463d"} Feb 27 17:21:36 crc kubenswrapper[4700]: I0227 17:21:36.770398 4700 scope.go:117] "RemoveContainer" containerID="6c88e3b04f19e6791ce8b218282ad3546f21efb28cfed9c5676d0da60d8b308b" Feb 27 17:21:37 crc kubenswrapper[4700]: I0227 17:21:37.878907 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-d7b5ff8c7-zbshw"] Feb 27 17:21:37 crc kubenswrapper[4700]: I0227 17:21:37.932437 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5b48777cdd-tqbkj"] Feb 27 17:21:37 crc kubenswrapper[4700]: E0227 17:21:37.932838 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfded842-e9db-4883-8501-76c145505e47" containerName="init" Feb 27 17:21:37 crc kubenswrapper[4700]: I0227 17:21:37.932851 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfded842-e9db-4883-8501-76c145505e47" containerName="init" Feb 27 17:21:37 crc kubenswrapper[4700]: I0227 17:21:37.933024 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfded842-e9db-4883-8501-76c145505e47" containerName="init" Feb 27 17:21:37 crc kubenswrapper[4700]: I0227 17:21:37.935138 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5b48777cdd-tqbkj" Feb 27 17:21:37 crc kubenswrapper[4700]: I0227 17:21:37.937384 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Feb 27 17:21:37 crc kubenswrapper[4700]: I0227 17:21:37.952615 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5b48777cdd-tqbkj"] Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.042374 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3989a3e-45ff-45d8-80a2-4dbe57663a57-logs\") pod \"horizon-5b48777cdd-tqbkj\" (UID: \"d3989a3e-45ff-45d8-80a2-4dbe57663a57\") " pod="openstack/horizon-5b48777cdd-tqbkj" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.042437 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d3989a3e-45ff-45d8-80a2-4dbe57663a57-scripts\") pod \"horizon-5b48777cdd-tqbkj\" (UID: \"d3989a3e-45ff-45d8-80a2-4dbe57663a57\") " pod="openstack/horizon-5b48777cdd-tqbkj" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.042477 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3989a3e-45ff-45d8-80a2-4dbe57663a57-horizon-tls-certs\") pod \"horizon-5b48777cdd-tqbkj\" (UID: \"d3989a3e-45ff-45d8-80a2-4dbe57663a57\") " pod="openstack/horizon-5b48777cdd-tqbkj" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.042548 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d3989a3e-45ff-45d8-80a2-4dbe57663a57-horizon-secret-key\") pod \"horizon-5b48777cdd-tqbkj\" (UID: \"d3989a3e-45ff-45d8-80a2-4dbe57663a57\") " pod="openstack/horizon-5b48777cdd-tqbkj" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.042574 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d3989a3e-45ff-45d8-80a2-4dbe57663a57-config-data\") pod \"horizon-5b48777cdd-tqbkj\" (UID: \"d3989a3e-45ff-45d8-80a2-4dbe57663a57\") " pod="openstack/horizon-5b48777cdd-tqbkj" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.042590 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8v9ht\" (UniqueName: \"kubernetes.io/projected/d3989a3e-45ff-45d8-80a2-4dbe57663a57-kube-api-access-8v9ht\") pod \"horizon-5b48777cdd-tqbkj\" (UID: \"d3989a3e-45ff-45d8-80a2-4dbe57663a57\") " pod="openstack/horizon-5b48777cdd-tqbkj" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.042629 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3989a3e-45ff-45d8-80a2-4dbe57663a57-combined-ca-bundle\") pod \"horizon-5b48777cdd-tqbkj\" (UID: \"d3989a3e-45ff-45d8-80a2-4dbe57663a57\") " pod="openstack/horizon-5b48777cdd-tqbkj" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.042743 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5bc46db9f7-zczkg"] Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.073570 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6cc494ff8-ggq9w"] Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.075447 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6cc494ff8-ggq9w" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.093796 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6cc494ff8-ggq9w"] Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.145066 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3989a3e-45ff-45d8-80a2-4dbe57663a57-combined-ca-bundle\") pod \"horizon-5b48777cdd-tqbkj\" (UID: \"d3989a3e-45ff-45d8-80a2-4dbe57663a57\") " pod="openstack/horizon-5b48777cdd-tqbkj" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.145513 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3989a3e-45ff-45d8-80a2-4dbe57663a57-logs\") pod \"horizon-5b48777cdd-tqbkj\" (UID: \"d3989a3e-45ff-45d8-80a2-4dbe57663a57\") " pod="openstack/horizon-5b48777cdd-tqbkj" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.145544 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d3989a3e-45ff-45d8-80a2-4dbe57663a57-scripts\") pod \"horizon-5b48777cdd-tqbkj\" (UID: \"d3989a3e-45ff-45d8-80a2-4dbe57663a57\") " pod="openstack/horizon-5b48777cdd-tqbkj" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.145578 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3989a3e-45ff-45d8-80a2-4dbe57663a57-horizon-tls-certs\") pod \"horizon-5b48777cdd-tqbkj\" (UID: \"d3989a3e-45ff-45d8-80a2-4dbe57663a57\") " pod="openstack/horizon-5b48777cdd-tqbkj" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.145651 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d3989a3e-45ff-45d8-80a2-4dbe57663a57-horizon-secret-key\") pod \"horizon-5b48777cdd-tqbkj\" (UID: \"d3989a3e-45ff-45d8-80a2-4dbe57663a57\") " pod="openstack/horizon-5b48777cdd-tqbkj" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.145679 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8v9ht\" (UniqueName: \"kubernetes.io/projected/d3989a3e-45ff-45d8-80a2-4dbe57663a57-kube-api-access-8v9ht\") pod \"horizon-5b48777cdd-tqbkj\" (UID: \"d3989a3e-45ff-45d8-80a2-4dbe57663a57\") " pod="openstack/horizon-5b48777cdd-tqbkj" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.145699 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d3989a3e-45ff-45d8-80a2-4dbe57663a57-config-data\") pod \"horizon-5b48777cdd-tqbkj\" (UID: \"d3989a3e-45ff-45d8-80a2-4dbe57663a57\") " pod="openstack/horizon-5b48777cdd-tqbkj" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.147451 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3989a3e-45ff-45d8-80a2-4dbe57663a57-logs\") pod \"horizon-5b48777cdd-tqbkj\" (UID: \"d3989a3e-45ff-45d8-80a2-4dbe57663a57\") " pod="openstack/horizon-5b48777cdd-tqbkj" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.148092 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d3989a3e-45ff-45d8-80a2-4dbe57663a57-scripts\") pod \"horizon-5b48777cdd-tqbkj\" (UID: \"d3989a3e-45ff-45d8-80a2-4dbe57663a57\") " pod="openstack/horizon-5b48777cdd-tqbkj" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.153991 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d3989a3e-45ff-45d8-80a2-4dbe57663a57-config-data\") pod \"horizon-5b48777cdd-tqbkj\" (UID: \"d3989a3e-45ff-45d8-80a2-4dbe57663a57\") " pod="openstack/horizon-5b48777cdd-tqbkj" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.157412 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d3989a3e-45ff-45d8-80a2-4dbe57663a57-horizon-secret-key\") pod \"horizon-5b48777cdd-tqbkj\" (UID: \"d3989a3e-45ff-45d8-80a2-4dbe57663a57\") " pod="openstack/horizon-5b48777cdd-tqbkj" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.200925 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3989a3e-45ff-45d8-80a2-4dbe57663a57-combined-ca-bundle\") pod \"horizon-5b48777cdd-tqbkj\" (UID: \"d3989a3e-45ff-45d8-80a2-4dbe57663a57\") " pod="openstack/horizon-5b48777cdd-tqbkj" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.203987 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3989a3e-45ff-45d8-80a2-4dbe57663a57-horizon-tls-certs\") pod \"horizon-5b48777cdd-tqbkj\" (UID: \"d3989a3e-45ff-45d8-80a2-4dbe57663a57\") " pod="openstack/horizon-5b48777cdd-tqbkj" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.208948 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8v9ht\" (UniqueName: \"kubernetes.io/projected/d3989a3e-45ff-45d8-80a2-4dbe57663a57-kube-api-access-8v9ht\") pod \"horizon-5b48777cdd-tqbkj\" (UID: \"d3989a3e-45ff-45d8-80a2-4dbe57663a57\") " pod="openstack/horizon-5b48777cdd-tqbkj" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.247875 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ed7353a3-748a-40cb-8865-c4badeb8e402-scripts\") pod \"horizon-6cc494ff8-ggq9w\" (UID: \"ed7353a3-748a-40cb-8865-c4badeb8e402\") " pod="openstack/horizon-6cc494ff8-ggq9w" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.250611 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed7353a3-748a-40cb-8865-c4badeb8e402-horizon-tls-certs\") pod \"horizon-6cc494ff8-ggq9w\" (UID: \"ed7353a3-748a-40cb-8865-c4badeb8e402\") " pod="openstack/horizon-6cc494ff8-ggq9w" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.250812 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25qx6\" (UniqueName: \"kubernetes.io/projected/ed7353a3-748a-40cb-8865-c4badeb8e402-kube-api-access-25qx6\") pod \"horizon-6cc494ff8-ggq9w\" (UID: \"ed7353a3-748a-40cb-8865-c4badeb8e402\") " pod="openstack/horizon-6cc494ff8-ggq9w" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.250992 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed7353a3-748a-40cb-8865-c4badeb8e402-logs\") pod \"horizon-6cc494ff8-ggq9w\" (UID: \"ed7353a3-748a-40cb-8865-c4badeb8e402\") " pod="openstack/horizon-6cc494ff8-ggq9w" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.251196 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed7353a3-748a-40cb-8865-c4badeb8e402-combined-ca-bundle\") pod \"horizon-6cc494ff8-ggq9w\" (UID: \"ed7353a3-748a-40cb-8865-c4badeb8e402\") " pod="openstack/horizon-6cc494ff8-ggq9w" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.251227 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ed7353a3-748a-40cb-8865-c4badeb8e402-config-data\") pod \"horizon-6cc494ff8-ggq9w\" (UID: \"ed7353a3-748a-40cb-8865-c4badeb8e402\") " pod="openstack/horizon-6cc494ff8-ggq9w" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.251319 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ed7353a3-748a-40cb-8865-c4badeb8e402-horizon-secret-key\") pod \"horizon-6cc494ff8-ggq9w\" (UID: \"ed7353a3-748a-40cb-8865-c4badeb8e402\") " pod="openstack/horizon-6cc494ff8-ggq9w" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.268138 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5b48777cdd-tqbkj" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.353841 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ed7353a3-748a-40cb-8865-c4badeb8e402-scripts\") pod \"horizon-6cc494ff8-ggq9w\" (UID: \"ed7353a3-748a-40cb-8865-c4badeb8e402\") " pod="openstack/horizon-6cc494ff8-ggq9w" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.353906 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed7353a3-748a-40cb-8865-c4badeb8e402-horizon-tls-certs\") pod \"horizon-6cc494ff8-ggq9w\" (UID: \"ed7353a3-748a-40cb-8865-c4badeb8e402\") " pod="openstack/horizon-6cc494ff8-ggq9w" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.353984 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25qx6\" (UniqueName: \"kubernetes.io/projected/ed7353a3-748a-40cb-8865-c4badeb8e402-kube-api-access-25qx6\") pod \"horizon-6cc494ff8-ggq9w\" (UID: \"ed7353a3-748a-40cb-8865-c4badeb8e402\") " pod="openstack/horizon-6cc494ff8-ggq9w" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.354036 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed7353a3-748a-40cb-8865-c4badeb8e402-logs\") pod \"horizon-6cc494ff8-ggq9w\" (UID: \"ed7353a3-748a-40cb-8865-c4badeb8e402\") " pod="openstack/horizon-6cc494ff8-ggq9w" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.354103 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed7353a3-748a-40cb-8865-c4badeb8e402-combined-ca-bundle\") pod \"horizon-6cc494ff8-ggq9w\" (UID: \"ed7353a3-748a-40cb-8865-c4badeb8e402\") " pod="openstack/horizon-6cc494ff8-ggq9w" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.354126 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ed7353a3-748a-40cb-8865-c4badeb8e402-config-data\") pod \"horizon-6cc494ff8-ggq9w\" (UID: \"ed7353a3-748a-40cb-8865-c4badeb8e402\") " pod="openstack/horizon-6cc494ff8-ggq9w" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.354166 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ed7353a3-748a-40cb-8865-c4badeb8e402-horizon-secret-key\") pod \"horizon-6cc494ff8-ggq9w\" (UID: \"ed7353a3-748a-40cb-8865-c4badeb8e402\") " pod="openstack/horizon-6cc494ff8-ggq9w" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.354688 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ed7353a3-748a-40cb-8865-c4badeb8e402-scripts\") pod \"horizon-6cc494ff8-ggq9w\" (UID: \"ed7353a3-748a-40cb-8865-c4badeb8e402\") " pod="openstack/horizon-6cc494ff8-ggq9w" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.355560 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed7353a3-748a-40cb-8865-c4badeb8e402-logs\") pod \"horizon-6cc494ff8-ggq9w\" (UID: \"ed7353a3-748a-40cb-8865-c4badeb8e402\") " pod="openstack/horizon-6cc494ff8-ggq9w" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.356235 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ed7353a3-748a-40cb-8865-c4badeb8e402-config-data\") pod \"horizon-6cc494ff8-ggq9w\" (UID: \"ed7353a3-748a-40cb-8865-c4badeb8e402\") " pod="openstack/horizon-6cc494ff8-ggq9w" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.358179 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ed7353a3-748a-40cb-8865-c4badeb8e402-horizon-secret-key\") pod \"horizon-6cc494ff8-ggq9w\" (UID: \"ed7353a3-748a-40cb-8865-c4badeb8e402\") " pod="openstack/horizon-6cc494ff8-ggq9w" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.364278 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed7353a3-748a-40cb-8865-c4badeb8e402-horizon-tls-certs\") pod \"horizon-6cc494ff8-ggq9w\" (UID: \"ed7353a3-748a-40cb-8865-c4badeb8e402\") " pod="openstack/horizon-6cc494ff8-ggq9w" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.369227 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed7353a3-748a-40cb-8865-c4badeb8e402-combined-ca-bundle\") pod \"horizon-6cc494ff8-ggq9w\" (UID: \"ed7353a3-748a-40cb-8865-c4badeb8e402\") " pod="openstack/horizon-6cc494ff8-ggq9w" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.371265 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25qx6\" (UniqueName: \"kubernetes.io/projected/ed7353a3-748a-40cb-8865-c4badeb8e402-kube-api-access-25qx6\") pod \"horizon-6cc494ff8-ggq9w\" (UID: \"ed7353a3-748a-40cb-8865-c4badeb8e402\") " pod="openstack/horizon-6cc494ff8-ggq9w" Feb 27 17:21:38 crc kubenswrapper[4700]: I0227 17:21:38.408746 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6cc494ff8-ggq9w" Feb 27 17:21:39 crc kubenswrapper[4700]: I0227 17:21:39.844657 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-787484f569-fhzcg" Feb 27 17:21:39 crc kubenswrapper[4700]: I0227 17:21:39.904068 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb78597c7-wtsvw"] Feb 27 17:21:39 crc kubenswrapper[4700]: I0227 17:21:39.904312 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7cb78597c7-wtsvw" podUID="a902211b-c06e-42af-a1c8-599ab09fbd23" containerName="dnsmasq-dns" containerID="cri-o://a2577f97fd441ece5be682f8b2be1537ddf7b8c6658c22448f59f2d3be4b41da" gracePeriod=10 Feb 27 17:21:40 crc kubenswrapper[4700]: E0227 17:21:40.131416 4700 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda902211b_c06e_42af_a1c8_599ab09fbd23.slice/crio-conmon-a2577f97fd441ece5be682f8b2be1537ddf7b8c6658c22448f59f2d3be4b41da.scope\": RecentStats: unable to find data in memory cache]" Feb 27 17:21:40 crc kubenswrapper[4700]: I0227 17:21:40.492717 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7cb78597c7-wtsvw" podUID="a902211b-c06e-42af-a1c8-599ab09fbd23" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.143:5353: connect: connection refused" Feb 27 17:21:40 crc kubenswrapper[4700]: I0227 17:21:40.818835 4700 generic.go:334] "Generic (PLEG): container finished" podID="a902211b-c06e-42af-a1c8-599ab09fbd23" containerID="a2577f97fd441ece5be682f8b2be1537ddf7b8c6658c22448f59f2d3be4b41da" exitCode=0 Feb 27 17:21:40 crc kubenswrapper[4700]: I0227 17:21:40.818875 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb78597c7-wtsvw" event={"ID":"a902211b-c06e-42af-a1c8-599ab09fbd23","Type":"ContainerDied","Data":"a2577f97fd441ece5be682f8b2be1537ddf7b8c6658c22448f59f2d3be4b41da"} Feb 27 17:21:43 crc kubenswrapper[4700]: I0227 17:21:43.894218 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="08c973c7-192d-4326-9630-aa18aea6a484" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.155:9322/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 27 17:21:45 crc kubenswrapper[4700]: I0227 17:21:45.200149 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9hq27" Feb 27 17:21:45 crc kubenswrapper[4700]: I0227 17:21:45.327326 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/bb266c49-545c-4e58-a9f8-6dede53accdf-credential-keys\") pod \"bb266c49-545c-4e58-a9f8-6dede53accdf\" (UID: \"bb266c49-545c-4e58-a9f8-6dede53accdf\") " Feb 27 17:21:45 crc kubenswrapper[4700]: I0227 17:21:45.327409 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb266c49-545c-4e58-a9f8-6dede53accdf-combined-ca-bundle\") pod \"bb266c49-545c-4e58-a9f8-6dede53accdf\" (UID: \"bb266c49-545c-4e58-a9f8-6dede53accdf\") " Feb 27 17:21:45 crc kubenswrapper[4700]: I0227 17:21:45.327519 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dbd5\" (UniqueName: \"kubernetes.io/projected/bb266c49-545c-4e58-a9f8-6dede53accdf-kube-api-access-5dbd5\") pod \"bb266c49-545c-4e58-a9f8-6dede53accdf\" (UID: \"bb266c49-545c-4e58-a9f8-6dede53accdf\") " Feb 27 17:21:45 crc kubenswrapper[4700]: I0227 17:21:45.327567 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bb266c49-545c-4e58-a9f8-6dede53accdf-fernet-keys\") pod \"bb266c49-545c-4e58-a9f8-6dede53accdf\" (UID: \"bb266c49-545c-4e58-a9f8-6dede53accdf\") " Feb 27 17:21:45 crc kubenswrapper[4700]: I0227 17:21:45.327696 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb266c49-545c-4e58-a9f8-6dede53accdf-config-data\") pod \"bb266c49-545c-4e58-a9f8-6dede53accdf\" (UID: \"bb266c49-545c-4e58-a9f8-6dede53accdf\") " Feb 27 17:21:45 crc kubenswrapper[4700]: I0227 17:21:45.327776 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb266c49-545c-4e58-a9f8-6dede53accdf-scripts\") pod \"bb266c49-545c-4e58-a9f8-6dede53accdf\" (UID: \"bb266c49-545c-4e58-a9f8-6dede53accdf\") " Feb 27 17:21:45 crc kubenswrapper[4700]: I0227 17:21:45.334567 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb266c49-545c-4e58-a9f8-6dede53accdf-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "bb266c49-545c-4e58-a9f8-6dede53accdf" (UID: "bb266c49-545c-4e58-a9f8-6dede53accdf"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:21:45 crc kubenswrapper[4700]: I0227 17:21:45.334607 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb266c49-545c-4e58-a9f8-6dede53accdf-kube-api-access-5dbd5" (OuterVolumeSpecName: "kube-api-access-5dbd5") pod "bb266c49-545c-4e58-a9f8-6dede53accdf" (UID: "bb266c49-545c-4e58-a9f8-6dede53accdf"). InnerVolumeSpecName "kube-api-access-5dbd5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:21:45 crc kubenswrapper[4700]: I0227 17:21:45.334690 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb266c49-545c-4e58-a9f8-6dede53accdf-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "bb266c49-545c-4e58-a9f8-6dede53accdf" (UID: "bb266c49-545c-4e58-a9f8-6dede53accdf"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:21:45 crc kubenswrapper[4700]: I0227 17:21:45.335625 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb266c49-545c-4e58-a9f8-6dede53accdf-scripts" (OuterVolumeSpecName: "scripts") pod "bb266c49-545c-4e58-a9f8-6dede53accdf" (UID: "bb266c49-545c-4e58-a9f8-6dede53accdf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:21:45 crc kubenswrapper[4700]: I0227 17:21:45.355678 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb266c49-545c-4e58-a9f8-6dede53accdf-config-data" (OuterVolumeSpecName: "config-data") pod "bb266c49-545c-4e58-a9f8-6dede53accdf" (UID: "bb266c49-545c-4e58-a9f8-6dede53accdf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:21:45 crc kubenswrapper[4700]: I0227 17:21:45.362548 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb266c49-545c-4e58-a9f8-6dede53accdf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bb266c49-545c-4e58-a9f8-6dede53accdf" (UID: "bb266c49-545c-4e58-a9f8-6dede53accdf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:21:45 crc kubenswrapper[4700]: I0227 17:21:45.436657 4700 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb266c49-545c-4e58-a9f8-6dede53accdf-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:45 crc kubenswrapper[4700]: I0227 17:21:45.436690 4700 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/bb266c49-545c-4e58-a9f8-6dede53accdf-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:45 crc kubenswrapper[4700]: I0227 17:21:45.436702 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb266c49-545c-4e58-a9f8-6dede53accdf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:45 crc kubenswrapper[4700]: I0227 17:21:45.436710 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dbd5\" (UniqueName: \"kubernetes.io/projected/bb266c49-545c-4e58-a9f8-6dede53accdf-kube-api-access-5dbd5\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:45 crc kubenswrapper[4700]: I0227 17:21:45.436719 4700 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bb266c49-545c-4e58-a9f8-6dede53accdf-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:45 crc kubenswrapper[4700]: I0227 17:21:45.436726 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb266c49-545c-4e58-a9f8-6dede53accdf-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:45 crc kubenswrapper[4700]: I0227 17:21:45.492347 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7cb78597c7-wtsvw" podUID="a902211b-c06e-42af-a1c8-599ab09fbd23" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.143:5353: connect: connection refused" Feb 27 17:21:45 crc kubenswrapper[4700]: I0227 17:21:45.878022 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9hq27" event={"ID":"bb266c49-545c-4e58-a9f8-6dede53accdf","Type":"ContainerDied","Data":"b279799e1dae229de7d1b0cbba974b8e92040a99b241f0836a1b27c9a37daf90"} Feb 27 17:21:45 crc kubenswrapper[4700]: I0227 17:21:45.878082 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b279799e1dae229de7d1b0cbba974b8e92040a99b241f0836a1b27c9a37daf90" Feb 27 17:21:45 crc kubenswrapper[4700]: I0227 17:21:45.878141 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9hq27" Feb 27 17:21:46 crc kubenswrapper[4700]: I0227 17:21:46.287398 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-9hq27"] Feb 27 17:21:46 crc kubenswrapper[4700]: I0227 17:21:46.298205 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-9hq27"] Feb 27 17:21:46 crc kubenswrapper[4700]: I0227 17:21:46.387118 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-8xqhb"] Feb 27 17:21:46 crc kubenswrapper[4700]: E0227 17:21:46.388001 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb266c49-545c-4e58-a9f8-6dede53accdf" containerName="keystone-bootstrap" Feb 27 17:21:46 crc kubenswrapper[4700]: I0227 17:21:46.388029 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb266c49-545c-4e58-a9f8-6dede53accdf" containerName="keystone-bootstrap" Feb 27 17:21:46 crc kubenswrapper[4700]: I0227 17:21:46.388278 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb266c49-545c-4e58-a9f8-6dede53accdf" containerName="keystone-bootstrap" Feb 27 17:21:46 crc kubenswrapper[4700]: I0227 17:21:46.390067 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8xqhb" Feb 27 17:21:46 crc kubenswrapper[4700]: I0227 17:21:46.393437 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 27 17:21:46 crc kubenswrapper[4700]: I0227 17:21:46.393437 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 27 17:21:46 crc kubenswrapper[4700]: I0227 17:21:46.393677 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 27 17:21:46 crc kubenswrapper[4700]: I0227 17:21:46.393883 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-llljd" Feb 27 17:21:46 crc kubenswrapper[4700]: I0227 17:21:46.393890 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 27 17:21:46 crc kubenswrapper[4700]: I0227 17:21:46.399994 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-8xqhb"] Feb 27 17:21:46 crc kubenswrapper[4700]: I0227 17:21:46.563545 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84901b1e-0055-4489-8009-711c1676e51f-scripts\") pod \"keystone-bootstrap-8xqhb\" (UID: \"84901b1e-0055-4489-8009-711c1676e51f\") " pod="openstack/keystone-bootstrap-8xqhb" Feb 27 17:21:46 crc kubenswrapper[4700]: I0227 17:21:46.563623 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/84901b1e-0055-4489-8009-711c1676e51f-fernet-keys\") pod \"keystone-bootstrap-8xqhb\" (UID: \"84901b1e-0055-4489-8009-711c1676e51f\") " pod="openstack/keystone-bootstrap-8xqhb" Feb 27 17:21:46 crc kubenswrapper[4700]: I0227 17:21:46.563656 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84901b1e-0055-4489-8009-711c1676e51f-combined-ca-bundle\") pod \"keystone-bootstrap-8xqhb\" (UID: \"84901b1e-0055-4489-8009-711c1676e51f\") " pod="openstack/keystone-bootstrap-8xqhb" Feb 27 17:21:46 crc kubenswrapper[4700]: I0227 17:21:46.563748 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/84901b1e-0055-4489-8009-711c1676e51f-credential-keys\") pod \"keystone-bootstrap-8xqhb\" (UID: \"84901b1e-0055-4489-8009-711c1676e51f\") " pod="openstack/keystone-bootstrap-8xqhb" Feb 27 17:21:46 crc kubenswrapper[4700]: I0227 17:21:46.564122 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbklk\" (UniqueName: \"kubernetes.io/projected/84901b1e-0055-4489-8009-711c1676e51f-kube-api-access-qbklk\") pod \"keystone-bootstrap-8xqhb\" (UID: \"84901b1e-0055-4489-8009-711c1676e51f\") " pod="openstack/keystone-bootstrap-8xqhb" Feb 27 17:21:46 crc kubenswrapper[4700]: I0227 17:21:46.564181 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84901b1e-0055-4489-8009-711c1676e51f-config-data\") pod \"keystone-bootstrap-8xqhb\" (UID: \"84901b1e-0055-4489-8009-711c1676e51f\") " pod="openstack/keystone-bootstrap-8xqhb" Feb 27 17:21:46 crc kubenswrapper[4700]: I0227 17:21:46.665976 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbklk\" (UniqueName: \"kubernetes.io/projected/84901b1e-0055-4489-8009-711c1676e51f-kube-api-access-qbklk\") pod \"keystone-bootstrap-8xqhb\" (UID: \"84901b1e-0055-4489-8009-711c1676e51f\") " pod="openstack/keystone-bootstrap-8xqhb" Feb 27 17:21:46 crc kubenswrapper[4700]: I0227 17:21:46.666022 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84901b1e-0055-4489-8009-711c1676e51f-config-data\") pod \"keystone-bootstrap-8xqhb\" (UID: \"84901b1e-0055-4489-8009-711c1676e51f\") " pod="openstack/keystone-bootstrap-8xqhb" Feb 27 17:21:46 crc kubenswrapper[4700]: I0227 17:21:46.666102 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84901b1e-0055-4489-8009-711c1676e51f-scripts\") pod \"keystone-bootstrap-8xqhb\" (UID: \"84901b1e-0055-4489-8009-711c1676e51f\") " pod="openstack/keystone-bootstrap-8xqhb" Feb 27 17:21:46 crc kubenswrapper[4700]: I0227 17:21:46.666128 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/84901b1e-0055-4489-8009-711c1676e51f-fernet-keys\") pod \"keystone-bootstrap-8xqhb\" (UID: \"84901b1e-0055-4489-8009-711c1676e51f\") " pod="openstack/keystone-bootstrap-8xqhb" Feb 27 17:21:46 crc kubenswrapper[4700]: I0227 17:21:46.666143 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84901b1e-0055-4489-8009-711c1676e51f-combined-ca-bundle\") pod \"keystone-bootstrap-8xqhb\" (UID: \"84901b1e-0055-4489-8009-711c1676e51f\") " pod="openstack/keystone-bootstrap-8xqhb" Feb 27 17:21:46 crc kubenswrapper[4700]: I0227 17:21:46.666167 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/84901b1e-0055-4489-8009-711c1676e51f-credential-keys\") pod \"keystone-bootstrap-8xqhb\" (UID: \"84901b1e-0055-4489-8009-711c1676e51f\") " pod="openstack/keystone-bootstrap-8xqhb" Feb 27 17:21:46 crc kubenswrapper[4700]: I0227 17:21:46.681329 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84901b1e-0055-4489-8009-711c1676e51f-scripts\") pod \"keystone-bootstrap-8xqhb\" (UID: \"84901b1e-0055-4489-8009-711c1676e51f\") " pod="openstack/keystone-bootstrap-8xqhb" Feb 27 17:21:46 crc kubenswrapper[4700]: I0227 17:21:46.681380 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/84901b1e-0055-4489-8009-711c1676e51f-credential-keys\") pod \"keystone-bootstrap-8xqhb\" (UID: \"84901b1e-0055-4489-8009-711c1676e51f\") " pod="openstack/keystone-bootstrap-8xqhb" Feb 27 17:21:46 crc kubenswrapper[4700]: I0227 17:21:46.683515 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84901b1e-0055-4489-8009-711c1676e51f-config-data\") pod \"keystone-bootstrap-8xqhb\" (UID: \"84901b1e-0055-4489-8009-711c1676e51f\") " pod="openstack/keystone-bootstrap-8xqhb" Feb 27 17:21:46 crc kubenswrapper[4700]: I0227 17:21:46.685221 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/84901b1e-0055-4489-8009-711c1676e51f-fernet-keys\") pod \"keystone-bootstrap-8xqhb\" (UID: \"84901b1e-0055-4489-8009-711c1676e51f\") " pod="openstack/keystone-bootstrap-8xqhb" Feb 27 17:21:46 crc kubenswrapper[4700]: I0227 17:21:46.685899 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84901b1e-0055-4489-8009-711c1676e51f-combined-ca-bundle\") pod \"keystone-bootstrap-8xqhb\" (UID: \"84901b1e-0055-4489-8009-711c1676e51f\") " pod="openstack/keystone-bootstrap-8xqhb" Feb 27 17:21:46 crc kubenswrapper[4700]: I0227 17:21:46.687197 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbklk\" (UniqueName: \"kubernetes.io/projected/84901b1e-0055-4489-8009-711c1676e51f-kube-api-access-qbklk\") pod \"keystone-bootstrap-8xqhb\" (UID: \"84901b1e-0055-4489-8009-711c1676e51f\") " pod="openstack/keystone-bootstrap-8xqhb" Feb 27 17:21:46 crc kubenswrapper[4700]: I0227 17:21:46.713078 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8xqhb" Feb 27 17:21:46 crc kubenswrapper[4700]: I0227 17:21:46.992753 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb266c49-545c-4e58-a9f8-6dede53accdf" path="/var/lib/kubelet/pods/bb266c49-545c-4e58-a9f8-6dede53accdf/volumes" Feb 27 17:21:48 crc kubenswrapper[4700]: I0227 17:21:48.895409 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="08c973c7-192d-4326-9630-aa18aea6a484" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.155:9322/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 27 17:21:50 crc kubenswrapper[4700]: E0227 17:21:50.430961 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.145:5001/podified-master-centos10/openstack-horizon:watcher_latest" Feb 27 17:21:50 crc kubenswrapper[4700]: E0227 17:21:50.431006 4700 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.145:5001/podified-master-centos10/openstack-horizon:watcher_latest" Feb 27 17:21:50 crc kubenswrapper[4700]: E0227 17:21:50.431123 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:38.102.83.145:5001/podified-master-centos10/openstack-horizon:watcher_latest,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n65ch58h5bch64dh578h564h597h576h5c5hc5h58h8dh586h5d8hcdh5f4h548h79h657h546hbh5cfhb6h659h586h56chd8h5b8h94h665h5cbh665q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:yes,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-d7xnh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-d7b5ff8c7-zbshw_openstack(94005464-5a50-42ba-bb5f-b69b66cdec58): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 17:21:50 crc kubenswrapper[4700]: E0227 17:21:50.443184 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.145:5001/podified-master-centos10/openstack-horizon:watcher_latest" Feb 27 17:21:50 crc kubenswrapper[4700]: E0227 17:21:50.443240 4700 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.145:5001/podified-master-centos10/openstack-horizon:watcher_latest" Feb 27 17:21:50 crc kubenswrapper[4700]: E0227 17:21:50.443399 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:38.102.83.145:5001/podified-master-centos10/openstack-horizon:watcher_latest,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n64fh696h555h678h566h4h568h697h564h5ch6h9dh575hd7h5bdhd6h66h8fh68dhc5h67bh699h646h56dh65bhdfh9dh676h8h685h65h8fq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:yes,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-65vhc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-6fbcf8869-6nmv8_openstack(9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 17:21:50 crc kubenswrapper[4700]: E0227 17:21:50.455575 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.145:5001/podified-master-centos10/openstack-horizon:watcher_latest\\\"\"]" pod="openstack/horizon-6fbcf8869-6nmv8" podUID="9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e" Feb 27 17:21:50 crc kubenswrapper[4700]: E0227 17:21:50.455868 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.145:5001/podified-master-centos10/openstack-horizon:watcher_latest\\\"\"]" pod="openstack/horizon-d7b5ff8c7-zbshw" podUID="94005464-5a50-42ba-bb5f-b69b66cdec58" Feb 27 17:21:50 crc kubenswrapper[4700]: E0227 17:21:50.459845 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.145:5001/podified-master-centos10/openstack-horizon:watcher_latest" Feb 27 17:21:50 crc kubenswrapper[4700]: E0227 17:21:50.459877 4700 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.145:5001/podified-master-centos10/openstack-horizon:watcher_latest" Feb 27 17:21:50 crc kubenswrapper[4700]: E0227 17:21:50.459960 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:38.102.83.145:5001/podified-master-centos10/openstack-horizon:watcher_latest,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5f6h69hbbhf7h659h5cfh5d5hfbh8dhc9h67dh5f8h5f6h5c9h698h556h85h675hf9h5fch579h8ch6fh644h68fh545h545hb6hdfh65ch78h664q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:yes,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fjgr4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-5bc46db9f7-zczkg_openstack(aaf44cea-4a4a-4103-a7bd-a7c7a6d97342): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 17:21:50 crc kubenswrapper[4700]: E0227 17:21:50.461880 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.145:5001/podified-master-centos10/openstack-horizon:watcher_latest\\\"\"]" pod="openstack/horizon-5bc46db9f7-zczkg" podUID="aaf44cea-4a4a-4103-a7bd-a7c7a6d97342" Feb 27 17:21:50 crc kubenswrapper[4700]: I0227 17:21:50.494267 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7cb78597c7-wtsvw" podUID="a902211b-c06e-42af-a1c8-599ab09fbd23" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.143:5353: connect: connection refused" Feb 27 17:21:50 crc kubenswrapper[4700]: I0227 17:21:50.494404 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7cb78597c7-wtsvw" Feb 27 17:21:50 crc kubenswrapper[4700]: I0227 17:21:50.563486 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Feb 27 17:21:50 crc kubenswrapper[4700]: I0227 17:21:50.644278 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfll7\" (UniqueName: \"kubernetes.io/projected/08c973c7-192d-4326-9630-aa18aea6a484-kube-api-access-qfll7\") pod \"08c973c7-192d-4326-9630-aa18aea6a484\" (UID: \"08c973c7-192d-4326-9630-aa18aea6a484\") " Feb 27 17:21:50 crc kubenswrapper[4700]: I0227 17:21:50.644502 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/08c973c7-192d-4326-9630-aa18aea6a484-custom-prometheus-ca\") pod \"08c973c7-192d-4326-9630-aa18aea6a484\" (UID: \"08c973c7-192d-4326-9630-aa18aea6a484\") " Feb 27 17:21:50 crc kubenswrapper[4700]: I0227 17:21:50.644538 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08c973c7-192d-4326-9630-aa18aea6a484-combined-ca-bundle\") pod \"08c973c7-192d-4326-9630-aa18aea6a484\" (UID: \"08c973c7-192d-4326-9630-aa18aea6a484\") " Feb 27 17:21:50 crc kubenswrapper[4700]: I0227 17:21:50.644562 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08c973c7-192d-4326-9630-aa18aea6a484-config-data\") pod \"08c973c7-192d-4326-9630-aa18aea6a484\" (UID: \"08c973c7-192d-4326-9630-aa18aea6a484\") " Feb 27 17:21:50 crc kubenswrapper[4700]: I0227 17:21:50.644616 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08c973c7-192d-4326-9630-aa18aea6a484-logs\") pod \"08c973c7-192d-4326-9630-aa18aea6a484\" (UID: \"08c973c7-192d-4326-9630-aa18aea6a484\") " Feb 27 17:21:50 crc kubenswrapper[4700]: I0227 17:21:50.645096 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08c973c7-192d-4326-9630-aa18aea6a484-logs" (OuterVolumeSpecName: "logs") pod "08c973c7-192d-4326-9630-aa18aea6a484" (UID: "08c973c7-192d-4326-9630-aa18aea6a484"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:21:50 crc kubenswrapper[4700]: I0227 17:21:50.660684 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08c973c7-192d-4326-9630-aa18aea6a484-kube-api-access-qfll7" (OuterVolumeSpecName: "kube-api-access-qfll7") pod "08c973c7-192d-4326-9630-aa18aea6a484" (UID: "08c973c7-192d-4326-9630-aa18aea6a484"). InnerVolumeSpecName "kube-api-access-qfll7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:21:50 crc kubenswrapper[4700]: I0227 17:21:50.672428 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08c973c7-192d-4326-9630-aa18aea6a484-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "08c973c7-192d-4326-9630-aa18aea6a484" (UID: "08c973c7-192d-4326-9630-aa18aea6a484"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:21:50 crc kubenswrapper[4700]: I0227 17:21:50.684747 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08c973c7-192d-4326-9630-aa18aea6a484-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "08c973c7-192d-4326-9630-aa18aea6a484" (UID: "08c973c7-192d-4326-9630-aa18aea6a484"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:21:50 crc kubenswrapper[4700]: I0227 17:21:50.709701 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08c973c7-192d-4326-9630-aa18aea6a484-config-data" (OuterVolumeSpecName: "config-data") pod "08c973c7-192d-4326-9630-aa18aea6a484" (UID: "08c973c7-192d-4326-9630-aa18aea6a484"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:21:50 crc kubenswrapper[4700]: I0227 17:21:50.747285 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfll7\" (UniqueName: \"kubernetes.io/projected/08c973c7-192d-4326-9630-aa18aea6a484-kube-api-access-qfll7\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:50 crc kubenswrapper[4700]: I0227 17:21:50.747328 4700 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/08c973c7-192d-4326-9630-aa18aea6a484-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:50 crc kubenswrapper[4700]: I0227 17:21:50.747342 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08c973c7-192d-4326-9630-aa18aea6a484-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:50 crc kubenswrapper[4700]: I0227 17:21:50.747355 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08c973c7-192d-4326-9630-aa18aea6a484-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:50 crc kubenswrapper[4700]: I0227 17:21:50.747369 4700 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08c973c7-192d-4326-9630-aa18aea6a484-logs\") on node \"crc\" DevicePath \"\"" Feb 27 17:21:50 crc kubenswrapper[4700]: I0227 17:21:50.922685 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"08c973c7-192d-4326-9630-aa18aea6a484","Type":"ContainerDied","Data":"b08e4a607ca6eeb709af3df47cd5219301ea2dba85282c6bd9478f4ed577e16f"} Feb 27 17:21:50 crc kubenswrapper[4700]: I0227 17:21:50.922858 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Feb 27 17:21:51 crc kubenswrapper[4700]: I0227 17:21:51.066982 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Feb 27 17:21:51 crc kubenswrapper[4700]: I0227 17:21:51.067040 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-api-0"] Feb 27 17:21:51 crc kubenswrapper[4700]: I0227 17:21:51.092160 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-api-0"] Feb 27 17:21:51 crc kubenswrapper[4700]: E0227 17:21:51.154104 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08c973c7-192d-4326-9630-aa18aea6a484" containerName="watcher-api" Feb 27 17:21:51 crc kubenswrapper[4700]: I0227 17:21:51.154158 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="08c973c7-192d-4326-9630-aa18aea6a484" containerName="watcher-api" Feb 27 17:21:51 crc kubenswrapper[4700]: E0227 17:21:51.154201 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08c973c7-192d-4326-9630-aa18aea6a484" containerName="watcher-api-log" Feb 27 17:21:51 crc kubenswrapper[4700]: I0227 17:21:51.154207 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="08c973c7-192d-4326-9630-aa18aea6a484" containerName="watcher-api-log" Feb 27 17:21:51 crc kubenswrapper[4700]: I0227 17:21:51.154588 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="08c973c7-192d-4326-9630-aa18aea6a484" containerName="watcher-api" Feb 27 17:21:51 crc kubenswrapper[4700]: I0227 17:21:51.154627 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="08c973c7-192d-4326-9630-aa18aea6a484" containerName="watcher-api-log" Feb 27 17:21:51 crc kubenswrapper[4700]: I0227 17:21:51.156005 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Feb 27 17:21:51 crc kubenswrapper[4700]: I0227 17:21:51.160887 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-api-config-data" Feb 27 17:21:51 crc kubenswrapper[4700]: I0227 17:21:51.163166 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9822c457-a600-4d01-9b94-b7b899ef2102-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"9822c457-a600-4d01-9b94-b7b899ef2102\") " pod="openstack/watcher-api-0" Feb 27 17:21:51 crc kubenswrapper[4700]: I0227 17:21:51.163245 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rhx5\" (UniqueName: \"kubernetes.io/projected/9822c457-a600-4d01-9b94-b7b899ef2102-kube-api-access-8rhx5\") pod \"watcher-api-0\" (UID: \"9822c457-a600-4d01-9b94-b7b899ef2102\") " pod="openstack/watcher-api-0" Feb 27 17:21:51 crc kubenswrapper[4700]: I0227 17:21:51.163284 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9822c457-a600-4d01-9b94-b7b899ef2102-logs\") pod \"watcher-api-0\" (UID: \"9822c457-a600-4d01-9b94-b7b899ef2102\") " pod="openstack/watcher-api-0" Feb 27 17:21:51 crc kubenswrapper[4700]: I0227 17:21:51.163333 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/9822c457-a600-4d01-9b94-b7b899ef2102-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"9822c457-a600-4d01-9b94-b7b899ef2102\") " pod="openstack/watcher-api-0" Feb 27 17:21:51 crc kubenswrapper[4700]: I0227 17:21:51.163366 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9822c457-a600-4d01-9b94-b7b899ef2102-config-data\") pod \"watcher-api-0\" (UID: \"9822c457-a600-4d01-9b94-b7b899ef2102\") " pod="openstack/watcher-api-0" Feb 27 17:21:51 crc kubenswrapper[4700]: I0227 17:21:51.225581 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Feb 27 17:21:51 crc kubenswrapper[4700]: I0227 17:21:51.263934 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9822c457-a600-4d01-9b94-b7b899ef2102-logs\") pod \"watcher-api-0\" (UID: \"9822c457-a600-4d01-9b94-b7b899ef2102\") " pod="openstack/watcher-api-0" Feb 27 17:21:51 crc kubenswrapper[4700]: I0227 17:21:51.263999 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/9822c457-a600-4d01-9b94-b7b899ef2102-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"9822c457-a600-4d01-9b94-b7b899ef2102\") " pod="openstack/watcher-api-0" Feb 27 17:21:51 crc kubenswrapper[4700]: I0227 17:21:51.264028 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9822c457-a600-4d01-9b94-b7b899ef2102-config-data\") pod \"watcher-api-0\" (UID: \"9822c457-a600-4d01-9b94-b7b899ef2102\") " pod="openstack/watcher-api-0" Feb 27 17:21:51 crc kubenswrapper[4700]: I0227 17:21:51.264085 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9822c457-a600-4d01-9b94-b7b899ef2102-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"9822c457-a600-4d01-9b94-b7b899ef2102\") " pod="openstack/watcher-api-0" Feb 27 17:21:51 crc kubenswrapper[4700]: I0227 17:21:51.264125 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rhx5\" (UniqueName: \"kubernetes.io/projected/9822c457-a600-4d01-9b94-b7b899ef2102-kube-api-access-8rhx5\") pod \"watcher-api-0\" (UID: \"9822c457-a600-4d01-9b94-b7b899ef2102\") " pod="openstack/watcher-api-0" Feb 27 17:21:51 crc kubenswrapper[4700]: I0227 17:21:51.264763 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9822c457-a600-4d01-9b94-b7b899ef2102-logs\") pod \"watcher-api-0\" (UID: \"9822c457-a600-4d01-9b94-b7b899ef2102\") " pod="openstack/watcher-api-0" Feb 27 17:21:51 crc kubenswrapper[4700]: I0227 17:21:51.269078 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/9822c457-a600-4d01-9b94-b7b899ef2102-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"9822c457-a600-4d01-9b94-b7b899ef2102\") " pod="openstack/watcher-api-0" Feb 27 17:21:51 crc kubenswrapper[4700]: I0227 17:21:51.269255 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9822c457-a600-4d01-9b94-b7b899ef2102-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"9822c457-a600-4d01-9b94-b7b899ef2102\") " pod="openstack/watcher-api-0" Feb 27 17:21:51 crc kubenswrapper[4700]: I0227 17:21:51.273985 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9822c457-a600-4d01-9b94-b7b899ef2102-config-data\") pod \"watcher-api-0\" (UID: \"9822c457-a600-4d01-9b94-b7b899ef2102\") " pod="openstack/watcher-api-0" Feb 27 17:21:51 crc kubenswrapper[4700]: I0227 17:21:51.279349 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rhx5\" (UniqueName: \"kubernetes.io/projected/9822c457-a600-4d01-9b94-b7b899ef2102-kube-api-access-8rhx5\") pod \"watcher-api-0\" (UID: \"9822c457-a600-4d01-9b94-b7b899ef2102\") " pod="openstack/watcher-api-0" Feb 27 17:21:51 crc kubenswrapper[4700]: E0227 17:21:51.386129 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.145:5001/podified-master-centos10/openstack-barbican-api:watcher_latest" Feb 27 17:21:51 crc kubenswrapper[4700]: E0227 17:21:51.386168 4700 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.145:5001/podified-master-centos10/openstack-barbican-api:watcher_latest" Feb 27 17:21:51 crc kubenswrapper[4700]: E0227 17:21:51.386276 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:38.102.83.145:5001/podified-master-centos10/openstack-barbican-api:watcher_latest,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sssn7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-vrb9h_openstack(f72833f0-cf5f-4bf8-bb5f-1f65477c4b76): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 17:21:51 crc kubenswrapper[4700]: E0227 17:21:51.387529 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-vrb9h" podUID="f72833f0-cf5f-4bf8-bb5f-1f65477c4b76" Feb 27 17:21:51 crc kubenswrapper[4700]: I0227 17:21:51.530471 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Feb 27 17:21:51 crc kubenswrapper[4700]: E0227 17:21:51.934955 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.145:5001/podified-master-centos10/openstack-barbican-api:watcher_latest\\\"\"" pod="openstack/barbican-db-sync-vrb9h" podUID="f72833f0-cf5f-4bf8-bb5f-1f65477c4b76" Feb 27 17:21:52 crc kubenswrapper[4700]: I0227 17:21:52.994886 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08c973c7-192d-4326-9630-aa18aea6a484" path="/var/lib/kubelet/pods/08c973c7-192d-4326-9630-aa18aea6a484/volumes" Feb 27 17:21:53 crc kubenswrapper[4700]: I0227 17:21:53.896689 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="08c973c7-192d-4326-9630-aa18aea6a484" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.155:9322/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 27 17:21:55 crc kubenswrapper[4700]: I0227 17:21:55.493300 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7cb78597c7-wtsvw" podUID="a902211b-c06e-42af-a1c8-599ab09fbd23" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.143:5353: connect: connection refused" Feb 27 17:22:00 crc kubenswrapper[4700]: I0227 17:22:00.148261 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536882-csprw"] Feb 27 17:22:00 crc kubenswrapper[4700]: I0227 17:22:00.151171 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536882-csprw" Feb 27 17:22:00 crc kubenswrapper[4700]: I0227 17:22:00.155985 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:22:00 crc kubenswrapper[4700]: I0227 17:22:00.157153 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:22:00 crc kubenswrapper[4700]: I0227 17:22:00.157674 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d62kc" Feb 27 17:22:00 crc kubenswrapper[4700]: I0227 17:22:00.164321 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536882-csprw"] Feb 27 17:22:00 crc kubenswrapper[4700]: I0227 17:22:00.289963 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwd6w\" (UniqueName: \"kubernetes.io/projected/58023608-7e65-4290-b24c-ea314a3fb15b-kube-api-access-lwd6w\") pod \"auto-csr-approver-29536882-csprw\" (UID: \"58023608-7e65-4290-b24c-ea314a3fb15b\") " pod="openshift-infra/auto-csr-approver-29536882-csprw" Feb 27 17:22:00 crc kubenswrapper[4700]: I0227 17:22:00.391839 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwd6w\" (UniqueName: \"kubernetes.io/projected/58023608-7e65-4290-b24c-ea314a3fb15b-kube-api-access-lwd6w\") pod \"auto-csr-approver-29536882-csprw\" (UID: \"58023608-7e65-4290-b24c-ea314a3fb15b\") " pod="openshift-infra/auto-csr-approver-29536882-csprw" Feb 27 17:22:00 crc kubenswrapper[4700]: I0227 17:22:00.423891 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwd6w\" (UniqueName: \"kubernetes.io/projected/58023608-7e65-4290-b24c-ea314a3fb15b-kube-api-access-lwd6w\") pod \"auto-csr-approver-29536882-csprw\" (UID: \"58023608-7e65-4290-b24c-ea314a3fb15b\") " pod="openshift-infra/auto-csr-approver-29536882-csprw" Feb 27 17:22:00 crc kubenswrapper[4700]: I0227 17:22:00.541018 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536882-csprw" Feb 27 17:22:01 crc kubenswrapper[4700]: E0227 17:22:01.997823 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.145:5001/podified-master-centos10/openstack-ceilometer-central:watcher_latest" Feb 27 17:22:02 crc kubenswrapper[4700]: E0227 17:22:02.000679 4700 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.145:5001/podified-master-centos10/openstack-ceilometer-central:watcher_latest" Feb 27 17:22:02 crc kubenswrapper[4700]: E0227 17:22:02.000917 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:38.102.83.145:5001/podified-master-centos10/openstack-ceilometer-central:watcher_latest,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n645h56fh564hfh5cbh57fh64bh54bh64h666hcbh9bh5bfh99hc4h684h565hfdh595h565h97hbfh5b6h75h64bh5b4h6bh9fh685h567h5f8h557q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hm7np,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(61baccca-f7e6-4442-b612-8ab97b82036d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.033210 4700 scope.go:117] "RemoveContainer" containerID="0c16b57672e1a5f873e00382c5b2e11ad43caaad41b66d7f5c14965ad07952c4" Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.176285 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d7b5ff8c7-zbshw" Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.181226 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bc46db9f7-zczkg" Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.207259 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6fbcf8869-6nmv8" Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.228345 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/94005464-5a50-42ba-bb5f-b69b66cdec58-horizon-secret-key\") pod \"94005464-5a50-42ba-bb5f-b69b66cdec58\" (UID: \"94005464-5a50-42ba-bb5f-b69b66cdec58\") " Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.229527 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/aaf44cea-4a4a-4103-a7bd-a7c7a6d97342-horizon-secret-key\") pod \"aaf44cea-4a4a-4103-a7bd-a7c7a6d97342\" (UID: \"aaf44cea-4a4a-4103-a7bd-a7c7a6d97342\") " Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.229909 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d7xnh\" (UniqueName: \"kubernetes.io/projected/94005464-5a50-42ba-bb5f-b69b66cdec58-kube-api-access-d7xnh\") pod \"94005464-5a50-42ba-bb5f-b69b66cdec58\" (UID: \"94005464-5a50-42ba-bb5f-b69b66cdec58\") " Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.230059 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/aaf44cea-4a4a-4103-a7bd-a7c7a6d97342-config-data\") pod \"aaf44cea-4a4a-4103-a7bd-a7c7a6d97342\" (UID: \"aaf44cea-4a4a-4103-a7bd-a7c7a6d97342\") " Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.230083 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/94005464-5a50-42ba-bb5f-b69b66cdec58-scripts\") pod \"94005464-5a50-42ba-bb5f-b69b66cdec58\" (UID: \"94005464-5a50-42ba-bb5f-b69b66cdec58\") " Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.230119 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjgr4\" (UniqueName: \"kubernetes.io/projected/aaf44cea-4a4a-4103-a7bd-a7c7a6d97342-kube-api-access-fjgr4\") pod \"aaf44cea-4a4a-4103-a7bd-a7c7a6d97342\" (UID: \"aaf44cea-4a4a-4103-a7bd-a7c7a6d97342\") " Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.230158 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/94005464-5a50-42ba-bb5f-b69b66cdec58-config-data\") pod \"94005464-5a50-42ba-bb5f-b69b66cdec58\" (UID: \"94005464-5a50-42ba-bb5f-b69b66cdec58\") " Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.230175 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aaf44cea-4a4a-4103-a7bd-a7c7a6d97342-scripts\") pod \"aaf44cea-4a4a-4103-a7bd-a7c7a6d97342\" (UID: \"aaf44cea-4a4a-4103-a7bd-a7c7a6d97342\") " Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.230200 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94005464-5a50-42ba-bb5f-b69b66cdec58-logs\") pod \"94005464-5a50-42ba-bb5f-b69b66cdec58\" (UID: \"94005464-5a50-42ba-bb5f-b69b66cdec58\") " Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.230298 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aaf44cea-4a4a-4103-a7bd-a7c7a6d97342-logs\") pod \"aaf44cea-4a4a-4103-a7bd-a7c7a6d97342\" (UID: \"aaf44cea-4a4a-4103-a7bd-a7c7a6d97342\") " Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.231074 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94005464-5a50-42ba-bb5f-b69b66cdec58-scripts" (OuterVolumeSpecName: "scripts") pod "94005464-5a50-42ba-bb5f-b69b66cdec58" (UID: "94005464-5a50-42ba-bb5f-b69b66cdec58"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.231322 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94005464-5a50-42ba-bb5f-b69b66cdec58-config-data" (OuterVolumeSpecName: "config-data") pod "94005464-5a50-42ba-bb5f-b69b66cdec58" (UID: "94005464-5a50-42ba-bb5f-b69b66cdec58"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.231829 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94005464-5a50-42ba-bb5f-b69b66cdec58-logs" (OuterVolumeSpecName: "logs") pod "94005464-5a50-42ba-bb5f-b69b66cdec58" (UID: "94005464-5a50-42ba-bb5f-b69b66cdec58"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.232694 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aaf44cea-4a4a-4103-a7bd-a7c7a6d97342-scripts" (OuterVolumeSpecName: "scripts") pod "aaf44cea-4a4a-4103-a7bd-a7c7a6d97342" (UID: "aaf44cea-4a4a-4103-a7bd-a7c7a6d97342"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.232977 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aaf44cea-4a4a-4103-a7bd-a7c7a6d97342-logs" (OuterVolumeSpecName: "logs") pod "aaf44cea-4a4a-4103-a7bd-a7c7a6d97342" (UID: "aaf44cea-4a4a-4103-a7bd-a7c7a6d97342"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.233235 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aaf44cea-4a4a-4103-a7bd-a7c7a6d97342-config-data" (OuterVolumeSpecName: "config-data") pod "aaf44cea-4a4a-4103-a7bd-a7c7a6d97342" (UID: "aaf44cea-4a4a-4103-a7bd-a7c7a6d97342"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.235435 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94005464-5a50-42ba-bb5f-b69b66cdec58-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "94005464-5a50-42ba-bb5f-b69b66cdec58" (UID: "94005464-5a50-42ba-bb5f-b69b66cdec58"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.235918 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94005464-5a50-42ba-bb5f-b69b66cdec58-kube-api-access-d7xnh" (OuterVolumeSpecName: "kube-api-access-d7xnh") pod "94005464-5a50-42ba-bb5f-b69b66cdec58" (UID: "94005464-5a50-42ba-bb5f-b69b66cdec58"). InnerVolumeSpecName "kube-api-access-d7xnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.245126 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aaf44cea-4a4a-4103-a7bd-a7c7a6d97342-kube-api-access-fjgr4" (OuterVolumeSpecName: "kube-api-access-fjgr4") pod "aaf44cea-4a4a-4103-a7bd-a7c7a6d97342" (UID: "aaf44cea-4a4a-4103-a7bd-a7c7a6d97342"). InnerVolumeSpecName "kube-api-access-fjgr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.245828 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aaf44cea-4a4a-4103-a7bd-a7c7a6d97342-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "aaf44cea-4a4a-4103-a7bd-a7c7a6d97342" (UID: "aaf44cea-4a4a-4103-a7bd-a7c7a6d97342"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.332309 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e-config-data\") pod \"9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e\" (UID: \"9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e\") " Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.332439 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e-logs\") pod \"9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e\" (UID: \"9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e\") " Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.332568 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e-horizon-secret-key\") pod \"9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e\" (UID: \"9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e\") " Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.332676 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e-scripts\") pod \"9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e\" (UID: \"9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e\") " Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.332826 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65vhc\" (UniqueName: \"kubernetes.io/projected/9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e-kube-api-access-65vhc\") pod \"9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e\" (UID: \"9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e\") " Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.332966 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e-logs" (OuterVolumeSpecName: "logs") pod "9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e" (UID: "9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.333210 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e-config-data" (OuterVolumeSpecName: "config-data") pod "9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e" (UID: "9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.333305 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e-scripts" (OuterVolumeSpecName: "scripts") pod "9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e" (UID: "9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.333533 4700 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/94005464-5a50-42ba-bb5f-b69b66cdec58-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.333553 4700 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e-logs\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.333568 4700 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/aaf44cea-4a4a-4103-a7bd-a7c7a6d97342-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.333581 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d7xnh\" (UniqueName: \"kubernetes.io/projected/94005464-5a50-42ba-bb5f-b69b66cdec58-kube-api-access-d7xnh\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.333593 4700 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.333603 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/aaf44cea-4a4a-4103-a7bd-a7c7a6d97342-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.333613 4700 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/94005464-5a50-42ba-bb5f-b69b66cdec58-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.333622 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjgr4\" (UniqueName: \"kubernetes.io/projected/aaf44cea-4a4a-4103-a7bd-a7c7a6d97342-kube-api-access-fjgr4\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.333631 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/94005464-5a50-42ba-bb5f-b69b66cdec58-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.333641 4700 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aaf44cea-4a4a-4103-a7bd-a7c7a6d97342-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.333650 4700 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94005464-5a50-42ba-bb5f-b69b66cdec58-logs\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.333659 4700 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aaf44cea-4a4a-4103-a7bd-a7c7a6d97342-logs\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.333668 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.337774 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e" (UID: "9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.338185 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e-kube-api-access-65vhc" (OuterVolumeSpecName: "kube-api-access-65vhc") pod "9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e" (UID: "9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e"). InnerVolumeSpecName "kube-api-access-65vhc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.434762 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65vhc\" (UniqueName: \"kubernetes.io/projected/9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e-kube-api-access-65vhc\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:02 crc kubenswrapper[4700]: I0227 17:22:02.434790 4700 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:03 crc kubenswrapper[4700]: I0227 17:22:03.046970 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6fbcf8869-6nmv8" event={"ID":"9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e","Type":"ContainerDied","Data":"8335e89a460f97fa470a541a43c201d64acc7a6bf00853fec92ad673d370cb50"} Feb 27 17:22:03 crc kubenswrapper[4700]: I0227 17:22:03.047056 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6fbcf8869-6nmv8" Feb 27 17:22:03 crc kubenswrapper[4700]: I0227 17:22:03.052782 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d7b5ff8c7-zbshw" event={"ID":"94005464-5a50-42ba-bb5f-b69b66cdec58","Type":"ContainerDied","Data":"fb75180431b1242f03100293dddb74fedf5d27f5d6c2c5a0b72f0c9fed00201a"} Feb 27 17:22:03 crc kubenswrapper[4700]: I0227 17:22:03.052888 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d7b5ff8c7-zbshw" Feb 27 17:22:03 crc kubenswrapper[4700]: I0227 17:22:03.059967 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bc46db9f7-zczkg" event={"ID":"aaf44cea-4a4a-4103-a7bd-a7c7a6d97342","Type":"ContainerDied","Data":"6e77efac0053700f53620ec14cb309c5b7ab3a05d32f5660b466ab8225b88276"} Feb 27 17:22:03 crc kubenswrapper[4700]: I0227 17:22:03.060110 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bc46db9f7-zczkg" Feb 27 17:22:03 crc kubenswrapper[4700]: I0227 17:22:03.175776 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5bc46db9f7-zczkg"] Feb 27 17:22:03 crc kubenswrapper[4700]: I0227 17:22:03.188936 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5bc46db9f7-zczkg"] Feb 27 17:22:03 crc kubenswrapper[4700]: I0227 17:22:03.224613 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6fbcf8869-6nmv8"] Feb 27 17:22:03 crc kubenswrapper[4700]: I0227 17:22:03.233093 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6fbcf8869-6nmv8"] Feb 27 17:22:03 crc kubenswrapper[4700]: I0227 17:22:03.267809 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-d7b5ff8c7-zbshw"] Feb 27 17:22:03 crc kubenswrapper[4700]: I0227 17:22:03.275255 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-d7b5ff8c7-zbshw"] Feb 27 17:22:03 crc kubenswrapper[4700]: E0227 17:22:03.305920 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.145:5001/podified-master-centos10/openstack-cinder-api:watcher_latest" Feb 27 17:22:03 crc kubenswrapper[4700]: E0227 17:22:03.305964 4700 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.145:5001/podified-master-centos10/openstack-cinder-api:watcher_latest" Feb 27 17:22:03 crc kubenswrapper[4700]: E0227 17:22:03.306065 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:38.102.83.145:5001/podified-master-centos10/openstack-cinder-api:watcher_latest,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9nrcq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-mhcf5_openstack(dd3817a7-c2ab-43f0-9c74-b477ac59fa52): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 17:22:03 crc kubenswrapper[4700]: E0227 17:22:03.307811 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-mhcf5" podUID="dd3817a7-c2ab-43f0-9c74-b477ac59fa52" Feb 27 17:22:03 crc kubenswrapper[4700]: I0227 17:22:03.407518 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb78597c7-wtsvw" Feb 27 17:22:03 crc kubenswrapper[4700]: I0227 17:22:03.454420 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a902211b-c06e-42af-a1c8-599ab09fbd23-dns-swift-storage-0\") pod \"a902211b-c06e-42af-a1c8-599ab09fbd23\" (UID: \"a902211b-c06e-42af-a1c8-599ab09fbd23\") " Feb 27 17:22:03 crc kubenswrapper[4700]: I0227 17:22:03.457019 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a902211b-c06e-42af-a1c8-599ab09fbd23-ovsdbserver-sb\") pod \"a902211b-c06e-42af-a1c8-599ab09fbd23\" (UID: \"a902211b-c06e-42af-a1c8-599ab09fbd23\") " Feb 27 17:22:03 crc kubenswrapper[4700]: I0227 17:22:03.457119 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a902211b-c06e-42af-a1c8-599ab09fbd23-ovsdbserver-nb\") pod \"a902211b-c06e-42af-a1c8-599ab09fbd23\" (UID: \"a902211b-c06e-42af-a1c8-599ab09fbd23\") " Feb 27 17:22:03 crc kubenswrapper[4700]: I0227 17:22:03.457193 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a902211b-c06e-42af-a1c8-599ab09fbd23-config\") pod \"a902211b-c06e-42af-a1c8-599ab09fbd23\" (UID: \"a902211b-c06e-42af-a1c8-599ab09fbd23\") " Feb 27 17:22:03 crc kubenswrapper[4700]: I0227 17:22:03.457295 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sntls\" (UniqueName: \"kubernetes.io/projected/a902211b-c06e-42af-a1c8-599ab09fbd23-kube-api-access-sntls\") pod \"a902211b-c06e-42af-a1c8-599ab09fbd23\" (UID: \"a902211b-c06e-42af-a1c8-599ab09fbd23\") " Feb 27 17:22:03 crc kubenswrapper[4700]: I0227 17:22:03.457375 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a902211b-c06e-42af-a1c8-599ab09fbd23-dns-svc\") pod \"a902211b-c06e-42af-a1c8-599ab09fbd23\" (UID: \"a902211b-c06e-42af-a1c8-599ab09fbd23\") " Feb 27 17:22:03 crc kubenswrapper[4700]: I0227 17:22:03.479479 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a902211b-c06e-42af-a1c8-599ab09fbd23-kube-api-access-sntls" (OuterVolumeSpecName: "kube-api-access-sntls") pod "a902211b-c06e-42af-a1c8-599ab09fbd23" (UID: "a902211b-c06e-42af-a1c8-599ab09fbd23"). InnerVolumeSpecName "kube-api-access-sntls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:22:03 crc kubenswrapper[4700]: I0227 17:22:03.511757 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a902211b-c06e-42af-a1c8-599ab09fbd23-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a902211b-c06e-42af-a1c8-599ab09fbd23" (UID: "a902211b-c06e-42af-a1c8-599ab09fbd23"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:22:03 crc kubenswrapper[4700]: I0227 17:22:03.514804 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a902211b-c06e-42af-a1c8-599ab09fbd23-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a902211b-c06e-42af-a1c8-599ab09fbd23" (UID: "a902211b-c06e-42af-a1c8-599ab09fbd23"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:22:03 crc kubenswrapper[4700]: I0227 17:22:03.532024 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a902211b-c06e-42af-a1c8-599ab09fbd23-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a902211b-c06e-42af-a1c8-599ab09fbd23" (UID: "a902211b-c06e-42af-a1c8-599ab09fbd23"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:22:03 crc kubenswrapper[4700]: I0227 17:22:03.532177 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a902211b-c06e-42af-a1c8-599ab09fbd23-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a902211b-c06e-42af-a1c8-599ab09fbd23" (UID: "a902211b-c06e-42af-a1c8-599ab09fbd23"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:22:03 crc kubenswrapper[4700]: I0227 17:22:03.535342 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a902211b-c06e-42af-a1c8-599ab09fbd23-config" (OuterVolumeSpecName: "config") pod "a902211b-c06e-42af-a1c8-599ab09fbd23" (UID: "a902211b-c06e-42af-a1c8-599ab09fbd23"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:22:03 crc kubenswrapper[4700]: I0227 17:22:03.562297 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sntls\" (UniqueName: \"kubernetes.io/projected/a902211b-c06e-42af-a1c8-599ab09fbd23-kube-api-access-sntls\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:03 crc kubenswrapper[4700]: I0227 17:22:03.562328 4700 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a902211b-c06e-42af-a1c8-599ab09fbd23-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:03 crc kubenswrapper[4700]: I0227 17:22:03.562341 4700 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a902211b-c06e-42af-a1c8-599ab09fbd23-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:03 crc kubenswrapper[4700]: I0227 17:22:03.562350 4700 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a902211b-c06e-42af-a1c8-599ab09fbd23-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:03 crc kubenswrapper[4700]: I0227 17:22:03.562360 4700 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a902211b-c06e-42af-a1c8-599ab09fbd23-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:03 crc kubenswrapper[4700]: I0227 17:22:03.562368 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a902211b-c06e-42af-a1c8-599ab09fbd23-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:03 crc kubenswrapper[4700]: I0227 17:22:03.774038 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5b48777cdd-tqbkj"] Feb 27 17:22:03 crc kubenswrapper[4700]: I0227 17:22:03.810636 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6cc494ff8-ggq9w"] Feb 27 17:22:03 crc kubenswrapper[4700]: I0227 17:22:03.833689 4700 scope.go:117] "RemoveContainer" containerID="c9fac31959703f02774d5d27d913f34715a9cc958dba2d9feb8cabb37e8fa868" Feb 27 17:22:04 crc kubenswrapper[4700]: I0227 17:22:04.105724 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6cc494ff8-ggq9w" event={"ID":"ed7353a3-748a-40cb-8865-c4badeb8e402","Type":"ContainerStarted","Data":"12bfa59835fa4db6f7a70bc5afa0d19d87b695ad96aa64501b327f108fbbee16"} Feb 27 17:22:04 crc kubenswrapper[4700]: I0227 17:22:04.109036 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5b48777cdd-tqbkj" event={"ID":"d3989a3e-45ff-45d8-80a2-4dbe57663a57","Type":"ContainerStarted","Data":"1aa8f44a853b1bb812ffcf4b2d862218885ccb5e6483fcdaa92691b146310aab"} Feb 27 17:22:04 crc kubenswrapper[4700]: I0227 17:22:04.119006 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerStarted","Data":"7f3dc0b799607b839ef8d740e7569600055eb53b8ad9a51f0d8e9d8e3a22c10d"} Feb 27 17:22:04 crc kubenswrapper[4700]: I0227 17:22:04.134732 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb78597c7-wtsvw" Feb 27 17:22:04 crc kubenswrapper[4700]: I0227 17:22:04.138146 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb78597c7-wtsvw" event={"ID":"a902211b-c06e-42af-a1c8-599ab09fbd23","Type":"ContainerDied","Data":"697573a9dd246961321407de070eb79aa12d2c399c7054ca5394cc74119e2ae5"} Feb 27 17:22:04 crc kubenswrapper[4700]: I0227 17:22:04.138819 4700 scope.go:117] "RemoveContainer" containerID="a2577f97fd441ece5be682f8b2be1537ddf7b8c6658c22448f59f2d3be4b41da" Feb 27 17:22:04 crc kubenswrapper[4700]: E0227 17:22:04.139009 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.145:5001/podified-master-centos10/openstack-cinder-api:watcher_latest\\\"\"" pod="openstack/cinder-db-sync-mhcf5" podUID="dd3817a7-c2ab-43f0-9c74-b477ac59fa52" Feb 27 17:22:04 crc kubenswrapper[4700]: I0227 17:22:04.263777 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-8xqhb"] Feb 27 17:22:04 crc kubenswrapper[4700]: W0227 17:22:04.285501 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod84901b1e_0055_4489_8009_711c1676e51f.slice/crio-9d7d6854081b95b98be8993482e03c572e56d28f096ac89d8569715004cea457 WatchSource:0}: Error finding container 9d7d6854081b95b98be8993482e03c572e56d28f096ac89d8569715004cea457: Status 404 returned error can't find the container with id 9d7d6854081b95b98be8993482e03c572e56d28f096ac89d8569715004cea457 Feb 27 17:22:04 crc kubenswrapper[4700]: I0227 17:22:04.354129 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb78597c7-wtsvw"] Feb 27 17:22:04 crc kubenswrapper[4700]: I0227 17:22:04.355599 4700 scope.go:117] "RemoveContainer" containerID="c825731a613b3749d9fee537385612f6b55244071f3b74ed1a4fab210445940c" Feb 27 17:22:04 crc kubenswrapper[4700]: I0227 17:22:04.363492 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cb78597c7-wtsvw"] Feb 27 17:22:04 crc kubenswrapper[4700]: I0227 17:22:04.382841 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Feb 27 17:22:04 crc kubenswrapper[4700]: I0227 17:22:04.444821 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536882-csprw"] Feb 27 17:22:04 crc kubenswrapper[4700]: W0227 17:22:04.463405 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9822c457_a600_4d01_9b94_b7b899ef2102.slice/crio-658bae7ec88504e1af3430564755bf8041919992ca9fbcb804d379bf8a4b5692 WatchSource:0}: Error finding container 658bae7ec88504e1af3430564755bf8041919992ca9fbcb804d379bf8a4b5692: Status 404 returned error can't find the container with id 658bae7ec88504e1af3430564755bf8041919992ca9fbcb804d379bf8a4b5692 Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.023541 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94005464-5a50-42ba-bb5f-b69b66cdec58" path="/var/lib/kubelet/pods/94005464-5a50-42ba-bb5f-b69b66cdec58/volumes" Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.025365 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e" path="/var/lib/kubelet/pods/9e0da06c-4410-4fe1-bc2e-1c08e25a1a7e/volumes" Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.025797 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a902211b-c06e-42af-a1c8-599ab09fbd23" path="/var/lib/kubelet/pods/a902211b-c06e-42af-a1c8-599ab09fbd23/volumes" Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.026496 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aaf44cea-4a4a-4103-a7bd-a7c7a6d97342" path="/var/lib/kubelet/pods/aaf44cea-4a4a-4103-a7bd-a7c7a6d97342/volumes" Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.168182 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"9822c457-a600-4d01-9b94-b7b899ef2102","Type":"ContainerStarted","Data":"68026171c113261075b633116718884cfbc6f7e2bc616208c246fd9b332e8dab"} Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.168497 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"9822c457-a600-4d01-9b94-b7b899ef2102","Type":"ContainerStarted","Data":"4593dfb0b9741e168873d46911f0902e7eb87fedecd51d7a9be60ebce9628d31"} Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.168529 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"9822c457-a600-4d01-9b94-b7b899ef2102","Type":"ContainerStarted","Data":"658bae7ec88504e1af3430564755bf8041919992ca9fbcb804d379bf8a4b5692"} Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.168764 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.171003 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="9822c457-a600-4d01-9b94-b7b899ef2102" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.171:9322/\": dial tcp 10.217.0.171:9322: connect: connection refused" Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.187222 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536882-csprw" event={"ID":"58023608-7e65-4290-b24c-ea314a3fb15b","Type":"ContainerStarted","Data":"2915d1805d5e772beb347ce1b08adde377f9bbae29d3bdcef031e9b0e31b44f1"} Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.188764 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-api-0" podStartSLOduration=15.188753042 podStartE2EDuration="15.188753042s" podCreationTimestamp="2026-02-27 17:21:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:22:05.184782137 +0000 UTC m=+1285.170094884" watchObservedRunningTime="2026-02-27 17:22:05.188753042 +0000 UTC m=+1285.174065799" Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.189346 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8xqhb" event={"ID":"84901b1e-0055-4489-8009-711c1676e51f","Type":"ContainerStarted","Data":"3caf20a4365af9022601cc805ec79dcb3f0bc006f30ea14208f9a3981739664e"} Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.207180 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8xqhb" event={"ID":"84901b1e-0055-4489-8009-711c1676e51f","Type":"ContainerStarted","Data":"9d7d6854081b95b98be8993482e03c572e56d28f096ac89d8569715004cea457"} Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.226024 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5aae014e-6d34-47fb-9af7-3b1901025e51","Type":"ContainerStarted","Data":"8fbadff5661ea993281ad23be148279bc1f422c331d42e15f1e294ab8b80a273"} Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.226212 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5aae014e-6d34-47fb-9af7-3b1901025e51" containerName="glance-log" containerID="cri-o://481cd96f6890e8291a0d0587ef75e75c787464a33a4382b61f8a8f3c9dd0148f" gracePeriod=30 Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.226812 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5aae014e-6d34-47fb-9af7-3b1901025e51" containerName="glance-httpd" containerID="cri-o://8fbadff5661ea993281ad23be148279bc1f422c331d42e15f1e294ab8b80a273" gracePeriod=30 Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.276278 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-8xqhb" podStartSLOduration=19.276256736 podStartE2EDuration="19.276256736s" podCreationTimestamp="2026-02-27 17:21:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:22:05.229570511 +0000 UTC m=+1285.214883258" watchObservedRunningTime="2026-02-27 17:22:05.276256736 +0000 UTC m=+1285.261569483" Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.280605 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5810478f-8400-4b6d-b647-e7294aa087ba","Type":"ContainerStarted","Data":"0c12db820cf315188ff119c527bebdfb4c167c6ab9335a7466d8af31acf213ae"} Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.280792 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="5810478f-8400-4b6d-b647-e7294aa087ba" containerName="glance-log" containerID="cri-o://9560399dcda8d30788a5744b22b16ffe16678cbfd42a8ef1cf0f18a207bafcdf" gracePeriod=30 Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.281162 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=37.281152186 podStartE2EDuration="37.281152186s" podCreationTimestamp="2026-02-27 17:21:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:22:05.247868685 +0000 UTC m=+1285.233181432" watchObservedRunningTime="2026-02-27 17:22:05.281152186 +0000 UTC m=+1285.266464933" Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.281420 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="5810478f-8400-4b6d-b647-e7294aa087ba" containerName="glance-httpd" containerID="cri-o://0c12db820cf315188ff119c527bebdfb4c167c6ab9335a7466d8af31acf213ae" gracePeriod=30 Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.322957 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5b48777cdd-tqbkj" event={"ID":"d3989a3e-45ff-45d8-80a2-4dbe57663a57","Type":"ContainerStarted","Data":"cf749444d7c76d9317d18a0299503d9ccecc0cfd7ec5ffd81bd3d244886124dc"} Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.323001 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5b48777cdd-tqbkj" event={"ID":"d3989a3e-45ff-45d8-80a2-4dbe57663a57","Type":"ContainerStarted","Data":"2b6e983100e515e7297f4f273adfbef4735de5dcf00d127f6ac2101806ecb734"} Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.325916 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=37.32590487 podStartE2EDuration="37.32590487s" podCreationTimestamp="2026-02-27 17:21:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:22:05.325691864 +0000 UTC m=+1285.311004611" watchObservedRunningTime="2026-02-27 17:22:05.32590487 +0000 UTC m=+1285.311217617" Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.328141 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dmf4m" event={"ID":"989eacab-9c48-4375-b7a0-f2c7f40d17ca","Type":"ContainerStarted","Data":"4bcb91689f3aad19cad5a2592cf44a074600f816d8f0720250702e244a8da0db"} Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.338761 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"d8e03836-cf5d-4ac9-80fe-97556117e65e","Type":"ContainerStarted","Data":"0f536762f0202ca79c315c2b9f19d46b59472ea624b34e548af06e9322bb143e"} Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.356421 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5b48777cdd-tqbkj" podStartSLOduration=27.982733162 podStartE2EDuration="28.356403826s" podCreationTimestamp="2026-02-27 17:21:37 +0000 UTC" firstStartedPulling="2026-02-27 17:22:03.871330583 +0000 UTC m=+1283.856643330" lastFinishedPulling="2026-02-27 17:22:04.245001247 +0000 UTC m=+1284.230313994" observedRunningTime="2026-02-27 17:22:05.346479344 +0000 UTC m=+1285.331792091" watchObservedRunningTime="2026-02-27 17:22:05.356403826 +0000 UTC m=+1285.341716573" Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.358023 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"61baccca-f7e6-4442-b612-8ab97b82036d","Type":"ContainerStarted","Data":"cc73bd6dc0d5c792aec617a0df2f5575d481047b7e4f6a3c68c681a14ff1c241"} Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.389332 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6cc494ff8-ggq9w" event={"ID":"ed7353a3-748a-40cb-8865-c4badeb8e402","Type":"ContainerStarted","Data":"b5383652824fd95c82f9cc552832f56d2207860a853cf74e27f90f0f09d6a099"} Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.389370 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6cc494ff8-ggq9w" event={"ID":"ed7353a3-748a-40cb-8865-c4badeb8e402","Type":"ContainerStarted","Data":"1bf550724ad1bdfd4b2bb082a26cef13640e77c19305db4e038b9f9450aeb307"} Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.402017 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-decision-engine-0" podStartSLOduration=6.093562859 podStartE2EDuration="37.401998083s" podCreationTimestamp="2026-02-27 17:21:28 +0000 UTC" firstStartedPulling="2026-02-27 17:21:30.691746623 +0000 UTC m=+1250.677059370" lastFinishedPulling="2026-02-27 17:22:02.000181767 +0000 UTC m=+1281.985494594" observedRunningTime="2026-02-27 17:22:05.382030994 +0000 UTC m=+1285.367343741" watchObservedRunningTime="2026-02-27 17:22:05.401998083 +0000 UTC m=+1285.387310830" Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.403266 4700 generic.go:334] "Generic (PLEG): container finished" podID="3f209c04-79ea-4048-9d31-6222874de091" containerID="1995dc0741ace72e33bf0776b3003b5630b6888aacbfc518cefa2974898a9358" exitCode=0 Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.403325 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-6b5ck" event={"ID":"3f209c04-79ea-4048-9d31-6222874de091","Type":"ContainerDied","Data":"1995dc0741ace72e33bf0776b3003b5630b6888aacbfc518cefa2974898a9358"} Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.415107 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-dmf4m" podStartSLOduration=6.343815858 podStartE2EDuration="37.415089029s" podCreationTimestamp="2026-02-27 17:21:28 +0000 UTC" firstStartedPulling="2026-02-27 17:21:30.929837451 +0000 UTC m=+1250.915150198" lastFinishedPulling="2026-02-27 17:22:02.001110582 +0000 UTC m=+1281.986423369" observedRunningTime="2026-02-27 17:22:05.399486776 +0000 UTC m=+1285.384799523" watchObservedRunningTime="2026-02-27 17:22:05.415089029 +0000 UTC m=+1285.400401776" Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.433324 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6cc494ff8-ggq9w" podStartSLOduration=27.152053841 podStartE2EDuration="27.433308361s" podCreationTimestamp="2026-02-27 17:21:38 +0000 UTC" firstStartedPulling="2026-02-27 17:22:03.962630478 +0000 UTC m=+1283.947943225" lastFinishedPulling="2026-02-27 17:22:04.243884998 +0000 UTC m=+1284.229197745" observedRunningTime="2026-02-27 17:22:05.42305338 +0000 UTC m=+1285.408366117" watchObservedRunningTime="2026-02-27 17:22:05.433308361 +0000 UTC m=+1285.418621108" Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.435514 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"4099c554-cf42-48b3-88cb-e672a66f1a05","Type":"ContainerStarted","Data":"790d64d806f8cf030b5fa73b8846ad67dd2fb782f8fff50bd6090997023776d8"} Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.481561 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-applier-0" podStartSLOduration=17.72709235 podStartE2EDuration="37.481542527s" podCreationTimestamp="2026-02-27 17:21:28 +0000 UTC" firstStartedPulling="2026-02-27 17:21:30.678198555 +0000 UTC m=+1250.663511302" lastFinishedPulling="2026-02-27 17:21:50.432648732 +0000 UTC m=+1270.417961479" observedRunningTime="2026-02-27 17:22:05.471901802 +0000 UTC m=+1285.457214549" watchObservedRunningTime="2026-02-27 17:22:05.481542527 +0000 UTC m=+1285.466855274" Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.492700 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7cb78597c7-wtsvw" podUID="a902211b-c06e-42af-a1c8-599ab09fbd23" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.143:5353: i/o timeout" Feb 27 17:22:05 crc kubenswrapper[4700]: I0227 17:22:05.924132 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.033361 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"5810478f-8400-4b6d-b647-e7294aa087ba\" (UID: \"5810478f-8400-4b6d-b647-e7294aa087ba\") " Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.033531 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5810478f-8400-4b6d-b647-e7294aa087ba-httpd-run\") pod \"5810478f-8400-4b6d-b647-e7294aa087ba\" (UID: \"5810478f-8400-4b6d-b647-e7294aa087ba\") " Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.033554 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5810478f-8400-4b6d-b647-e7294aa087ba-scripts\") pod \"5810478f-8400-4b6d-b647-e7294aa087ba\" (UID: \"5810478f-8400-4b6d-b647-e7294aa087ba\") " Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.033588 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5810478f-8400-4b6d-b647-e7294aa087ba-combined-ca-bundle\") pod \"5810478f-8400-4b6d-b647-e7294aa087ba\" (UID: \"5810478f-8400-4b6d-b647-e7294aa087ba\") " Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.033680 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5810478f-8400-4b6d-b647-e7294aa087ba-logs\") pod \"5810478f-8400-4b6d-b647-e7294aa087ba\" (UID: \"5810478f-8400-4b6d-b647-e7294aa087ba\") " Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.033739 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5rmn\" (UniqueName: \"kubernetes.io/projected/5810478f-8400-4b6d-b647-e7294aa087ba-kube-api-access-q5rmn\") pod \"5810478f-8400-4b6d-b647-e7294aa087ba\" (UID: \"5810478f-8400-4b6d-b647-e7294aa087ba\") " Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.033778 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5810478f-8400-4b6d-b647-e7294aa087ba-config-data\") pod \"5810478f-8400-4b6d-b647-e7294aa087ba\" (UID: \"5810478f-8400-4b6d-b647-e7294aa087ba\") " Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.033813 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5810478f-8400-4b6d-b647-e7294aa087ba-public-tls-certs\") pod \"5810478f-8400-4b6d-b647-e7294aa087ba\" (UID: \"5810478f-8400-4b6d-b647-e7294aa087ba\") " Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.036973 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5810478f-8400-4b6d-b647-e7294aa087ba-logs" (OuterVolumeSpecName: "logs") pod "5810478f-8400-4b6d-b647-e7294aa087ba" (UID: "5810478f-8400-4b6d-b647-e7294aa087ba"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.037290 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5810478f-8400-4b6d-b647-e7294aa087ba-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5810478f-8400-4b6d-b647-e7294aa087ba" (UID: "5810478f-8400-4b6d-b647-e7294aa087ba"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.063167 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5810478f-8400-4b6d-b647-e7294aa087ba-kube-api-access-q5rmn" (OuterVolumeSpecName: "kube-api-access-q5rmn") pod "5810478f-8400-4b6d-b647-e7294aa087ba" (UID: "5810478f-8400-4b6d-b647-e7294aa087ba"). InnerVolumeSpecName "kube-api-access-q5rmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.077656 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "5810478f-8400-4b6d-b647-e7294aa087ba" (UID: "5810478f-8400-4b6d-b647-e7294aa087ba"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.091976 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5810478f-8400-4b6d-b647-e7294aa087ba-scripts" (OuterVolumeSpecName: "scripts") pod "5810478f-8400-4b6d-b647-e7294aa087ba" (UID: "5810478f-8400-4b6d-b647-e7294aa087ba"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.139691 4700 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.139739 4700 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5810478f-8400-4b6d-b647-e7294aa087ba-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.139748 4700 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5810478f-8400-4b6d-b647-e7294aa087ba-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.139756 4700 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5810478f-8400-4b6d-b647-e7294aa087ba-logs\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.139764 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5rmn\" (UniqueName: \"kubernetes.io/projected/5810478f-8400-4b6d-b647-e7294aa087ba-kube-api-access-q5rmn\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.179330 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5810478f-8400-4b6d-b647-e7294aa087ba-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "5810478f-8400-4b6d-b647-e7294aa087ba" (UID: "5810478f-8400-4b6d-b647-e7294aa087ba"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.220626 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5810478f-8400-4b6d-b647-e7294aa087ba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5810478f-8400-4b6d-b647-e7294aa087ba" (UID: "5810478f-8400-4b6d-b647-e7294aa087ba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.236431 4700 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.244745 4700 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5810478f-8400-4b6d-b647-e7294aa087ba-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.244785 4700 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.244797 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5810478f-8400-4b6d-b647-e7294aa087ba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.268611 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5810478f-8400-4b6d-b647-e7294aa087ba-config-data" (OuterVolumeSpecName: "config-data") pod "5810478f-8400-4b6d-b647-e7294aa087ba" (UID: "5810478f-8400-4b6d-b647-e7294aa087ba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.337648 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.348776 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5810478f-8400-4b6d-b647-e7294aa087ba-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.449538 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bzfhv\" (UniqueName: \"kubernetes.io/projected/5aae014e-6d34-47fb-9af7-3b1901025e51-kube-api-access-bzfhv\") pod \"5aae014e-6d34-47fb-9af7-3b1901025e51\" (UID: \"5aae014e-6d34-47fb-9af7-3b1901025e51\") " Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.449591 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"5aae014e-6d34-47fb-9af7-3b1901025e51\" (UID: \"5aae014e-6d34-47fb-9af7-3b1901025e51\") " Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.449678 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5aae014e-6d34-47fb-9af7-3b1901025e51-internal-tls-certs\") pod \"5aae014e-6d34-47fb-9af7-3b1901025e51\" (UID: \"5aae014e-6d34-47fb-9af7-3b1901025e51\") " Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.449705 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5aae014e-6d34-47fb-9af7-3b1901025e51-logs\") pod \"5aae014e-6d34-47fb-9af7-3b1901025e51\" (UID: \"5aae014e-6d34-47fb-9af7-3b1901025e51\") " Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.449770 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5aae014e-6d34-47fb-9af7-3b1901025e51-scripts\") pod \"5aae014e-6d34-47fb-9af7-3b1901025e51\" (UID: \"5aae014e-6d34-47fb-9af7-3b1901025e51\") " Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.449834 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aae014e-6d34-47fb-9af7-3b1901025e51-combined-ca-bundle\") pod \"5aae014e-6d34-47fb-9af7-3b1901025e51\" (UID: \"5aae014e-6d34-47fb-9af7-3b1901025e51\") " Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.449906 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5aae014e-6d34-47fb-9af7-3b1901025e51-config-data\") pod \"5aae014e-6d34-47fb-9af7-3b1901025e51\" (UID: \"5aae014e-6d34-47fb-9af7-3b1901025e51\") " Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.449929 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5aae014e-6d34-47fb-9af7-3b1901025e51-httpd-run\") pod \"5aae014e-6d34-47fb-9af7-3b1901025e51\" (UID: \"5aae014e-6d34-47fb-9af7-3b1901025e51\") " Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.451156 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5aae014e-6d34-47fb-9af7-3b1901025e51-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5aae014e-6d34-47fb-9af7-3b1901025e51" (UID: "5aae014e-6d34-47fb-9af7-3b1901025e51"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.451374 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5aae014e-6d34-47fb-9af7-3b1901025e51-logs" (OuterVolumeSpecName: "logs") pod "5aae014e-6d34-47fb-9af7-3b1901025e51" (UID: "5aae014e-6d34-47fb-9af7-3b1901025e51"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.459295 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5aae014e-6d34-47fb-9af7-3b1901025e51-kube-api-access-bzfhv" (OuterVolumeSpecName: "kube-api-access-bzfhv") pod "5aae014e-6d34-47fb-9af7-3b1901025e51" (UID: "5aae014e-6d34-47fb-9af7-3b1901025e51"). InnerVolumeSpecName "kube-api-access-bzfhv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.461688 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "5aae014e-6d34-47fb-9af7-3b1901025e51" (UID: "5aae014e-6d34-47fb-9af7-3b1901025e51"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.463736 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5aae014e-6d34-47fb-9af7-3b1901025e51-scripts" (OuterVolumeSpecName: "scripts") pod "5aae014e-6d34-47fb-9af7-3b1901025e51" (UID: "5aae014e-6d34-47fb-9af7-3b1901025e51"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.475982 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.477159 4700 generic.go:334] "Generic (PLEG): container finished" podID="5810478f-8400-4b6d-b647-e7294aa087ba" containerID="0c12db820cf315188ff119c527bebdfb4c167c6ab9335a7466d8af31acf213ae" exitCode=0 Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.477216 4700 generic.go:334] "Generic (PLEG): container finished" podID="5810478f-8400-4b6d-b647-e7294aa087ba" containerID="9560399dcda8d30788a5744b22b16ffe16678cbfd42a8ef1cf0f18a207bafcdf" exitCode=143 Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.477283 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5810478f-8400-4b6d-b647-e7294aa087ba","Type":"ContainerDied","Data":"0c12db820cf315188ff119c527bebdfb4c167c6ab9335a7466d8af31acf213ae"} Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.477316 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5810478f-8400-4b6d-b647-e7294aa087ba","Type":"ContainerDied","Data":"9560399dcda8d30788a5744b22b16ffe16678cbfd42a8ef1cf0f18a207bafcdf"} Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.477331 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5810478f-8400-4b6d-b647-e7294aa087ba","Type":"ContainerDied","Data":"7349ff6f87ac9b272915a050d3c4968356affd17eed934469ccf251f668fb25d"} Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.477351 4700 scope.go:117] "RemoveContainer" containerID="0c12db820cf315188ff119c527bebdfb4c167c6ab9335a7466d8af31acf213ae" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.478686 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5aae014e-6d34-47fb-9af7-3b1901025e51-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5aae014e-6d34-47fb-9af7-3b1901025e51" (UID: "5aae014e-6d34-47fb-9af7-3b1901025e51"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.482757 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-vrb9h" event={"ID":"f72833f0-cf5f-4bf8-bb5f-1f65477c4b76","Type":"ContainerStarted","Data":"0f2f183bfd914f6252dd78997763889b1e5cc74a29e4d9af31c6456aceb88f96"} Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.492768 4700 generic.go:334] "Generic (PLEG): container finished" podID="5aae014e-6d34-47fb-9af7-3b1901025e51" containerID="8fbadff5661ea993281ad23be148279bc1f422c331d42e15f1e294ab8b80a273" exitCode=0 Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.492797 4700 generic.go:334] "Generic (PLEG): container finished" podID="5aae014e-6d34-47fb-9af7-3b1901025e51" containerID="481cd96f6890e8291a0d0587ef75e75c787464a33a4382b61f8a8f3c9dd0148f" exitCode=143 Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.493574 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.493980 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5aae014e-6d34-47fb-9af7-3b1901025e51","Type":"ContainerDied","Data":"8fbadff5661ea993281ad23be148279bc1f422c331d42e15f1e294ab8b80a273"} Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.494013 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5aae014e-6d34-47fb-9af7-3b1901025e51","Type":"ContainerDied","Data":"481cd96f6890e8291a0d0587ef75e75c787464a33a4382b61f8a8f3c9dd0148f"} Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.494023 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5aae014e-6d34-47fb-9af7-3b1901025e51","Type":"ContainerDied","Data":"6ae74025b872ed814351ea8c739b32a31304a07fdc82802f42b24034bdccc514"} Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.531936 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.533093 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5aae014e-6d34-47fb-9af7-3b1901025e51-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5aae014e-6d34-47fb-9af7-3b1901025e51" (UID: "5aae014e-6d34-47fb-9af7-3b1901025e51"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.533474 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-vrb9h" podStartSLOduration=4.205028362 podStartE2EDuration="38.533443141s" podCreationTimestamp="2026-02-27 17:21:28 +0000 UTC" firstStartedPulling="2026-02-27 17:21:30.765574556 +0000 UTC m=+1250.750887303" lastFinishedPulling="2026-02-27 17:22:05.093989335 +0000 UTC m=+1285.079302082" observedRunningTime="2026-02-27 17:22:06.503039517 +0000 UTC m=+1286.488352264" watchObservedRunningTime="2026-02-27 17:22:06.533443141 +0000 UTC m=+1286.518755888" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.554099 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.560595 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aae014e-6d34-47fb-9af7-3b1901025e51-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.560624 4700 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5aae014e-6d34-47fb-9af7-3b1901025e51-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.560634 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bzfhv\" (UniqueName: \"kubernetes.io/projected/5aae014e-6d34-47fb-9af7-3b1901025e51-kube-api-access-bzfhv\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.560656 4700 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.560666 4700 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5aae014e-6d34-47fb-9af7-3b1901025e51-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.560675 4700 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5aae014e-6d34-47fb-9af7-3b1901025e51-logs\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.560683 4700 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5aae014e-6d34-47fb-9af7-3b1901025e51-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.567171 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.570916 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5aae014e-6d34-47fb-9af7-3b1901025e51-config-data" (OuterVolumeSpecName: "config-data") pod "5aae014e-6d34-47fb-9af7-3b1901025e51" (UID: "5aae014e-6d34-47fb-9af7-3b1901025e51"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.575078 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 17:22:06 crc kubenswrapper[4700]: E0227 17:22:06.575831 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5810478f-8400-4b6d-b647-e7294aa087ba" containerName="glance-httpd" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.575848 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="5810478f-8400-4b6d-b647-e7294aa087ba" containerName="glance-httpd" Feb 27 17:22:06 crc kubenswrapper[4700]: E0227 17:22:06.575866 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a902211b-c06e-42af-a1c8-599ab09fbd23" containerName="dnsmasq-dns" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.575872 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="a902211b-c06e-42af-a1c8-599ab09fbd23" containerName="dnsmasq-dns" Feb 27 17:22:06 crc kubenswrapper[4700]: E0227 17:22:06.575884 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5aae014e-6d34-47fb-9af7-3b1901025e51" containerName="glance-httpd" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.575890 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="5aae014e-6d34-47fb-9af7-3b1901025e51" containerName="glance-httpd" Feb 27 17:22:06 crc kubenswrapper[4700]: E0227 17:22:06.575919 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5aae014e-6d34-47fb-9af7-3b1901025e51" containerName="glance-log" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.575925 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="5aae014e-6d34-47fb-9af7-3b1901025e51" containerName="glance-log" Feb 27 17:22:06 crc kubenswrapper[4700]: E0227 17:22:06.575940 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a902211b-c06e-42af-a1c8-599ab09fbd23" containerName="init" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.575947 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="a902211b-c06e-42af-a1c8-599ab09fbd23" containerName="init" Feb 27 17:22:06 crc kubenswrapper[4700]: E0227 17:22:06.575960 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5810478f-8400-4b6d-b647-e7294aa087ba" containerName="glance-log" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.575965 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="5810478f-8400-4b6d-b647-e7294aa087ba" containerName="glance-log" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.576119 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="5810478f-8400-4b6d-b647-e7294aa087ba" containerName="glance-httpd" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.576131 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="5aae014e-6d34-47fb-9af7-3b1901025e51" containerName="glance-log" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.576137 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="a902211b-c06e-42af-a1c8-599ab09fbd23" containerName="dnsmasq-dns" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.576153 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="5aae014e-6d34-47fb-9af7-3b1901025e51" containerName="glance-httpd" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.576165 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="5810478f-8400-4b6d-b647-e7294aa087ba" containerName="glance-log" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.577070 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.591110 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.591351 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.599971 4700 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.616805 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.667553 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-logs\") pod \"glance-default-external-api-0\" (UID: \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\") " pod="openstack/glance-default-external-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.667855 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-scripts\") pod \"glance-default-external-api-0\" (UID: \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\") " pod="openstack/glance-default-external-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.667885 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\") " pod="openstack/glance-default-external-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.667924 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\") " pod="openstack/glance-default-external-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.667946 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-config-data\") pod \"glance-default-external-api-0\" (UID: \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\") " pod="openstack/glance-default-external-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.668033 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\") " pod="openstack/glance-default-external-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.668052 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\") " pod="openstack/glance-default-external-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.668078 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lt89w\" (UniqueName: \"kubernetes.io/projected/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-kube-api-access-lt89w\") pod \"glance-default-external-api-0\" (UID: \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\") " pod="openstack/glance-default-external-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.668172 4700 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.668185 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5aae014e-6d34-47fb-9af7-3b1901025e51-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.772849 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\") " pod="openstack/glance-default-external-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.772890 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-config-data\") pod \"glance-default-external-api-0\" (UID: \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\") " pod="openstack/glance-default-external-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.772970 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\") " pod="openstack/glance-default-external-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.772994 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\") " pod="openstack/glance-default-external-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.773028 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lt89w\" (UniqueName: \"kubernetes.io/projected/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-kube-api-access-lt89w\") pod \"glance-default-external-api-0\" (UID: \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\") " pod="openstack/glance-default-external-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.773109 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-logs\") pod \"glance-default-external-api-0\" (UID: \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\") " pod="openstack/glance-default-external-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.773140 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-scripts\") pod \"glance-default-external-api-0\" (UID: \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\") " pod="openstack/glance-default-external-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.773167 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\") " pod="openstack/glance-default-external-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.773377 4700 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.773504 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\") " pod="openstack/glance-default-external-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.774288 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-logs\") pod \"glance-default-external-api-0\" (UID: \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\") " pod="openstack/glance-default-external-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.779123 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-scripts\") pod \"glance-default-external-api-0\" (UID: \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\") " pod="openstack/glance-default-external-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.782092 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-config-data\") pod \"glance-default-external-api-0\" (UID: \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\") " pod="openstack/glance-default-external-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.782615 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\") " pod="openstack/glance-default-external-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.787434 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\") " pod="openstack/glance-default-external-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.816314 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lt89w\" (UniqueName: \"kubernetes.io/projected/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-kube-api-access-lt89w\") pod \"glance-default-external-api-0\" (UID: \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\") " pod="openstack/glance-default-external-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.857773 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\") " pod="openstack/glance-default-external-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.913524 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.925561 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.928286 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.938810 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.940438 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.945060 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.949027 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.960266 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.978502 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.978547 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-config-data\") pod \"glance-default-internal-api-0\" (UID: \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.978618 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.978650 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-scripts\") pod \"glance-default-internal-api-0\" (UID: \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.978667 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-logs\") pod \"glance-default-internal-api-0\" (UID: \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.978689 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.978720 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zr8ct\" (UniqueName: \"kubernetes.io/projected/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-kube-api-access-zr8ct\") pod \"glance-default-internal-api-0\" (UID: \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.978782 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.998094 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5810478f-8400-4b6d-b647-e7294aa087ba" path="/var/lib/kubelet/pods/5810478f-8400-4b6d-b647-e7294aa087ba/volumes" Feb 27 17:22:06 crc kubenswrapper[4700]: I0227 17:22:06.999135 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5aae014e-6d34-47fb-9af7-3b1901025e51" path="/var/lib/kubelet/pods/5aae014e-6d34-47fb-9af7-3b1901025e51/volumes" Feb 27 17:22:07 crc kubenswrapper[4700]: I0227 17:22:07.081152 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:22:07 crc kubenswrapper[4700]: I0227 17:22:07.081718 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:22:07 crc kubenswrapper[4700]: I0227 17:22:07.081819 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-config-data\") pod \"glance-default-internal-api-0\" (UID: \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:22:07 crc kubenswrapper[4700]: I0227 17:22:07.081973 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:22:07 crc kubenswrapper[4700]: I0227 17:22:07.082058 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-scripts\") pod \"glance-default-internal-api-0\" (UID: \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:22:07 crc kubenswrapper[4700]: I0227 17:22:07.082138 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-logs\") pod \"glance-default-internal-api-0\" (UID: \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:22:07 crc kubenswrapper[4700]: I0227 17:22:07.082217 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:22:07 crc kubenswrapper[4700]: I0227 17:22:07.082299 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zr8ct\" (UniqueName: \"kubernetes.io/projected/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-kube-api-access-zr8ct\") pod \"glance-default-internal-api-0\" (UID: \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:22:07 crc kubenswrapper[4700]: I0227 17:22:07.082445 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:22:07 crc kubenswrapper[4700]: I0227 17:22:07.083313 4700 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Feb 27 17:22:07 crc kubenswrapper[4700]: I0227 17:22:07.083332 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-logs\") pod \"glance-default-internal-api-0\" (UID: \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:22:07 crc kubenswrapper[4700]: I0227 17:22:07.091150 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:22:07 crc kubenswrapper[4700]: I0227 17:22:07.091533 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-scripts\") pod \"glance-default-internal-api-0\" (UID: \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:22:07 crc kubenswrapper[4700]: I0227 17:22:07.091642 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:22:07 crc kubenswrapper[4700]: I0227 17:22:07.092536 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-config-data\") pod \"glance-default-internal-api-0\" (UID: \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:22:07 crc kubenswrapper[4700]: I0227 17:22:07.111311 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zr8ct\" (UniqueName: \"kubernetes.io/projected/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-kube-api-access-zr8ct\") pod \"glance-default-internal-api-0\" (UID: \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:22:07 crc kubenswrapper[4700]: I0227 17:22:07.148655 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:22:07 crc kubenswrapper[4700]: I0227 17:22:07.272162 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 27 17:22:07 crc kubenswrapper[4700]: I0227 17:22:07.929080 4700 scope.go:117] "RemoveContainer" containerID="9560399dcda8d30788a5744b22b16ffe16678cbfd42a8ef1cf0f18a207bafcdf" Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.073403 4700 scope.go:117] "RemoveContainer" containerID="0c12db820cf315188ff119c527bebdfb4c167c6ab9335a7466d8af31acf213ae" Feb 27 17:22:08 crc kubenswrapper[4700]: E0227 17:22:08.074024 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c12db820cf315188ff119c527bebdfb4c167c6ab9335a7466d8af31acf213ae\": container with ID starting with 0c12db820cf315188ff119c527bebdfb4c167c6ab9335a7466d8af31acf213ae not found: ID does not exist" containerID="0c12db820cf315188ff119c527bebdfb4c167c6ab9335a7466d8af31acf213ae" Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.074083 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c12db820cf315188ff119c527bebdfb4c167c6ab9335a7466d8af31acf213ae"} err="failed to get container status \"0c12db820cf315188ff119c527bebdfb4c167c6ab9335a7466d8af31acf213ae\": rpc error: code = NotFound desc = could not find container \"0c12db820cf315188ff119c527bebdfb4c167c6ab9335a7466d8af31acf213ae\": container with ID starting with 0c12db820cf315188ff119c527bebdfb4c167c6ab9335a7466d8af31acf213ae not found: ID does not exist" Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.074138 4700 scope.go:117] "RemoveContainer" containerID="9560399dcda8d30788a5744b22b16ffe16678cbfd42a8ef1cf0f18a207bafcdf" Feb 27 17:22:08 crc kubenswrapper[4700]: E0227 17:22:08.074565 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9560399dcda8d30788a5744b22b16ffe16678cbfd42a8ef1cf0f18a207bafcdf\": container with ID starting with 9560399dcda8d30788a5744b22b16ffe16678cbfd42a8ef1cf0f18a207bafcdf not found: ID does not exist" containerID="9560399dcda8d30788a5744b22b16ffe16678cbfd42a8ef1cf0f18a207bafcdf" Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.074598 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9560399dcda8d30788a5744b22b16ffe16678cbfd42a8ef1cf0f18a207bafcdf"} err="failed to get container status \"9560399dcda8d30788a5744b22b16ffe16678cbfd42a8ef1cf0f18a207bafcdf\": rpc error: code = NotFound desc = could not find container \"9560399dcda8d30788a5744b22b16ffe16678cbfd42a8ef1cf0f18a207bafcdf\": container with ID starting with 9560399dcda8d30788a5744b22b16ffe16678cbfd42a8ef1cf0f18a207bafcdf not found: ID does not exist" Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.074618 4700 scope.go:117] "RemoveContainer" containerID="0c12db820cf315188ff119c527bebdfb4c167c6ab9335a7466d8af31acf213ae" Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.075067 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c12db820cf315188ff119c527bebdfb4c167c6ab9335a7466d8af31acf213ae"} err="failed to get container status \"0c12db820cf315188ff119c527bebdfb4c167c6ab9335a7466d8af31acf213ae\": rpc error: code = NotFound desc = could not find container \"0c12db820cf315188ff119c527bebdfb4c167c6ab9335a7466d8af31acf213ae\": container with ID starting with 0c12db820cf315188ff119c527bebdfb4c167c6ab9335a7466d8af31acf213ae not found: ID does not exist" Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.075110 4700 scope.go:117] "RemoveContainer" containerID="9560399dcda8d30788a5744b22b16ffe16678cbfd42a8ef1cf0f18a207bafcdf" Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.075623 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9560399dcda8d30788a5744b22b16ffe16678cbfd42a8ef1cf0f18a207bafcdf"} err="failed to get container status \"9560399dcda8d30788a5744b22b16ffe16678cbfd42a8ef1cf0f18a207bafcdf\": rpc error: code = NotFound desc = could not find container \"9560399dcda8d30788a5744b22b16ffe16678cbfd42a8ef1cf0f18a207bafcdf\": container with ID starting with 9560399dcda8d30788a5744b22b16ffe16678cbfd42a8ef1cf0f18a207bafcdf not found: ID does not exist" Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.075643 4700 scope.go:117] "RemoveContainer" containerID="8fbadff5661ea993281ad23be148279bc1f422c331d42e15f1e294ab8b80a273" Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.105796 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-6b5ck" Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.184267 4700 scope.go:117] "RemoveContainer" containerID="481cd96f6890e8291a0d0587ef75e75c787464a33a4382b61f8a8f3c9dd0148f" Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.204497 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3f209c04-79ea-4048-9d31-6222874de091-config\") pod \"3f209c04-79ea-4048-9d31-6222874de091\" (UID: \"3f209c04-79ea-4048-9d31-6222874de091\") " Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.204661 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f209c04-79ea-4048-9d31-6222874de091-combined-ca-bundle\") pod \"3f209c04-79ea-4048-9d31-6222874de091\" (UID: \"3f209c04-79ea-4048-9d31-6222874de091\") " Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.204693 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-glksm\" (UniqueName: \"kubernetes.io/projected/3f209c04-79ea-4048-9d31-6222874de091-kube-api-access-glksm\") pod \"3f209c04-79ea-4048-9d31-6222874de091\" (UID: \"3f209c04-79ea-4048-9d31-6222874de091\") " Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.237754 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f209c04-79ea-4048-9d31-6222874de091-kube-api-access-glksm" (OuterVolumeSpecName: "kube-api-access-glksm") pod "3f209c04-79ea-4048-9d31-6222874de091" (UID: "3f209c04-79ea-4048-9d31-6222874de091"). InnerVolumeSpecName "kube-api-access-glksm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.257342 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f209c04-79ea-4048-9d31-6222874de091-config" (OuterVolumeSpecName: "config") pod "3f209c04-79ea-4048-9d31-6222874de091" (UID: "3f209c04-79ea-4048-9d31-6222874de091"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.270688 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5b48777cdd-tqbkj" Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.270740 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5b48777cdd-tqbkj" Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.281004 4700 scope.go:117] "RemoveContainer" containerID="8fbadff5661ea993281ad23be148279bc1f422c331d42e15f1e294ab8b80a273" Feb 27 17:22:08 crc kubenswrapper[4700]: E0227 17:22:08.283154 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8fbadff5661ea993281ad23be148279bc1f422c331d42e15f1e294ab8b80a273\": container with ID starting with 8fbadff5661ea993281ad23be148279bc1f422c331d42e15f1e294ab8b80a273 not found: ID does not exist" containerID="8fbadff5661ea993281ad23be148279bc1f422c331d42e15f1e294ab8b80a273" Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.283184 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fbadff5661ea993281ad23be148279bc1f422c331d42e15f1e294ab8b80a273"} err="failed to get container status \"8fbadff5661ea993281ad23be148279bc1f422c331d42e15f1e294ab8b80a273\": rpc error: code = NotFound desc = could not find container \"8fbadff5661ea993281ad23be148279bc1f422c331d42e15f1e294ab8b80a273\": container with ID starting with 8fbadff5661ea993281ad23be148279bc1f422c331d42e15f1e294ab8b80a273 not found: ID does not exist" Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.283202 4700 scope.go:117] "RemoveContainer" containerID="481cd96f6890e8291a0d0587ef75e75c787464a33a4382b61f8a8f3c9dd0148f" Feb 27 17:22:08 crc kubenswrapper[4700]: E0227 17:22:08.285100 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"481cd96f6890e8291a0d0587ef75e75c787464a33a4382b61f8a8f3c9dd0148f\": container with ID starting with 481cd96f6890e8291a0d0587ef75e75c787464a33a4382b61f8a8f3c9dd0148f not found: ID does not exist" containerID="481cd96f6890e8291a0d0587ef75e75c787464a33a4382b61f8a8f3c9dd0148f" Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.285148 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"481cd96f6890e8291a0d0587ef75e75c787464a33a4382b61f8a8f3c9dd0148f"} err="failed to get container status \"481cd96f6890e8291a0d0587ef75e75c787464a33a4382b61f8a8f3c9dd0148f\": rpc error: code = NotFound desc = could not find container \"481cd96f6890e8291a0d0587ef75e75c787464a33a4382b61f8a8f3c9dd0148f\": container with ID starting with 481cd96f6890e8291a0d0587ef75e75c787464a33a4382b61f8a8f3c9dd0148f not found: ID does not exist" Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.285181 4700 scope.go:117] "RemoveContainer" containerID="8fbadff5661ea993281ad23be148279bc1f422c331d42e15f1e294ab8b80a273" Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.290991 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fbadff5661ea993281ad23be148279bc1f422c331d42e15f1e294ab8b80a273"} err="failed to get container status \"8fbadff5661ea993281ad23be148279bc1f422c331d42e15f1e294ab8b80a273\": rpc error: code = NotFound desc = could not find container \"8fbadff5661ea993281ad23be148279bc1f422c331d42e15f1e294ab8b80a273\": container with ID starting with 8fbadff5661ea993281ad23be148279bc1f422c331d42e15f1e294ab8b80a273 not found: ID does not exist" Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.291024 4700 scope.go:117] "RemoveContainer" containerID="481cd96f6890e8291a0d0587ef75e75c787464a33a4382b61f8a8f3c9dd0148f" Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.294527 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"481cd96f6890e8291a0d0587ef75e75c787464a33a4382b61f8a8f3c9dd0148f"} err="failed to get container status \"481cd96f6890e8291a0d0587ef75e75c787464a33a4382b61f8a8f3c9dd0148f\": rpc error: code = NotFound desc = could not find container \"481cd96f6890e8291a0d0587ef75e75c787464a33a4382b61f8a8f3c9dd0148f\": container with ID starting with 481cd96f6890e8291a0d0587ef75e75c787464a33a4382b61f8a8f3c9dd0148f not found: ID does not exist" Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.307547 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f209c04-79ea-4048-9d31-6222874de091-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3f209c04-79ea-4048-9d31-6222874de091" (UID: "3f209c04-79ea-4048-9d31-6222874de091"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.308706 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f209c04-79ea-4048-9d31-6222874de091-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.308719 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-glksm\" (UniqueName: \"kubernetes.io/projected/3f209c04-79ea-4048-9d31-6222874de091-kube-api-access-glksm\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.308729 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/3f209c04-79ea-4048-9d31-6222874de091-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.411533 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6cc494ff8-ggq9w" Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.411992 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6cc494ff8-ggq9w" Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.517645 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-6b5ck" event={"ID":"3f209c04-79ea-4048-9d31-6222874de091","Type":"ContainerDied","Data":"e717d818bfbb46c1a1c7498bbf3e72872eae0252d5af74b4ba64523538038f72"} Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.517682 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e717d818bfbb46c1a1c7498bbf3e72872eae0252d5af74b4ba64523538038f72" Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.517725 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-6b5ck" Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.529385 4700 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.668840 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 17:22:08 crc kubenswrapper[4700]: W0227 17:22:08.678009 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod345c36c6_a3e3_4ece_9ac8_d59a7b31b366.slice/crio-a98a8e42a97e9e2975b87c4917ec6b1475249cb171de127d2dc86371a5f57eda WatchSource:0}: Error finding container a98a8e42a97e9e2975b87c4917ec6b1475249cb171de127d2dc86371a5f57eda: Status 404 returned error can't find the container with id a98a8e42a97e9e2975b87c4917ec6b1475249cb171de127d2dc86371a5f57eda Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.763656 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.831444 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.876279 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-applier-0" Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.876312 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-applier-0" Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.890588 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-decision-engine-0" Feb 27 17:22:08 crc kubenswrapper[4700]: W0227 17:22:08.921403 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3a7b7812_6e5f_4c57_a8d1_507e3dc2217c.slice/crio-2db193f558ee0dfae7f8503c267a33ea0acab870e1ad92fbb4a0b47590f23bc3 WatchSource:0}: Error finding container 2db193f558ee0dfae7f8503c267a33ea0acab870e1ad92fbb4a0b47590f23bc3: Status 404 returned error can't find the container with id 2db193f558ee0dfae7f8503c267a33ea0acab870e1ad92fbb4a0b47590f23bc3 Feb 27 17:22:08 crc kubenswrapper[4700]: I0227 17:22:08.946730 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-applier-0" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.466679 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7744b9dc6c-qnsm2"] Feb 27 17:22:09 crc kubenswrapper[4700]: E0227 17:22:09.467905 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f209c04-79ea-4048-9d31-6222874de091" containerName="neutron-db-sync" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.467968 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f209c04-79ea-4048-9d31-6222874de091" containerName="neutron-db-sync" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.475057 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f209c04-79ea-4048-9d31-6222874de091" containerName="neutron-db-sync" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.476096 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7744b9dc6c-qnsm2" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.507396 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7744b9dc6c-qnsm2"] Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.619931 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"345c36c6-a3e3-4ece-9ac8-d59a7b31b366","Type":"ContainerStarted","Data":"a98a8e42a97e9e2975b87c4917ec6b1475249cb171de127d2dc86371a5f57eda"} Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.628844 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c","Type":"ContainerStarted","Data":"2db193f558ee0dfae7f8503c267a33ea0acab870e1ad92fbb4a0b47590f23bc3"} Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.635371 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536882-csprw" event={"ID":"58023608-7e65-4290-b24c-ea314a3fb15b","Type":"ContainerStarted","Data":"1b851f98ecd070cf7e19c017388d8ba7cc882c7f9ed67c744e2f4332334b168f"} Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.635424 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.661727 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc-config\") pod \"dnsmasq-dns-7744b9dc6c-qnsm2\" (UID: \"5ca33b95-1b7a-4d1c-a3b8-65a5969945bc\") " pod="openstack/dnsmasq-dns-7744b9dc6c-qnsm2" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.661793 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc-ovsdbserver-nb\") pod \"dnsmasq-dns-7744b9dc6c-qnsm2\" (UID: \"5ca33b95-1b7a-4d1c-a3b8-65a5969945bc\") " pod="openstack/dnsmasq-dns-7744b9dc6c-qnsm2" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.661846 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc-ovsdbserver-sb\") pod \"dnsmasq-dns-7744b9dc6c-qnsm2\" (UID: \"5ca33b95-1b7a-4d1c-a3b8-65a5969945bc\") " pod="openstack/dnsmasq-dns-7744b9dc6c-qnsm2" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.661880 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc-dns-swift-storage-0\") pod \"dnsmasq-dns-7744b9dc6c-qnsm2\" (UID: \"5ca33b95-1b7a-4d1c-a3b8-65a5969945bc\") " pod="openstack/dnsmasq-dns-7744b9dc6c-qnsm2" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.661944 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97lrn\" (UniqueName: \"kubernetes.io/projected/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc-kube-api-access-97lrn\") pod \"dnsmasq-dns-7744b9dc6c-qnsm2\" (UID: \"5ca33b95-1b7a-4d1c-a3b8-65a5969945bc\") " pod="openstack/dnsmasq-dns-7744b9dc6c-qnsm2" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.661964 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc-dns-svc\") pod \"dnsmasq-dns-7744b9dc6c-qnsm2\" (UID: \"5ca33b95-1b7a-4d1c-a3b8-65a5969945bc\") " pod="openstack/dnsmasq-dns-7744b9dc6c-qnsm2" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.699242 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7db57cbc8-t9p8l"] Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.705137 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7db57cbc8-t9p8l" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.710916 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.711069 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.711190 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-qxjr4" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.714020 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.720763 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536882-csprw" podStartSLOduration=8.505773633 podStartE2EDuration="9.720744262s" podCreationTimestamp="2026-02-27 17:22:00 +0000 UTC" firstStartedPulling="2026-02-27 17:22:04.461765351 +0000 UTC m=+1284.447078088" lastFinishedPulling="2026-02-27 17:22:05.67673596 +0000 UTC m=+1285.662048717" observedRunningTime="2026-02-27 17:22:09.679187002 +0000 UTC m=+1289.664499739" watchObservedRunningTime="2026-02-27 17:22:09.720744262 +0000 UTC m=+1289.706057009" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.742118 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7db57cbc8-t9p8l"] Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.762815 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/24cf75b6-81b2-43e8-b75e-c41171313a0f-config\") pod \"neutron-7db57cbc8-t9p8l\" (UID: \"24cf75b6-81b2-43e8-b75e-c41171313a0f\") " pod="openstack/neutron-7db57cbc8-t9p8l" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.762858 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97lrn\" (UniqueName: \"kubernetes.io/projected/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc-kube-api-access-97lrn\") pod \"dnsmasq-dns-7744b9dc6c-qnsm2\" (UID: \"5ca33b95-1b7a-4d1c-a3b8-65a5969945bc\") " pod="openstack/dnsmasq-dns-7744b9dc6c-qnsm2" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.762884 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc-dns-svc\") pod \"dnsmasq-dns-7744b9dc6c-qnsm2\" (UID: \"5ca33b95-1b7a-4d1c-a3b8-65a5969945bc\") " pod="openstack/dnsmasq-dns-7744b9dc6c-qnsm2" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.762906 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24cf75b6-81b2-43e8-b75e-c41171313a0f-combined-ca-bundle\") pod \"neutron-7db57cbc8-t9p8l\" (UID: \"24cf75b6-81b2-43e8-b75e-c41171313a0f\") " pod="openstack/neutron-7db57cbc8-t9p8l" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.762933 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmmqq\" (UniqueName: \"kubernetes.io/projected/24cf75b6-81b2-43e8-b75e-c41171313a0f-kube-api-access-xmmqq\") pod \"neutron-7db57cbc8-t9p8l\" (UID: \"24cf75b6-81b2-43e8-b75e-c41171313a0f\") " pod="openstack/neutron-7db57cbc8-t9p8l" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.762987 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/24cf75b6-81b2-43e8-b75e-c41171313a0f-httpd-config\") pod \"neutron-7db57cbc8-t9p8l\" (UID: \"24cf75b6-81b2-43e8-b75e-c41171313a0f\") " pod="openstack/neutron-7db57cbc8-t9p8l" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.763012 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc-config\") pod \"dnsmasq-dns-7744b9dc6c-qnsm2\" (UID: \"5ca33b95-1b7a-4d1c-a3b8-65a5969945bc\") " pod="openstack/dnsmasq-dns-7744b9dc6c-qnsm2" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.763033 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc-ovsdbserver-nb\") pod \"dnsmasq-dns-7744b9dc6c-qnsm2\" (UID: \"5ca33b95-1b7a-4d1c-a3b8-65a5969945bc\") " pod="openstack/dnsmasq-dns-7744b9dc6c-qnsm2" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.763050 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/24cf75b6-81b2-43e8-b75e-c41171313a0f-ovndb-tls-certs\") pod \"neutron-7db57cbc8-t9p8l\" (UID: \"24cf75b6-81b2-43e8-b75e-c41171313a0f\") " pod="openstack/neutron-7db57cbc8-t9p8l" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.763091 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc-ovsdbserver-sb\") pod \"dnsmasq-dns-7744b9dc6c-qnsm2\" (UID: \"5ca33b95-1b7a-4d1c-a3b8-65a5969945bc\") " pod="openstack/dnsmasq-dns-7744b9dc6c-qnsm2" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.763113 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc-dns-swift-storage-0\") pod \"dnsmasq-dns-7744b9dc6c-qnsm2\" (UID: \"5ca33b95-1b7a-4d1c-a3b8-65a5969945bc\") " pod="openstack/dnsmasq-dns-7744b9dc6c-qnsm2" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.770415 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc-dns-swift-storage-0\") pod \"dnsmasq-dns-7744b9dc6c-qnsm2\" (UID: \"5ca33b95-1b7a-4d1c-a3b8-65a5969945bc\") " pod="openstack/dnsmasq-dns-7744b9dc6c-qnsm2" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.771161 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc-dns-svc\") pod \"dnsmasq-dns-7744b9dc6c-qnsm2\" (UID: \"5ca33b95-1b7a-4d1c-a3b8-65a5969945bc\") " pod="openstack/dnsmasq-dns-7744b9dc6c-qnsm2" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.772289 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc-ovsdbserver-nb\") pod \"dnsmasq-dns-7744b9dc6c-qnsm2\" (UID: \"5ca33b95-1b7a-4d1c-a3b8-65a5969945bc\") " pod="openstack/dnsmasq-dns-7744b9dc6c-qnsm2" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.774156 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc-ovsdbserver-sb\") pod \"dnsmasq-dns-7744b9dc6c-qnsm2\" (UID: \"5ca33b95-1b7a-4d1c-a3b8-65a5969945bc\") " pod="openstack/dnsmasq-dns-7744b9dc6c-qnsm2" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.782988 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc-config\") pod \"dnsmasq-dns-7744b9dc6c-qnsm2\" (UID: \"5ca33b95-1b7a-4d1c-a3b8-65a5969945bc\") " pod="openstack/dnsmasq-dns-7744b9dc6c-qnsm2" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.787730 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-applier-0" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.847442 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97lrn\" (UniqueName: \"kubernetes.io/projected/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc-kube-api-access-97lrn\") pod \"dnsmasq-dns-7744b9dc6c-qnsm2\" (UID: \"5ca33b95-1b7a-4d1c-a3b8-65a5969945bc\") " pod="openstack/dnsmasq-dns-7744b9dc6c-qnsm2" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.847877 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7744b9dc6c-qnsm2" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.864288 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/24cf75b6-81b2-43e8-b75e-c41171313a0f-config\") pod \"neutron-7db57cbc8-t9p8l\" (UID: \"24cf75b6-81b2-43e8-b75e-c41171313a0f\") " pod="openstack/neutron-7db57cbc8-t9p8l" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.864344 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24cf75b6-81b2-43e8-b75e-c41171313a0f-combined-ca-bundle\") pod \"neutron-7db57cbc8-t9p8l\" (UID: \"24cf75b6-81b2-43e8-b75e-c41171313a0f\") " pod="openstack/neutron-7db57cbc8-t9p8l" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.864370 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmmqq\" (UniqueName: \"kubernetes.io/projected/24cf75b6-81b2-43e8-b75e-c41171313a0f-kube-api-access-xmmqq\") pod \"neutron-7db57cbc8-t9p8l\" (UID: \"24cf75b6-81b2-43e8-b75e-c41171313a0f\") " pod="openstack/neutron-7db57cbc8-t9p8l" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.864425 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/24cf75b6-81b2-43e8-b75e-c41171313a0f-httpd-config\") pod \"neutron-7db57cbc8-t9p8l\" (UID: \"24cf75b6-81b2-43e8-b75e-c41171313a0f\") " pod="openstack/neutron-7db57cbc8-t9p8l" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.864453 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/24cf75b6-81b2-43e8-b75e-c41171313a0f-ovndb-tls-certs\") pod \"neutron-7db57cbc8-t9p8l\" (UID: \"24cf75b6-81b2-43e8-b75e-c41171313a0f\") " pod="openstack/neutron-7db57cbc8-t9p8l" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.869709 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-decision-engine-0" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.873293 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-applier-0"] Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.874983 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/24cf75b6-81b2-43e8-b75e-c41171313a0f-ovndb-tls-certs\") pod \"neutron-7db57cbc8-t9p8l\" (UID: \"24cf75b6-81b2-43e8-b75e-c41171313a0f\") " pod="openstack/neutron-7db57cbc8-t9p8l" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.879148 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/24cf75b6-81b2-43e8-b75e-c41171313a0f-httpd-config\") pod \"neutron-7db57cbc8-t9p8l\" (UID: \"24cf75b6-81b2-43e8-b75e-c41171313a0f\") " pod="openstack/neutron-7db57cbc8-t9p8l" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.888443 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24cf75b6-81b2-43e8-b75e-c41171313a0f-combined-ca-bundle\") pod \"neutron-7db57cbc8-t9p8l\" (UID: \"24cf75b6-81b2-43e8-b75e-c41171313a0f\") " pod="openstack/neutron-7db57cbc8-t9p8l" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.896721 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/24cf75b6-81b2-43e8-b75e-c41171313a0f-config\") pod \"neutron-7db57cbc8-t9p8l\" (UID: \"24cf75b6-81b2-43e8-b75e-c41171313a0f\") " pod="openstack/neutron-7db57cbc8-t9p8l" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.899991 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmmqq\" (UniqueName: \"kubernetes.io/projected/24cf75b6-81b2-43e8-b75e-c41171313a0f-kube-api-access-xmmqq\") pod \"neutron-7db57cbc8-t9p8l\" (UID: \"24cf75b6-81b2-43e8-b75e-c41171313a0f\") " pod="openstack/neutron-7db57cbc8-t9p8l" Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.925798 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-decision-engine-0"] Feb 27 17:22:09 crc kubenswrapper[4700]: I0227 17:22:09.982194 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Feb 27 17:22:10 crc kubenswrapper[4700]: I0227 17:22:10.071790 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7db57cbc8-t9p8l" Feb 27 17:22:10 crc kubenswrapper[4700]: I0227 17:22:10.644358 4700 generic.go:334] "Generic (PLEG): container finished" podID="58023608-7e65-4290-b24c-ea314a3fb15b" containerID="1b851f98ecd070cf7e19c017388d8ba7cc882c7f9ed67c744e2f4332334b168f" exitCode=0 Feb 27 17:22:10 crc kubenswrapper[4700]: I0227 17:22:10.644447 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536882-csprw" event={"ID":"58023608-7e65-4290-b24c-ea314a3fb15b","Type":"ContainerDied","Data":"1b851f98ecd070cf7e19c017388d8ba7cc882c7f9ed67c744e2f4332334b168f"} Feb 27 17:22:10 crc kubenswrapper[4700]: I0227 17:22:10.654795 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"345c36c6-a3e3-4ece-9ac8-d59a7b31b366","Type":"ContainerStarted","Data":"3796d0ad84376c90605ec60063d83fa92dd2c330c228af0dd64e6141ad2e7bd0"} Feb 27 17:22:10 crc kubenswrapper[4700]: I0227 17:22:10.657341 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c","Type":"ContainerStarted","Data":"71cbecfa6096e107490dd03a4e89a8a7ddf94f1f8f4c66dccfeb084fb8fb85c9"} Feb 27 17:22:10 crc kubenswrapper[4700]: E0227 17:22:10.860244 4700 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod84901b1e_0055_4489_8009_711c1676e51f.slice/crio-3caf20a4365af9022601cc805ec79dcb3f0bc006f30ea14208f9a3981739664e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod84901b1e_0055_4489_8009_711c1676e51f.slice/crio-conmon-3caf20a4365af9022601cc805ec79dcb3f0bc006f30ea14208f9a3981739664e.scope\": RecentStats: unable to find data in memory cache]" Feb 27 17:22:11 crc kubenswrapper[4700]: I0227 17:22:11.531640 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-api-0" Feb 27 17:22:11 crc kubenswrapper[4700]: I0227 17:22:11.554006 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-api-0" Feb 27 17:22:11 crc kubenswrapper[4700]: I0227 17:22:11.668124 4700 generic.go:334] "Generic (PLEG): container finished" podID="84901b1e-0055-4489-8009-711c1676e51f" containerID="3caf20a4365af9022601cc805ec79dcb3f0bc006f30ea14208f9a3981739664e" exitCode=0 Feb 27 17:22:11 crc kubenswrapper[4700]: I0227 17:22:11.669110 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-applier-0" podUID="4099c554-cf42-48b3-88cb-e672a66f1a05" containerName="watcher-applier" containerID="cri-o://790d64d806f8cf030b5fa73b8846ad67dd2fb782f8fff50bd6090997023776d8" gracePeriod=30 Feb 27 17:22:11 crc kubenswrapper[4700]: I0227 17:22:11.669946 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8xqhb" event={"ID":"84901b1e-0055-4489-8009-711c1676e51f","Type":"ContainerDied","Data":"3caf20a4365af9022601cc805ec79dcb3f0bc006f30ea14208f9a3981739664e"} Feb 27 17:22:11 crc kubenswrapper[4700]: I0227 17:22:11.672943 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-decision-engine-0" podUID="d8e03836-cf5d-4ac9-80fe-97556117e65e" containerName="watcher-decision-engine" containerID="cri-o://0f536762f0202ca79c315c2b9f19d46b59472ea624b34e548af06e9322bb143e" gracePeriod=30 Feb 27 17:22:11 crc kubenswrapper[4700]: I0227 17:22:11.683966 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Feb 27 17:22:11 crc kubenswrapper[4700]: I0227 17:22:11.874765 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-8946ccf97-n4k7p"] Feb 27 17:22:11 crc kubenswrapper[4700]: I0227 17:22:11.898292 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8946ccf97-n4k7p"] Feb 27 17:22:11 crc kubenswrapper[4700]: I0227 17:22:11.898408 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8946ccf97-n4k7p" Feb 27 17:22:11 crc kubenswrapper[4700]: I0227 17:22:11.903502 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Feb 27 17:22:11 crc kubenswrapper[4700]: I0227 17:22:11.903736 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Feb 27 17:22:12 crc kubenswrapper[4700]: I0227 17:22:12.004799 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2szd\" (UniqueName: \"kubernetes.io/projected/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-kube-api-access-j2szd\") pod \"neutron-8946ccf97-n4k7p\" (UID: \"f0d00b91-0e33-4466-8c9f-6a1c1038d23f\") " pod="openstack/neutron-8946ccf97-n4k7p" Feb 27 17:22:12 crc kubenswrapper[4700]: I0227 17:22:12.004858 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-config\") pod \"neutron-8946ccf97-n4k7p\" (UID: \"f0d00b91-0e33-4466-8c9f-6a1c1038d23f\") " pod="openstack/neutron-8946ccf97-n4k7p" Feb 27 17:22:12 crc kubenswrapper[4700]: I0227 17:22:12.004920 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-combined-ca-bundle\") pod \"neutron-8946ccf97-n4k7p\" (UID: \"f0d00b91-0e33-4466-8c9f-6a1c1038d23f\") " pod="openstack/neutron-8946ccf97-n4k7p" Feb 27 17:22:12 crc kubenswrapper[4700]: I0227 17:22:12.004938 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-public-tls-certs\") pod \"neutron-8946ccf97-n4k7p\" (UID: \"f0d00b91-0e33-4466-8c9f-6a1c1038d23f\") " pod="openstack/neutron-8946ccf97-n4k7p" Feb 27 17:22:12 crc kubenswrapper[4700]: I0227 17:22:12.004954 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-internal-tls-certs\") pod \"neutron-8946ccf97-n4k7p\" (UID: \"f0d00b91-0e33-4466-8c9f-6a1c1038d23f\") " pod="openstack/neutron-8946ccf97-n4k7p" Feb 27 17:22:12 crc kubenswrapper[4700]: I0227 17:22:12.004968 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-ovndb-tls-certs\") pod \"neutron-8946ccf97-n4k7p\" (UID: \"f0d00b91-0e33-4466-8c9f-6a1c1038d23f\") " pod="openstack/neutron-8946ccf97-n4k7p" Feb 27 17:22:12 crc kubenswrapper[4700]: I0227 17:22:12.005013 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-httpd-config\") pod \"neutron-8946ccf97-n4k7p\" (UID: \"f0d00b91-0e33-4466-8c9f-6a1c1038d23f\") " pod="openstack/neutron-8946ccf97-n4k7p" Feb 27 17:22:12 crc kubenswrapper[4700]: I0227 17:22:12.106564 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2szd\" (UniqueName: \"kubernetes.io/projected/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-kube-api-access-j2szd\") pod \"neutron-8946ccf97-n4k7p\" (UID: \"f0d00b91-0e33-4466-8c9f-6a1c1038d23f\") " pod="openstack/neutron-8946ccf97-n4k7p" Feb 27 17:22:12 crc kubenswrapper[4700]: I0227 17:22:12.107089 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-config\") pod \"neutron-8946ccf97-n4k7p\" (UID: \"f0d00b91-0e33-4466-8c9f-6a1c1038d23f\") " pod="openstack/neutron-8946ccf97-n4k7p" Feb 27 17:22:12 crc kubenswrapper[4700]: I0227 17:22:12.107165 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-combined-ca-bundle\") pod \"neutron-8946ccf97-n4k7p\" (UID: \"f0d00b91-0e33-4466-8c9f-6a1c1038d23f\") " pod="openstack/neutron-8946ccf97-n4k7p" Feb 27 17:22:12 crc kubenswrapper[4700]: I0227 17:22:12.107184 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-public-tls-certs\") pod \"neutron-8946ccf97-n4k7p\" (UID: \"f0d00b91-0e33-4466-8c9f-6a1c1038d23f\") " pod="openstack/neutron-8946ccf97-n4k7p" Feb 27 17:22:12 crc kubenswrapper[4700]: I0227 17:22:12.107225 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-internal-tls-certs\") pod \"neutron-8946ccf97-n4k7p\" (UID: \"f0d00b91-0e33-4466-8c9f-6a1c1038d23f\") " pod="openstack/neutron-8946ccf97-n4k7p" Feb 27 17:22:12 crc kubenswrapper[4700]: I0227 17:22:12.107243 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-ovndb-tls-certs\") pod \"neutron-8946ccf97-n4k7p\" (UID: \"f0d00b91-0e33-4466-8c9f-6a1c1038d23f\") " pod="openstack/neutron-8946ccf97-n4k7p" Feb 27 17:22:12 crc kubenswrapper[4700]: I0227 17:22:12.107337 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-httpd-config\") pod \"neutron-8946ccf97-n4k7p\" (UID: \"f0d00b91-0e33-4466-8c9f-6a1c1038d23f\") " pod="openstack/neutron-8946ccf97-n4k7p" Feb 27 17:22:12 crc kubenswrapper[4700]: I0227 17:22:12.149615 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-combined-ca-bundle\") pod \"neutron-8946ccf97-n4k7p\" (UID: \"f0d00b91-0e33-4466-8c9f-6a1c1038d23f\") " pod="openstack/neutron-8946ccf97-n4k7p" Feb 27 17:22:12 crc kubenswrapper[4700]: I0227 17:22:12.151189 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2szd\" (UniqueName: \"kubernetes.io/projected/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-kube-api-access-j2szd\") pod \"neutron-8946ccf97-n4k7p\" (UID: \"f0d00b91-0e33-4466-8c9f-6a1c1038d23f\") " pod="openstack/neutron-8946ccf97-n4k7p" Feb 27 17:22:12 crc kubenswrapper[4700]: I0227 17:22:12.151991 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-httpd-config\") pod \"neutron-8946ccf97-n4k7p\" (UID: \"f0d00b91-0e33-4466-8c9f-6a1c1038d23f\") " pod="openstack/neutron-8946ccf97-n4k7p" Feb 27 17:22:12 crc kubenswrapper[4700]: I0227 17:22:12.164364 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-internal-tls-certs\") pod \"neutron-8946ccf97-n4k7p\" (UID: \"f0d00b91-0e33-4466-8c9f-6a1c1038d23f\") " pod="openstack/neutron-8946ccf97-n4k7p" Feb 27 17:22:12 crc kubenswrapper[4700]: I0227 17:22:12.165113 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-public-tls-certs\") pod \"neutron-8946ccf97-n4k7p\" (UID: \"f0d00b91-0e33-4466-8c9f-6a1c1038d23f\") " pod="openstack/neutron-8946ccf97-n4k7p" Feb 27 17:22:12 crc kubenswrapper[4700]: I0227 17:22:12.167531 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-ovndb-tls-certs\") pod \"neutron-8946ccf97-n4k7p\" (UID: \"f0d00b91-0e33-4466-8c9f-6a1c1038d23f\") " pod="openstack/neutron-8946ccf97-n4k7p" Feb 27 17:22:12 crc kubenswrapper[4700]: I0227 17:22:12.176235 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-config\") pod \"neutron-8946ccf97-n4k7p\" (UID: \"f0d00b91-0e33-4466-8c9f-6a1c1038d23f\") " pod="openstack/neutron-8946ccf97-n4k7p" Feb 27 17:22:12 crc kubenswrapper[4700]: I0227 17:22:12.252965 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8946ccf97-n4k7p" Feb 27 17:22:13 crc kubenswrapper[4700]: I0227 17:22:13.687014 4700 generic.go:334] "Generic (PLEG): container finished" podID="4099c554-cf42-48b3-88cb-e672a66f1a05" containerID="790d64d806f8cf030b5fa73b8846ad67dd2fb782f8fff50bd6090997023776d8" exitCode=0 Feb 27 17:22:13 crc kubenswrapper[4700]: I0227 17:22:13.687329 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"4099c554-cf42-48b3-88cb-e672a66f1a05","Type":"ContainerDied","Data":"790d64d806f8cf030b5fa73b8846ad67dd2fb782f8fff50bd6090997023776d8"} Feb 27 17:22:13 crc kubenswrapper[4700]: I0227 17:22:13.690155 4700 generic.go:334] "Generic (PLEG): container finished" podID="d8e03836-cf5d-4ac9-80fe-97556117e65e" containerID="0f536762f0202ca79c315c2b9f19d46b59472ea624b34e548af06e9322bb143e" exitCode=1 Feb 27 17:22:13 crc kubenswrapper[4700]: I0227 17:22:13.690174 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"d8e03836-cf5d-4ac9-80fe-97556117e65e","Type":"ContainerDied","Data":"0f536762f0202ca79c315c2b9f19d46b59472ea624b34e548af06e9322bb143e"} Feb 27 17:22:13 crc kubenswrapper[4700]: E0227 17:22:13.877357 4700 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 790d64d806f8cf030b5fa73b8846ad67dd2fb782f8fff50bd6090997023776d8 is running failed: container process not found" containerID="790d64d806f8cf030b5fa73b8846ad67dd2fb782f8fff50bd6090997023776d8" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Feb 27 17:22:13 crc kubenswrapper[4700]: E0227 17:22:13.877654 4700 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 790d64d806f8cf030b5fa73b8846ad67dd2fb782f8fff50bd6090997023776d8 is running failed: container process not found" containerID="790d64d806f8cf030b5fa73b8846ad67dd2fb782f8fff50bd6090997023776d8" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Feb 27 17:22:13 crc kubenswrapper[4700]: E0227 17:22:13.877894 4700 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 790d64d806f8cf030b5fa73b8846ad67dd2fb782f8fff50bd6090997023776d8 is running failed: container process not found" containerID="790d64d806f8cf030b5fa73b8846ad67dd2fb782f8fff50bd6090997023776d8" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Feb 27 17:22:13 crc kubenswrapper[4700]: E0227 17:22:13.877925 4700 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 790d64d806f8cf030b5fa73b8846ad67dd2fb782f8fff50bd6090997023776d8 is running failed: container process not found" probeType="Readiness" pod="openstack/watcher-applier-0" podUID="4099c554-cf42-48b3-88cb-e672a66f1a05" containerName="watcher-applier" Feb 27 17:22:14 crc kubenswrapper[4700]: I0227 17:22:14.701161 4700 generic.go:334] "Generic (PLEG): container finished" podID="989eacab-9c48-4375-b7a0-f2c7f40d17ca" containerID="4bcb91689f3aad19cad5a2592cf44a074600f816d8f0720250702e244a8da0db" exitCode=0 Feb 27 17:22:14 crc kubenswrapper[4700]: I0227 17:22:14.701208 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dmf4m" event={"ID":"989eacab-9c48-4375-b7a0-f2c7f40d17ca","Type":"ContainerDied","Data":"4bcb91689f3aad19cad5a2592cf44a074600f816d8f0720250702e244a8da0db"} Feb 27 17:22:15 crc kubenswrapper[4700]: I0227 17:22:15.597656 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Feb 27 17:22:15 crc kubenswrapper[4700]: I0227 17:22:15.598487 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="9822c457-a600-4d01-9b94-b7b899ef2102" containerName="watcher-api-log" containerID="cri-o://4593dfb0b9741e168873d46911f0902e7eb87fedecd51d7a9be60ebce9628d31" gracePeriod=30 Feb 27 17:22:15 crc kubenswrapper[4700]: I0227 17:22:15.599072 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="9822c457-a600-4d01-9b94-b7b899ef2102" containerName="watcher-api" containerID="cri-o://68026171c113261075b633116718884cfbc6f7e2bc616208c246fd9b332e8dab" gracePeriod=30 Feb 27 17:22:15 crc kubenswrapper[4700]: I0227 17:22:15.729384 4700 generic.go:334] "Generic (PLEG): container finished" podID="f72833f0-cf5f-4bf8-bb5f-1f65477c4b76" containerID="0f2f183bfd914f6252dd78997763889b1e5cc74a29e4d9af31c6456aceb88f96" exitCode=0 Feb 27 17:22:15 crc kubenswrapper[4700]: I0227 17:22:15.729481 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-vrb9h" event={"ID":"f72833f0-cf5f-4bf8-bb5f-1f65477c4b76","Type":"ContainerDied","Data":"0f2f183bfd914f6252dd78997763889b1e5cc74a29e4d9af31c6456aceb88f96"} Feb 27 17:22:15 crc kubenswrapper[4700]: I0227 17:22:15.996598 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8xqhb" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.046015 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536882-csprw" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.153285 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84901b1e-0055-4489-8009-711c1676e51f-combined-ca-bundle\") pod \"84901b1e-0055-4489-8009-711c1676e51f\" (UID: \"84901b1e-0055-4489-8009-711c1676e51f\") " Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.153686 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/84901b1e-0055-4489-8009-711c1676e51f-credential-keys\") pod \"84901b1e-0055-4489-8009-711c1676e51f\" (UID: \"84901b1e-0055-4489-8009-711c1676e51f\") " Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.153703 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84901b1e-0055-4489-8009-711c1676e51f-config-data\") pod \"84901b1e-0055-4489-8009-711c1676e51f\" (UID: \"84901b1e-0055-4489-8009-711c1676e51f\") " Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.153757 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbklk\" (UniqueName: \"kubernetes.io/projected/84901b1e-0055-4489-8009-711c1676e51f-kube-api-access-qbklk\") pod \"84901b1e-0055-4489-8009-711c1676e51f\" (UID: \"84901b1e-0055-4489-8009-711c1676e51f\") " Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.153783 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwd6w\" (UniqueName: \"kubernetes.io/projected/58023608-7e65-4290-b24c-ea314a3fb15b-kube-api-access-lwd6w\") pod \"58023608-7e65-4290-b24c-ea314a3fb15b\" (UID: \"58023608-7e65-4290-b24c-ea314a3fb15b\") " Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.153851 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/84901b1e-0055-4489-8009-711c1676e51f-fernet-keys\") pod \"84901b1e-0055-4489-8009-711c1676e51f\" (UID: \"84901b1e-0055-4489-8009-711c1676e51f\") " Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.153875 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84901b1e-0055-4489-8009-711c1676e51f-scripts\") pod \"84901b1e-0055-4489-8009-711c1676e51f\" (UID: \"84901b1e-0055-4489-8009-711c1676e51f\") " Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.203565 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84901b1e-0055-4489-8009-711c1676e51f-kube-api-access-qbklk" (OuterVolumeSpecName: "kube-api-access-qbklk") pod "84901b1e-0055-4489-8009-711c1676e51f" (UID: "84901b1e-0055-4489-8009-711c1676e51f"). InnerVolumeSpecName "kube-api-access-qbklk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.206924 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58023608-7e65-4290-b24c-ea314a3fb15b-kube-api-access-lwd6w" (OuterVolumeSpecName: "kube-api-access-lwd6w") pod "58023608-7e65-4290-b24c-ea314a3fb15b" (UID: "58023608-7e65-4290-b24c-ea314a3fb15b"). InnerVolumeSpecName "kube-api-access-lwd6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.207013 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84901b1e-0055-4489-8009-711c1676e51f-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "84901b1e-0055-4489-8009-711c1676e51f" (UID: "84901b1e-0055-4489-8009-711c1676e51f"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.210772 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84901b1e-0055-4489-8009-711c1676e51f-scripts" (OuterVolumeSpecName: "scripts") pod "84901b1e-0055-4489-8009-711c1676e51f" (UID: "84901b1e-0055-4489-8009-711c1676e51f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.223728 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84901b1e-0055-4489-8009-711c1676e51f-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "84901b1e-0055-4489-8009-711c1676e51f" (UID: "84901b1e-0055-4489-8009-711c1676e51f"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.231274 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84901b1e-0055-4489-8009-711c1676e51f-config-data" (OuterVolumeSpecName: "config-data") pod "84901b1e-0055-4489-8009-711c1676e51f" (UID: "84901b1e-0055-4489-8009-711c1676e51f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.260844 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84901b1e-0055-4489-8009-711c1676e51f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "84901b1e-0055-4489-8009-711c1676e51f" (UID: "84901b1e-0055-4489-8009-711c1676e51f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.267966 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbklk\" (UniqueName: \"kubernetes.io/projected/84901b1e-0055-4489-8009-711c1676e51f-kube-api-access-qbklk\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.267996 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwd6w\" (UniqueName: \"kubernetes.io/projected/58023608-7e65-4290-b24c-ea314a3fb15b-kube-api-access-lwd6w\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.268007 4700 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/84901b1e-0055-4489-8009-711c1676e51f-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.268019 4700 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84901b1e-0055-4489-8009-711c1676e51f-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.268029 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84901b1e-0055-4489-8009-711c1676e51f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.268037 4700 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/84901b1e-0055-4489-8009-711c1676e51f-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.268046 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84901b1e-0055-4489-8009-711c1676e51f-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.343509 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.373743 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dmf4m" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.385526 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.388552 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4099c554-cf42-48b3-88cb-e672a66f1a05-logs\") pod \"4099c554-cf42-48b3-88cb-e672a66f1a05\" (UID: \"4099c554-cf42-48b3-88cb-e672a66f1a05\") " Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.388592 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4099c554-cf42-48b3-88cb-e672a66f1a05-combined-ca-bundle\") pod \"4099c554-cf42-48b3-88cb-e672a66f1a05\" (UID: \"4099c554-cf42-48b3-88cb-e672a66f1a05\") " Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.388617 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcz7f\" (UniqueName: \"kubernetes.io/projected/4099c554-cf42-48b3-88cb-e672a66f1a05-kube-api-access-fcz7f\") pod \"4099c554-cf42-48b3-88cb-e672a66f1a05\" (UID: \"4099c554-cf42-48b3-88cb-e672a66f1a05\") " Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.388638 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4099c554-cf42-48b3-88cb-e672a66f1a05-config-data\") pod \"4099c554-cf42-48b3-88cb-e672a66f1a05\" (UID: \"4099c554-cf42-48b3-88cb-e672a66f1a05\") " Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.388682 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/989eacab-9c48-4375-b7a0-f2c7f40d17ca-logs\") pod \"989eacab-9c48-4375-b7a0-f2c7f40d17ca\" (UID: \"989eacab-9c48-4375-b7a0-f2c7f40d17ca\") " Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.388790 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/989eacab-9c48-4375-b7a0-f2c7f40d17ca-scripts\") pod \"989eacab-9c48-4375-b7a0-f2c7f40d17ca\" (UID: \"989eacab-9c48-4375-b7a0-f2c7f40d17ca\") " Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.388813 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/989eacab-9c48-4375-b7a0-f2c7f40d17ca-combined-ca-bundle\") pod \"989eacab-9c48-4375-b7a0-f2c7f40d17ca\" (UID: \"989eacab-9c48-4375-b7a0-f2c7f40d17ca\") " Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.388884 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/989eacab-9c48-4375-b7a0-f2c7f40d17ca-config-data\") pod \"989eacab-9c48-4375-b7a0-f2c7f40d17ca\" (UID: \"989eacab-9c48-4375-b7a0-f2c7f40d17ca\") " Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.388982 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkqvc\" (UniqueName: \"kubernetes.io/projected/989eacab-9c48-4375-b7a0-f2c7f40d17ca-kube-api-access-nkqvc\") pod \"989eacab-9c48-4375-b7a0-f2c7f40d17ca\" (UID: \"989eacab-9c48-4375-b7a0-f2c7f40d17ca\") " Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.401474 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/989eacab-9c48-4375-b7a0-f2c7f40d17ca-logs" (OuterVolumeSpecName: "logs") pod "989eacab-9c48-4375-b7a0-f2c7f40d17ca" (UID: "989eacab-9c48-4375-b7a0-f2c7f40d17ca"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.406902 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4099c554-cf42-48b3-88cb-e672a66f1a05-kube-api-access-fcz7f" (OuterVolumeSpecName: "kube-api-access-fcz7f") pod "4099c554-cf42-48b3-88cb-e672a66f1a05" (UID: "4099c554-cf42-48b3-88cb-e672a66f1a05"). InnerVolumeSpecName "kube-api-access-fcz7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.411539 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/989eacab-9c48-4375-b7a0-f2c7f40d17ca-kube-api-access-nkqvc" (OuterVolumeSpecName: "kube-api-access-nkqvc") pod "989eacab-9c48-4375-b7a0-f2c7f40d17ca" (UID: "989eacab-9c48-4375-b7a0-f2c7f40d17ca"). InnerVolumeSpecName "kube-api-access-nkqvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.414668 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4099c554-cf42-48b3-88cb-e672a66f1a05-logs" (OuterVolumeSpecName: "logs") pod "4099c554-cf42-48b3-88cb-e672a66f1a05" (UID: "4099c554-cf42-48b3-88cb-e672a66f1a05"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.415132 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/989eacab-9c48-4375-b7a0-f2c7f40d17ca-scripts" (OuterVolumeSpecName: "scripts") pod "989eacab-9c48-4375-b7a0-f2c7f40d17ca" (UID: "989eacab-9c48-4375-b7a0-f2c7f40d17ca"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.467570 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/989eacab-9c48-4375-b7a0-f2c7f40d17ca-config-data" (OuterVolumeSpecName: "config-data") pod "989eacab-9c48-4375-b7a0-f2c7f40d17ca" (UID: "989eacab-9c48-4375-b7a0-f2c7f40d17ca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.494401 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/989eacab-9c48-4375-b7a0-f2c7f40d17ca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "989eacab-9c48-4375-b7a0-f2c7f40d17ca" (UID: "989eacab-9c48-4375-b7a0-f2c7f40d17ca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.495202 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkqvc\" (UniqueName: \"kubernetes.io/projected/989eacab-9c48-4375-b7a0-f2c7f40d17ca-kube-api-access-nkqvc\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.495289 4700 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4099c554-cf42-48b3-88cb-e672a66f1a05-logs\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.495309 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcz7f\" (UniqueName: \"kubernetes.io/projected/4099c554-cf42-48b3-88cb-e672a66f1a05-kube-api-access-fcz7f\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.495350 4700 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/989eacab-9c48-4375-b7a0-f2c7f40d17ca-logs\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.495369 4700 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/989eacab-9c48-4375-b7a0-f2c7f40d17ca-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.495401 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/989eacab-9c48-4375-b7a0-f2c7f40d17ca-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.495890 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4099c554-cf42-48b3-88cb-e672a66f1a05-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4099c554-cf42-48b3-88cb-e672a66f1a05" (UID: "4099c554-cf42-48b3-88cb-e672a66f1a05"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.506691 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4099c554-cf42-48b3-88cb-e672a66f1a05-config-data" (OuterVolumeSpecName: "config-data") pod "4099c554-cf42-48b3-88cb-e672a66f1a05" (UID: "4099c554-cf42-48b3-88cb-e672a66f1a05"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.531254 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="9822c457-a600-4d01-9b94-b7b899ef2102" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.171:9322/\": dial tcp 10.217.0.171:9322: connect: connection refused" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.531309 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="9822c457-a600-4d01-9b94-b7b899ef2102" containerName="watcher-api-log" probeResult="failure" output="Get \"http://10.217.0.171:9322/\": dial tcp 10.217.0.171:9322: connect: connection refused" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.597043 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gnlv\" (UniqueName: \"kubernetes.io/projected/d8e03836-cf5d-4ac9-80fe-97556117e65e-kube-api-access-6gnlv\") pod \"d8e03836-cf5d-4ac9-80fe-97556117e65e\" (UID: \"d8e03836-cf5d-4ac9-80fe-97556117e65e\") " Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.597114 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8e03836-cf5d-4ac9-80fe-97556117e65e-logs\") pod \"d8e03836-cf5d-4ac9-80fe-97556117e65e\" (UID: \"d8e03836-cf5d-4ac9-80fe-97556117e65e\") " Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.597284 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8e03836-cf5d-4ac9-80fe-97556117e65e-config-data\") pod \"d8e03836-cf5d-4ac9-80fe-97556117e65e\" (UID: \"d8e03836-cf5d-4ac9-80fe-97556117e65e\") " Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.597372 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8e03836-cf5d-4ac9-80fe-97556117e65e-combined-ca-bundle\") pod \"d8e03836-cf5d-4ac9-80fe-97556117e65e\" (UID: \"d8e03836-cf5d-4ac9-80fe-97556117e65e\") " Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.597442 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/d8e03836-cf5d-4ac9-80fe-97556117e65e-custom-prometheus-ca\") pod \"d8e03836-cf5d-4ac9-80fe-97556117e65e\" (UID: \"d8e03836-cf5d-4ac9-80fe-97556117e65e\") " Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.597595 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8e03836-cf5d-4ac9-80fe-97556117e65e-logs" (OuterVolumeSpecName: "logs") pod "d8e03836-cf5d-4ac9-80fe-97556117e65e" (UID: "d8e03836-cf5d-4ac9-80fe-97556117e65e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.597889 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4099c554-cf42-48b3-88cb-e672a66f1a05-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.597907 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4099c554-cf42-48b3-88cb-e672a66f1a05-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.597917 4700 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8e03836-cf5d-4ac9-80fe-97556117e65e-logs\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.597925 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/989eacab-9c48-4375-b7a0-f2c7f40d17ca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.602417 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8e03836-cf5d-4ac9-80fe-97556117e65e-kube-api-access-6gnlv" (OuterVolumeSpecName: "kube-api-access-6gnlv") pod "d8e03836-cf5d-4ac9-80fe-97556117e65e" (UID: "d8e03836-cf5d-4ac9-80fe-97556117e65e"). InnerVolumeSpecName "kube-api-access-6gnlv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.657699 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8e03836-cf5d-4ac9-80fe-97556117e65e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d8e03836-cf5d-4ac9-80fe-97556117e65e" (UID: "d8e03836-cf5d-4ac9-80fe-97556117e65e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.670716 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8e03836-cf5d-4ac9-80fe-97556117e65e-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "d8e03836-cf5d-4ac9-80fe-97556117e65e" (UID: "d8e03836-cf5d-4ac9-80fe-97556117e65e"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.682139 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7744b9dc6c-qnsm2"] Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.687847 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8e03836-cf5d-4ac9-80fe-97556117e65e-config-data" (OuterVolumeSpecName: "config-data") pod "d8e03836-cf5d-4ac9-80fe-97556117e65e" (UID: "d8e03836-cf5d-4ac9-80fe-97556117e65e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.702002 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6gnlv\" (UniqueName: \"kubernetes.io/projected/d8e03836-cf5d-4ac9-80fe-97556117e65e-kube-api-access-6gnlv\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.702162 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8e03836-cf5d-4ac9-80fe-97556117e65e-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.702241 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8e03836-cf5d-4ac9-80fe-97556117e65e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.702303 4700 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/d8e03836-cf5d-4ac9-80fe-97556117e65e-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.737633 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7db57cbc8-t9p8l"] Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.752150 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"d8e03836-cf5d-4ac9-80fe-97556117e65e","Type":"ContainerDied","Data":"5e543ce39643617a2d1093e251750a3f81e9cf2abd59ae1d43c1c647ea768d2a"} Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.752211 4700 scope.go:117] "RemoveContainer" containerID="0f536762f0202ca79c315c2b9f19d46b59472ea624b34e548af06e9322bb143e" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.752389 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.795948 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"61baccca-f7e6-4442-b612-8ab97b82036d","Type":"ContainerStarted","Data":"6f9ba752e86e4641d39854d1aa0c23e4bc4cc492d9d9d37d744899ec3be2bd0c"} Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.801267 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7744b9dc6c-qnsm2" event={"ID":"5ca33b95-1b7a-4d1c-a3b8-65a5969945bc","Type":"ContainerStarted","Data":"0a2214348a1e10b67335e27f4e595d328ec5df3df2658744b07eac983036bf84"} Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.818192 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.818369 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"4099c554-cf42-48b3-88cb-e672a66f1a05","Type":"ContainerDied","Data":"133172d8e994f23225b6e8b5515cb0f7bbadf05464ecf9b4304224beae2992bd"} Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.837310 4700 generic.go:334] "Generic (PLEG): container finished" podID="9822c457-a600-4d01-9b94-b7b899ef2102" containerID="68026171c113261075b633116718884cfbc6f7e2bc616208c246fd9b332e8dab" exitCode=0 Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.837340 4700 generic.go:334] "Generic (PLEG): container finished" podID="9822c457-a600-4d01-9b94-b7b899ef2102" containerID="4593dfb0b9741e168873d46911f0902e7eb87fedecd51d7a9be60ebce9628d31" exitCode=143 Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.837406 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"9822c457-a600-4d01-9b94-b7b899ef2102","Type":"ContainerDied","Data":"68026171c113261075b633116718884cfbc6f7e2bc616208c246fd9b332e8dab"} Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.837432 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"9822c457-a600-4d01-9b94-b7b899ef2102","Type":"ContainerDied","Data":"4593dfb0b9741e168873d46911f0902e7eb87fedecd51d7a9be60ebce9628d31"} Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.851199 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dmf4m" event={"ID":"989eacab-9c48-4375-b7a0-f2c7f40d17ca","Type":"ContainerDied","Data":"fffa2885ffb0a6f235b8106c289c562b951a3d3d8c451756699d73c0ea0b6cdb"} Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.851240 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fffa2885ffb0a6f235b8106c289c562b951a3d3d8c451756699d73c0ea0b6cdb" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.851292 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dmf4m" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.853664 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8946ccf97-n4k7p"] Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.878071 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536882-csprw" event={"ID":"58023608-7e65-4290-b24c-ea314a3fb15b","Type":"ContainerDied","Data":"2915d1805d5e772beb347ce1b08adde377f9bbae29d3bdcef031e9b0e31b44f1"} Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.878127 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2915d1805d5e772beb347ce1b08adde377f9bbae29d3bdcef031e9b0e31b44f1" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.878219 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536882-csprw" Feb 27 17:22:16 crc kubenswrapper[4700]: W0227 17:22:16.902177 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf0d00b91_0e33_4466_8c9f_6a1c1038d23f.slice/crio-665272820cc2663fa668acd424d980adf81034d6df2f2bfb7a976ce5645487d8 WatchSource:0}: Error finding container 665272820cc2663fa668acd424d980adf81034d6df2f2bfb7a976ce5645487d8: Status 404 returned error can't find the container with id 665272820cc2663fa668acd424d980adf81034d6df2f2bfb7a976ce5645487d8 Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.907027 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-64fcf8677d-mvkvn"] Feb 27 17:22:16 crc kubenswrapper[4700]: E0227 17:22:16.907415 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84901b1e-0055-4489-8009-711c1676e51f" containerName="keystone-bootstrap" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.907431 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="84901b1e-0055-4489-8009-711c1676e51f" containerName="keystone-bootstrap" Feb 27 17:22:16 crc kubenswrapper[4700]: E0227 17:22:16.907447 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8e03836-cf5d-4ac9-80fe-97556117e65e" containerName="watcher-decision-engine" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.907468 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8e03836-cf5d-4ac9-80fe-97556117e65e" containerName="watcher-decision-engine" Feb 27 17:22:16 crc kubenswrapper[4700]: E0227 17:22:16.907486 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4099c554-cf42-48b3-88cb-e672a66f1a05" containerName="watcher-applier" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.907492 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="4099c554-cf42-48b3-88cb-e672a66f1a05" containerName="watcher-applier" Feb 27 17:22:16 crc kubenswrapper[4700]: E0227 17:22:16.907512 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58023608-7e65-4290-b24c-ea314a3fb15b" containerName="oc" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.907518 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="58023608-7e65-4290-b24c-ea314a3fb15b" containerName="oc" Feb 27 17:22:16 crc kubenswrapper[4700]: E0227 17:22:16.907527 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="989eacab-9c48-4375-b7a0-f2c7f40d17ca" containerName="placement-db-sync" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.907533 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="989eacab-9c48-4375-b7a0-f2c7f40d17ca" containerName="placement-db-sync" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.907695 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="989eacab-9c48-4375-b7a0-f2c7f40d17ca" containerName="placement-db-sync" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.907706 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="4099c554-cf42-48b3-88cb-e672a66f1a05" containerName="watcher-applier" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.907716 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="58023608-7e65-4290-b24c-ea314a3fb15b" containerName="oc" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.907728 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="84901b1e-0055-4489-8009-711c1676e51f" containerName="keystone-bootstrap" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.907740 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8e03836-cf5d-4ac9-80fe-97556117e65e" containerName="watcher-decision-engine" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.908682 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-64fcf8677d-mvkvn" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.930144 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-64fcf8677d-mvkvn"] Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.936364 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.936535 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.936667 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.936889 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-psmcn" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.937023 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.937760 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8xqhb" event={"ID":"84901b1e-0055-4489-8009-711c1676e51f","Type":"ContainerDied","Data":"9d7d6854081b95b98be8993482e03c572e56d28f096ac89d8569715004cea457"} Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.937829 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d7d6854081b95b98be8993482e03c572e56d28f096ac89d8569715004cea457" Feb 27 17:22:16 crc kubenswrapper[4700]: I0227 17:22:16.937812 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8xqhb" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.051446 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-applier-0"] Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.060281 4700 scope.go:117] "RemoveContainer" containerID="790d64d806f8cf030b5fa73b8846ad67dd2fb782f8fff50bd6090997023776d8" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.096896 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-applier-0"] Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.114850 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-combined-ca-bundle\") pod \"placement-64fcf8677d-mvkvn\" (UID: \"df366836-b6d3-42a1-ab1d-39f7d77f6cb5\") " pod="openstack/placement-64fcf8677d-mvkvn" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.115055 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-internal-tls-certs\") pod \"placement-64fcf8677d-mvkvn\" (UID: \"df366836-b6d3-42a1-ab1d-39f7d77f6cb5\") " pod="openstack/placement-64fcf8677d-mvkvn" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.115127 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-scripts\") pod \"placement-64fcf8677d-mvkvn\" (UID: \"df366836-b6d3-42a1-ab1d-39f7d77f6cb5\") " pod="openstack/placement-64fcf8677d-mvkvn" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.115207 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-public-tls-certs\") pod \"placement-64fcf8677d-mvkvn\" (UID: \"df366836-b6d3-42a1-ab1d-39f7d77f6cb5\") " pod="openstack/placement-64fcf8677d-mvkvn" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.115280 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lz8wr\" (UniqueName: \"kubernetes.io/projected/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-kube-api-access-lz8wr\") pod \"placement-64fcf8677d-mvkvn\" (UID: \"df366836-b6d3-42a1-ab1d-39f7d77f6cb5\") " pod="openstack/placement-64fcf8677d-mvkvn" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.115372 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-logs\") pod \"placement-64fcf8677d-mvkvn\" (UID: \"df366836-b6d3-42a1-ab1d-39f7d77f6cb5\") " pod="openstack/placement-64fcf8677d-mvkvn" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.115519 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-config-data\") pod \"placement-64fcf8677d-mvkvn\" (UID: \"df366836-b6d3-42a1-ab1d-39f7d77f6cb5\") " pod="openstack/placement-64fcf8677d-mvkvn" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.164941 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-decision-engine-0"] Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.218402 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-config-data\") pod \"placement-64fcf8677d-mvkvn\" (UID: \"df366836-b6d3-42a1-ab1d-39f7d77f6cb5\") " pod="openstack/placement-64fcf8677d-mvkvn" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.218488 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-combined-ca-bundle\") pod \"placement-64fcf8677d-mvkvn\" (UID: \"df366836-b6d3-42a1-ab1d-39f7d77f6cb5\") " pod="openstack/placement-64fcf8677d-mvkvn" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.218513 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-internal-tls-certs\") pod \"placement-64fcf8677d-mvkvn\" (UID: \"df366836-b6d3-42a1-ab1d-39f7d77f6cb5\") " pod="openstack/placement-64fcf8677d-mvkvn" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.218531 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-scripts\") pod \"placement-64fcf8677d-mvkvn\" (UID: \"df366836-b6d3-42a1-ab1d-39f7d77f6cb5\") " pod="openstack/placement-64fcf8677d-mvkvn" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.218561 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-public-tls-certs\") pod \"placement-64fcf8677d-mvkvn\" (UID: \"df366836-b6d3-42a1-ab1d-39f7d77f6cb5\") " pod="openstack/placement-64fcf8677d-mvkvn" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.218577 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lz8wr\" (UniqueName: \"kubernetes.io/projected/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-kube-api-access-lz8wr\") pod \"placement-64fcf8677d-mvkvn\" (UID: \"df366836-b6d3-42a1-ab1d-39f7d77f6cb5\") " pod="openstack/placement-64fcf8677d-mvkvn" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.218599 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-logs\") pod \"placement-64fcf8677d-mvkvn\" (UID: \"df366836-b6d3-42a1-ab1d-39f7d77f6cb5\") " pod="openstack/placement-64fcf8677d-mvkvn" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.219022 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-logs\") pod \"placement-64fcf8677d-mvkvn\" (UID: \"df366836-b6d3-42a1-ab1d-39f7d77f6cb5\") " pod="openstack/placement-64fcf8677d-mvkvn" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.249374 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-decision-engine-0"] Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.274306 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-internal-tls-certs\") pod \"placement-64fcf8677d-mvkvn\" (UID: \"df366836-b6d3-42a1-ab1d-39f7d77f6cb5\") " pod="openstack/placement-64fcf8677d-mvkvn" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.306820 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-applier-0"] Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.329847 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.336356 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-combined-ca-bundle\") pod \"placement-64fcf8677d-mvkvn\" (UID: \"df366836-b6d3-42a1-ab1d-39f7d77f6cb5\") " pod="openstack/placement-64fcf8677d-mvkvn" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.343818 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.356384 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-config-data\") pod \"placement-64fcf8677d-mvkvn\" (UID: \"df366836-b6d3-42a1-ab1d-39f7d77f6cb5\") " pod="openstack/placement-64fcf8677d-mvkvn" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.361726 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-public-tls-certs\") pod \"placement-64fcf8677d-mvkvn\" (UID: \"df366836-b6d3-42a1-ab1d-39f7d77f6cb5\") " pod="openstack/placement-64fcf8677d-mvkvn" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.364218 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-decision-engine-0"] Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.364488 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lz8wr\" (UniqueName: \"kubernetes.io/projected/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-kube-api-access-lz8wr\") pod \"placement-64fcf8677d-mvkvn\" (UID: \"df366836-b6d3-42a1-ab1d-39f7d77f6cb5\") " pod="openstack/placement-64fcf8677d-mvkvn" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.377919 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-scripts\") pod \"placement-64fcf8677d-mvkvn\" (UID: \"df366836-b6d3-42a1-ab1d-39f7d77f6cb5\") " pod="openstack/placement-64fcf8677d-mvkvn" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.399978 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-applier-config-data" Feb 27 17:22:17 crc kubenswrapper[4700]: E0227 17:22:17.401902 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9822c457-a600-4d01-9b94-b7b899ef2102" containerName="watcher-api" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.401941 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="9822c457-a600-4d01-9b94-b7b899ef2102" containerName="watcher-api" Feb 27 17:22:17 crc kubenswrapper[4700]: E0227 17:22:17.401970 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9822c457-a600-4d01-9b94-b7b899ef2102" containerName="watcher-api-log" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.401979 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="9822c457-a600-4d01-9b94-b7b899ef2102" containerName="watcher-api-log" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.404020 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="9822c457-a600-4d01-9b94-b7b899ef2102" containerName="watcher-api" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.404057 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="9822c457-a600-4d01-9b94-b7b899ef2102" containerName="watcher-api-log" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.404918 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.418959 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-decision-engine-config-data" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.496073 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.513682 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.553757 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9822c457-a600-4d01-9b94-b7b899ef2102-logs\") pod \"9822c457-a600-4d01-9b94-b7b899ef2102\" (UID: \"9822c457-a600-4d01-9b94-b7b899ef2102\") " Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.553808 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rhx5\" (UniqueName: \"kubernetes.io/projected/9822c457-a600-4d01-9b94-b7b899ef2102-kube-api-access-8rhx5\") pod \"9822c457-a600-4d01-9b94-b7b899ef2102\" (UID: \"9822c457-a600-4d01-9b94-b7b899ef2102\") " Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.553848 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9822c457-a600-4d01-9b94-b7b899ef2102-combined-ca-bundle\") pod \"9822c457-a600-4d01-9b94-b7b899ef2102\" (UID: \"9822c457-a600-4d01-9b94-b7b899ef2102\") " Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.553874 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/9822c457-a600-4d01-9b94-b7b899ef2102-custom-prometheus-ca\") pod \"9822c457-a600-4d01-9b94-b7b899ef2102\" (UID: \"9822c457-a600-4d01-9b94-b7b899ef2102\") " Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.554048 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9822c457-a600-4d01-9b94-b7b899ef2102-config-data\") pod \"9822c457-a600-4d01-9b94-b7b899ef2102\" (UID: \"9822c457-a600-4d01-9b94-b7b899ef2102\") " Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.554156 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9822c457-a600-4d01-9b94-b7b899ef2102-logs" (OuterVolumeSpecName: "logs") pod "9822c457-a600-4d01-9b94-b7b899ef2102" (UID: "9822c457-a600-4d01-9b94-b7b899ef2102"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.554374 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4\") " pod="openstack/watcher-applier-0" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.554393 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f1ae433-28e4-45cf-9635-dc6edbf16f6e-logs\") pod \"watcher-decision-engine-0\" (UID: \"0f1ae433-28e4-45cf-9635-dc6edbf16f6e\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.554468 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntfbz\" (UniqueName: \"kubernetes.io/projected/0f1ae433-28e4-45cf-9635-dc6edbf16f6e-kube-api-access-ntfbz\") pod \"watcher-decision-engine-0\" (UID: \"0f1ae433-28e4-45cf-9635-dc6edbf16f6e\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.554490 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f1ae433-28e4-45cf-9635-dc6edbf16f6e-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"0f1ae433-28e4-45cf-9635-dc6edbf16f6e\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.554515 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4-logs\") pod \"watcher-applier-0\" (UID: \"95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4\") " pod="openstack/watcher-applier-0" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.554530 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7zsv\" (UniqueName: \"kubernetes.io/projected/95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4-kube-api-access-n7zsv\") pod \"watcher-applier-0\" (UID: \"95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4\") " pod="openstack/watcher-applier-0" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.554594 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f1ae433-28e4-45cf-9635-dc6edbf16f6e-config-data\") pod \"watcher-decision-engine-0\" (UID: \"0f1ae433-28e4-45cf-9635-dc6edbf16f6e\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.554611 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/0f1ae433-28e4-45cf-9635-dc6edbf16f6e-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"0f1ae433-28e4-45cf-9635-dc6edbf16f6e\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.554642 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4-config-data\") pod \"watcher-applier-0\" (UID: \"95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4\") " pod="openstack/watcher-applier-0" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.554681 4700 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9822c457-a600-4d01-9b94-b7b899ef2102-logs\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.558799 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9822c457-a600-4d01-9b94-b7b899ef2102-kube-api-access-8rhx5" (OuterVolumeSpecName: "kube-api-access-8rhx5") pod "9822c457-a600-4d01-9b94-b7b899ef2102" (UID: "9822c457-a600-4d01-9b94-b7b899ef2102"). InnerVolumeSpecName "kube-api-access-8rhx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.560635 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7dcdff5d6d-9f49t"] Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.561890 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7dcdff5d6d-9f49t" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.564836 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.564957 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-llljd" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.565571 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.565712 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.565912 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.566105 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.579117 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7dcdff5d6d-9f49t"] Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.583926 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-64fcf8677d-mvkvn" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.588590 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9822c457-a600-4d01-9b94-b7b899ef2102-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "9822c457-a600-4d01-9b94-b7b899ef2102" (UID: "9822c457-a600-4d01-9b94-b7b899ef2102"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.589482 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536876-pxhvd"] Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.596911 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536876-pxhvd"] Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.598572 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9822c457-a600-4d01-9b94-b7b899ef2102-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9822c457-a600-4d01-9b94-b7b899ef2102" (UID: "9822c457-a600-4d01-9b94-b7b899ef2102"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.618083 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-vrb9h" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.660080 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntfbz\" (UniqueName: \"kubernetes.io/projected/0f1ae433-28e4-45cf-9635-dc6edbf16f6e-kube-api-access-ntfbz\") pod \"watcher-decision-engine-0\" (UID: \"0f1ae433-28e4-45cf-9635-dc6edbf16f6e\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.660123 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f1ae433-28e4-45cf-9635-dc6edbf16f6e-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"0f1ae433-28e4-45cf-9635-dc6edbf16f6e\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.660157 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4-logs\") pod \"watcher-applier-0\" (UID: \"95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4\") " pod="openstack/watcher-applier-0" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.660179 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7zsv\" (UniqueName: \"kubernetes.io/projected/95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4-kube-api-access-n7zsv\") pod \"watcher-applier-0\" (UID: \"95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4\") " pod="openstack/watcher-applier-0" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.660260 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/0f1ae433-28e4-45cf-9635-dc6edbf16f6e-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"0f1ae433-28e4-45cf-9635-dc6edbf16f6e\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.660282 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f1ae433-28e4-45cf-9635-dc6edbf16f6e-config-data\") pod \"watcher-decision-engine-0\" (UID: \"0f1ae433-28e4-45cf-9635-dc6edbf16f6e\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.660327 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4-config-data\") pod \"watcher-applier-0\" (UID: \"95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4\") " pod="openstack/watcher-applier-0" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.660359 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4\") " pod="openstack/watcher-applier-0" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.660380 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f1ae433-28e4-45cf-9635-dc6edbf16f6e-logs\") pod \"watcher-decision-engine-0\" (UID: \"0f1ae433-28e4-45cf-9635-dc6edbf16f6e\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.660438 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rhx5\" (UniqueName: \"kubernetes.io/projected/9822c457-a600-4d01-9b94-b7b899ef2102-kube-api-access-8rhx5\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.660453 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9822c457-a600-4d01-9b94-b7b899ef2102-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.660482 4700 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/9822c457-a600-4d01-9b94-b7b899ef2102-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.661340 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f1ae433-28e4-45cf-9635-dc6edbf16f6e-logs\") pod \"watcher-decision-engine-0\" (UID: \"0f1ae433-28e4-45cf-9635-dc6edbf16f6e\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.667419 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4-logs\") pod \"watcher-applier-0\" (UID: \"95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4\") " pod="openstack/watcher-applier-0" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.676153 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f1ae433-28e4-45cf-9635-dc6edbf16f6e-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"0f1ae433-28e4-45cf-9635-dc6edbf16f6e\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.681295 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/0f1ae433-28e4-45cf-9635-dc6edbf16f6e-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"0f1ae433-28e4-45cf-9635-dc6edbf16f6e\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.685063 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4-config-data\") pod \"watcher-applier-0\" (UID: \"95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4\") " pod="openstack/watcher-applier-0" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.685063 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f1ae433-28e4-45cf-9635-dc6edbf16f6e-config-data\") pod \"watcher-decision-engine-0\" (UID: \"0f1ae433-28e4-45cf-9635-dc6edbf16f6e\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.688160 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4\") " pod="openstack/watcher-applier-0" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.704324 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7zsv\" (UniqueName: \"kubernetes.io/projected/95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4-kube-api-access-n7zsv\") pod \"watcher-applier-0\" (UID: \"95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4\") " pod="openstack/watcher-applier-0" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.715067 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntfbz\" (UniqueName: \"kubernetes.io/projected/0f1ae433-28e4-45cf-9635-dc6edbf16f6e-kube-api-access-ntfbz\") pod \"watcher-decision-engine-0\" (UID: \"0f1ae433-28e4-45cf-9635-dc6edbf16f6e\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.753217 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9822c457-a600-4d01-9b94-b7b899ef2102-config-data" (OuterVolumeSpecName: "config-data") pod "9822c457-a600-4d01-9b94-b7b899ef2102" (UID: "9822c457-a600-4d01-9b94-b7b899ef2102"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.762131 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f72833f0-cf5f-4bf8-bb5f-1f65477c4b76-combined-ca-bundle\") pod \"f72833f0-cf5f-4bf8-bb5f-1f65477c4b76\" (UID: \"f72833f0-cf5f-4bf8-bb5f-1f65477c4b76\") " Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.762213 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f72833f0-cf5f-4bf8-bb5f-1f65477c4b76-db-sync-config-data\") pod \"f72833f0-cf5f-4bf8-bb5f-1f65477c4b76\" (UID: \"f72833f0-cf5f-4bf8-bb5f-1f65477c4b76\") " Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.762322 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sssn7\" (UniqueName: \"kubernetes.io/projected/f72833f0-cf5f-4bf8-bb5f-1f65477c4b76-kube-api-access-sssn7\") pod \"f72833f0-cf5f-4bf8-bb5f-1f65477c4b76\" (UID: \"f72833f0-cf5f-4bf8-bb5f-1f65477c4b76\") " Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.762584 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/47d9c3da-3a41-4ed3-93d9-1218f2d7a441-fernet-keys\") pod \"keystone-7dcdff5d6d-9f49t\" (UID: \"47d9c3da-3a41-4ed3-93d9-1218f2d7a441\") " pod="openstack/keystone-7dcdff5d6d-9f49t" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.762625 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hct6\" (UniqueName: \"kubernetes.io/projected/47d9c3da-3a41-4ed3-93d9-1218f2d7a441-kube-api-access-6hct6\") pod \"keystone-7dcdff5d6d-9f49t\" (UID: \"47d9c3da-3a41-4ed3-93d9-1218f2d7a441\") " pod="openstack/keystone-7dcdff5d6d-9f49t" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.762645 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47d9c3da-3a41-4ed3-93d9-1218f2d7a441-config-data\") pod \"keystone-7dcdff5d6d-9f49t\" (UID: \"47d9c3da-3a41-4ed3-93d9-1218f2d7a441\") " pod="openstack/keystone-7dcdff5d6d-9f49t" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.762666 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/47d9c3da-3a41-4ed3-93d9-1218f2d7a441-internal-tls-certs\") pod \"keystone-7dcdff5d6d-9f49t\" (UID: \"47d9c3da-3a41-4ed3-93d9-1218f2d7a441\") " pod="openstack/keystone-7dcdff5d6d-9f49t" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.762692 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/47d9c3da-3a41-4ed3-93d9-1218f2d7a441-public-tls-certs\") pod \"keystone-7dcdff5d6d-9f49t\" (UID: \"47d9c3da-3a41-4ed3-93d9-1218f2d7a441\") " pod="openstack/keystone-7dcdff5d6d-9f49t" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.762713 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/47d9c3da-3a41-4ed3-93d9-1218f2d7a441-credential-keys\") pod \"keystone-7dcdff5d6d-9f49t\" (UID: \"47d9c3da-3a41-4ed3-93d9-1218f2d7a441\") " pod="openstack/keystone-7dcdff5d6d-9f49t" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.762755 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d9c3da-3a41-4ed3-93d9-1218f2d7a441-combined-ca-bundle\") pod \"keystone-7dcdff5d6d-9f49t\" (UID: \"47d9c3da-3a41-4ed3-93d9-1218f2d7a441\") " pod="openstack/keystone-7dcdff5d6d-9f49t" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.762779 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47d9c3da-3a41-4ed3-93d9-1218f2d7a441-scripts\") pod \"keystone-7dcdff5d6d-9f49t\" (UID: \"47d9c3da-3a41-4ed3-93d9-1218f2d7a441\") " pod="openstack/keystone-7dcdff5d6d-9f49t" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.762858 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9822c457-a600-4d01-9b94-b7b899ef2102-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.782002 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f72833f0-cf5f-4bf8-bb5f-1f65477c4b76-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f72833f0-cf5f-4bf8-bb5f-1f65477c4b76" (UID: "f72833f0-cf5f-4bf8-bb5f-1f65477c4b76"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.782647 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f72833f0-cf5f-4bf8-bb5f-1f65477c4b76-kube-api-access-sssn7" (OuterVolumeSpecName: "kube-api-access-sssn7") pod "f72833f0-cf5f-4bf8-bb5f-1f65477c4b76" (UID: "f72833f0-cf5f-4bf8-bb5f-1f65477c4b76"). InnerVolumeSpecName "kube-api-access-sssn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.835567 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f72833f0-cf5f-4bf8-bb5f-1f65477c4b76-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f72833f0-cf5f-4bf8-bb5f-1f65477c4b76" (UID: "f72833f0-cf5f-4bf8-bb5f-1f65477c4b76"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.849897 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.856262 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.864366 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/47d9c3da-3a41-4ed3-93d9-1218f2d7a441-credential-keys\") pod \"keystone-7dcdff5d6d-9f49t\" (UID: \"47d9c3da-3a41-4ed3-93d9-1218f2d7a441\") " pod="openstack/keystone-7dcdff5d6d-9f49t" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.864433 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d9c3da-3a41-4ed3-93d9-1218f2d7a441-combined-ca-bundle\") pod \"keystone-7dcdff5d6d-9f49t\" (UID: \"47d9c3da-3a41-4ed3-93d9-1218f2d7a441\") " pod="openstack/keystone-7dcdff5d6d-9f49t" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.864539 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47d9c3da-3a41-4ed3-93d9-1218f2d7a441-scripts\") pod \"keystone-7dcdff5d6d-9f49t\" (UID: \"47d9c3da-3a41-4ed3-93d9-1218f2d7a441\") " pod="openstack/keystone-7dcdff5d6d-9f49t" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.864617 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/47d9c3da-3a41-4ed3-93d9-1218f2d7a441-fernet-keys\") pod \"keystone-7dcdff5d6d-9f49t\" (UID: \"47d9c3da-3a41-4ed3-93d9-1218f2d7a441\") " pod="openstack/keystone-7dcdff5d6d-9f49t" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.864650 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hct6\" (UniqueName: \"kubernetes.io/projected/47d9c3da-3a41-4ed3-93d9-1218f2d7a441-kube-api-access-6hct6\") pod \"keystone-7dcdff5d6d-9f49t\" (UID: \"47d9c3da-3a41-4ed3-93d9-1218f2d7a441\") " pod="openstack/keystone-7dcdff5d6d-9f49t" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.864670 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47d9c3da-3a41-4ed3-93d9-1218f2d7a441-config-data\") pod \"keystone-7dcdff5d6d-9f49t\" (UID: \"47d9c3da-3a41-4ed3-93d9-1218f2d7a441\") " pod="openstack/keystone-7dcdff5d6d-9f49t" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.864690 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/47d9c3da-3a41-4ed3-93d9-1218f2d7a441-internal-tls-certs\") pod \"keystone-7dcdff5d6d-9f49t\" (UID: \"47d9c3da-3a41-4ed3-93d9-1218f2d7a441\") " pod="openstack/keystone-7dcdff5d6d-9f49t" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.864717 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/47d9c3da-3a41-4ed3-93d9-1218f2d7a441-public-tls-certs\") pod \"keystone-7dcdff5d6d-9f49t\" (UID: \"47d9c3da-3a41-4ed3-93d9-1218f2d7a441\") " pod="openstack/keystone-7dcdff5d6d-9f49t" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.864761 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f72833f0-cf5f-4bf8-bb5f-1f65477c4b76-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.864771 4700 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f72833f0-cf5f-4bf8-bb5f-1f65477c4b76-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.864780 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sssn7\" (UniqueName: \"kubernetes.io/projected/f72833f0-cf5f-4bf8-bb5f-1f65477c4b76-kube-api-access-sssn7\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.887431 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/47d9c3da-3a41-4ed3-93d9-1218f2d7a441-public-tls-certs\") pod \"keystone-7dcdff5d6d-9f49t\" (UID: \"47d9c3da-3a41-4ed3-93d9-1218f2d7a441\") " pod="openstack/keystone-7dcdff5d6d-9f49t" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.907661 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/47d9c3da-3a41-4ed3-93d9-1218f2d7a441-credential-keys\") pod \"keystone-7dcdff5d6d-9f49t\" (UID: \"47d9c3da-3a41-4ed3-93d9-1218f2d7a441\") " pod="openstack/keystone-7dcdff5d6d-9f49t" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.919530 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/47d9c3da-3a41-4ed3-93d9-1218f2d7a441-fernet-keys\") pod \"keystone-7dcdff5d6d-9f49t\" (UID: \"47d9c3da-3a41-4ed3-93d9-1218f2d7a441\") " pod="openstack/keystone-7dcdff5d6d-9f49t" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.927850 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d9c3da-3a41-4ed3-93d9-1218f2d7a441-combined-ca-bundle\") pod \"keystone-7dcdff5d6d-9f49t\" (UID: \"47d9c3da-3a41-4ed3-93d9-1218f2d7a441\") " pod="openstack/keystone-7dcdff5d6d-9f49t" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.938897 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47d9c3da-3a41-4ed3-93d9-1218f2d7a441-scripts\") pod \"keystone-7dcdff5d6d-9f49t\" (UID: \"47d9c3da-3a41-4ed3-93d9-1218f2d7a441\") " pod="openstack/keystone-7dcdff5d6d-9f49t" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.939244 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hct6\" (UniqueName: \"kubernetes.io/projected/47d9c3da-3a41-4ed3-93d9-1218f2d7a441-kube-api-access-6hct6\") pod \"keystone-7dcdff5d6d-9f49t\" (UID: \"47d9c3da-3a41-4ed3-93d9-1218f2d7a441\") " pod="openstack/keystone-7dcdff5d6d-9f49t" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.940775 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/47d9c3da-3a41-4ed3-93d9-1218f2d7a441-internal-tls-certs\") pod \"keystone-7dcdff5d6d-9f49t\" (UID: \"47d9c3da-3a41-4ed3-93d9-1218f2d7a441\") " pod="openstack/keystone-7dcdff5d6d-9f49t" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.941313 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47d9c3da-3a41-4ed3-93d9-1218f2d7a441-config-data\") pod \"keystone-7dcdff5d6d-9f49t\" (UID: \"47d9c3da-3a41-4ed3-93d9-1218f2d7a441\") " pod="openstack/keystone-7dcdff5d6d-9f49t" Feb 27 17:22:17 crc kubenswrapper[4700]: I0227 17:22:17.968318 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7dcdff5d6d-9f49t" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.014668 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-7f87979786-hh2zw"] Feb 27 17:22:18 crc kubenswrapper[4700]: E0227 17:22:18.015020 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f72833f0-cf5f-4bf8-bb5f-1f65477c4b76" containerName="barbican-db-sync" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.015036 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="f72833f0-cf5f-4bf8-bb5f-1f65477c4b76" containerName="barbican-db-sync" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.015202 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="f72833f0-cf5f-4bf8-bb5f-1f65477c4b76" containerName="barbican-db-sync" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.021033 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7db57cbc8-t9p8l" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.021059 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7db57cbc8-t9p8l" event={"ID":"24cf75b6-81b2-43e8-b75e-c41171313a0f","Type":"ContainerStarted","Data":"994e8654f68b07a71d26635805e774bb44d66badb496942fec0b2817d51b9f27"} Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.021076 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7db57cbc8-t9p8l" event={"ID":"24cf75b6-81b2-43e8-b75e-c41171313a0f","Type":"ContainerStarted","Data":"3675602a5b830e0a8101ab8de56cf97a66e7e05593be9748eed8b289659a40a0"} Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.021086 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7db57cbc8-t9p8l" event={"ID":"24cf75b6-81b2-43e8-b75e-c41171313a0f","Type":"ContainerStarted","Data":"15635a8ed85c68ba78061124a91a84acf4bc7783426a91c3f6401f894c977b93"} Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.021154 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7f87979786-hh2zw" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.034674 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7f87979786-hh2zw"] Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.050084 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.076614 4700 generic.go:334] "Generic (PLEG): container finished" podID="5ca33b95-1b7a-4d1c-a3b8-65a5969945bc" containerID="d4f0a0eedccf69f640673b202c20c063df1d5604705e8c1a133236b1040464c7" exitCode=0 Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.076694 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7744b9dc6c-qnsm2" event={"ID":"5ca33b95-1b7a-4d1c-a3b8-65a5969945bc","Type":"ContainerDied","Data":"d4f0a0eedccf69f640673b202c20c063df1d5604705e8c1a133236b1040464c7"} Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.110657 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"345c36c6-a3e3-4ece-9ac8-d59a7b31b366","Type":"ContainerStarted","Data":"f3160c9d9756370992369dfa24520fea2a6602a148ba480ccc954e0d6a434526"} Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.132278 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-756b7f868f-z7qr7"] Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.134018 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-756b7f868f-z7qr7" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.143198 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.150840 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c","Type":"ContainerStarted","Data":"a6a7945d87c6b8692ed3f3f79afc16e0fa98c0e1a7600285da5d09af361e8f2c"} Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.159318 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-vrb9h" event={"ID":"f72833f0-cf5f-4bf8-bb5f-1f65477c4b76","Type":"ContainerDied","Data":"6c6f06c549b0df8963eef790078a533d0845705b5c6a418be97923315f7fd804"} Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.159359 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c6f06c549b0df8963eef790078a533d0845705b5c6a418be97923315f7fd804" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.159418 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-vrb9h" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.175231 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xn4pq\" (UniqueName: \"kubernetes.io/projected/3b6c33fa-4855-4b22-a634-cbc739ec0580-kube-api-access-xn4pq\") pod \"barbican-keystone-listener-7f87979786-hh2zw\" (UID: \"3b6c33fa-4855-4b22-a634-cbc739ec0580\") " pod="openstack/barbican-keystone-listener-7f87979786-hh2zw" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.175451 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b6c33fa-4855-4b22-a634-cbc739ec0580-logs\") pod \"barbican-keystone-listener-7f87979786-hh2zw\" (UID: \"3b6c33fa-4855-4b22-a634-cbc739ec0580\") " pod="openstack/barbican-keystone-listener-7f87979786-hh2zw" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.175594 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b6c33fa-4855-4b22-a634-cbc739ec0580-combined-ca-bundle\") pod \"barbican-keystone-listener-7f87979786-hh2zw\" (UID: \"3b6c33fa-4855-4b22-a634-cbc739ec0580\") " pod="openstack/barbican-keystone-listener-7f87979786-hh2zw" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.175661 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b6c33fa-4855-4b22-a634-cbc739ec0580-config-data\") pod \"barbican-keystone-listener-7f87979786-hh2zw\" (UID: \"3b6c33fa-4855-4b22-a634-cbc739ec0580\") " pod="openstack/barbican-keystone-listener-7f87979786-hh2zw" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.175688 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3b6c33fa-4855-4b22-a634-cbc739ec0580-config-data-custom\") pod \"barbican-keystone-listener-7f87979786-hh2zw\" (UID: \"3b6c33fa-4855-4b22-a634-cbc739ec0580\") " pod="openstack/barbican-keystone-listener-7f87979786-hh2zw" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.176169 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.176550 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"9822c457-a600-4d01-9b94-b7b899ef2102","Type":"ContainerDied","Data":"658bae7ec88504e1af3430564755bf8041919992ca9fbcb804d379bf8a4b5692"} Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.176596 4700 scope.go:117] "RemoveContainer" containerID="68026171c113261075b633116718884cfbc6f7e2bc616208c246fd9b332e8dab" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.189347 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8946ccf97-n4k7p" event={"ID":"f0d00b91-0e33-4466-8c9f-6a1c1038d23f","Type":"ContainerStarted","Data":"276f51e328821d543c3ac41196366ae8614a6c6d49c1d994e3ad93e0d3638898"} Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.189394 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8946ccf97-n4k7p" event={"ID":"f0d00b91-0e33-4466-8c9f-6a1c1038d23f","Type":"ContainerStarted","Data":"665272820cc2663fa668acd424d980adf81034d6df2f2bfb7a976ce5645487d8"} Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.212296 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-756b7f868f-z7qr7"] Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.249997 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7db57cbc8-t9p8l" podStartSLOduration=9.249974292 podStartE2EDuration="9.249974292s" podCreationTimestamp="2026-02-27 17:22:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:22:18.069585772 +0000 UTC m=+1298.054898519" watchObservedRunningTime="2026-02-27 17:22:18.249974292 +0000 UTC m=+1298.235287039" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.274675 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7744b9dc6c-qnsm2"] Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.278325 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1311d698-8af2-4960-9790-f470a67f8457-logs\") pod \"barbican-worker-756b7f868f-z7qr7\" (UID: \"1311d698-8af2-4960-9790-f470a67f8457\") " pod="openstack/barbican-worker-756b7f868f-z7qr7" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.278995 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b6c33fa-4855-4b22-a634-cbc739ec0580-logs\") pod \"barbican-keystone-listener-7f87979786-hh2zw\" (UID: \"3b6c33fa-4855-4b22-a634-cbc739ec0580\") " pod="openstack/barbican-keystone-listener-7f87979786-hh2zw" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.279118 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b6c33fa-4855-4b22-a634-cbc739ec0580-combined-ca-bundle\") pod \"barbican-keystone-listener-7f87979786-hh2zw\" (UID: \"3b6c33fa-4855-4b22-a634-cbc739ec0580\") " pod="openstack/barbican-keystone-listener-7f87979786-hh2zw" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.279149 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1311d698-8af2-4960-9790-f470a67f8457-config-data\") pod \"barbican-worker-756b7f868f-z7qr7\" (UID: \"1311d698-8af2-4960-9790-f470a67f8457\") " pod="openstack/barbican-worker-756b7f868f-z7qr7" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.279218 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1311d698-8af2-4960-9790-f470a67f8457-combined-ca-bundle\") pod \"barbican-worker-756b7f868f-z7qr7\" (UID: \"1311d698-8af2-4960-9790-f470a67f8457\") " pod="openstack/barbican-worker-756b7f868f-z7qr7" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.279238 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b6c33fa-4855-4b22-a634-cbc739ec0580-config-data\") pod \"barbican-keystone-listener-7f87979786-hh2zw\" (UID: \"3b6c33fa-4855-4b22-a634-cbc739ec0580\") " pod="openstack/barbican-keystone-listener-7f87979786-hh2zw" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.279257 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3b6c33fa-4855-4b22-a634-cbc739ec0580-config-data-custom\") pod \"barbican-keystone-listener-7f87979786-hh2zw\" (UID: \"3b6c33fa-4855-4b22-a634-cbc739ec0580\") " pod="openstack/barbican-keystone-listener-7f87979786-hh2zw" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.279312 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1311d698-8af2-4960-9790-f470a67f8457-config-data-custom\") pod \"barbican-worker-756b7f868f-z7qr7\" (UID: \"1311d698-8af2-4960-9790-f470a67f8457\") " pod="openstack/barbican-worker-756b7f868f-z7qr7" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.279380 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xn4pq\" (UniqueName: \"kubernetes.io/projected/3b6c33fa-4855-4b22-a634-cbc739ec0580-kube-api-access-xn4pq\") pod \"barbican-keystone-listener-7f87979786-hh2zw\" (UID: \"3b6c33fa-4855-4b22-a634-cbc739ec0580\") " pod="openstack/barbican-keystone-listener-7f87979786-hh2zw" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.290284 4700 scope.go:117] "RemoveContainer" containerID="4593dfb0b9741e168873d46911f0902e7eb87fedecd51d7a9be60ebce9628d31" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.290338 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3b6c33fa-4855-4b22-a634-cbc739ec0580-config-data-custom\") pod \"barbican-keystone-listener-7f87979786-hh2zw\" (UID: \"3b6c33fa-4855-4b22-a634-cbc739ec0580\") " pod="openstack/barbican-keystone-listener-7f87979786-hh2zw" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.294979 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b6c33fa-4855-4b22-a634-cbc739ec0580-logs\") pod \"barbican-keystone-listener-7f87979786-hh2zw\" (UID: \"3b6c33fa-4855-4b22-a634-cbc739ec0580\") " pod="openstack/barbican-keystone-listener-7f87979786-hh2zw" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.298857 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rbgs\" (UniqueName: \"kubernetes.io/projected/1311d698-8af2-4960-9790-f470a67f8457-kube-api-access-8rbgs\") pod \"barbican-worker-756b7f868f-z7qr7\" (UID: \"1311d698-8af2-4960-9790-f470a67f8457\") " pod="openstack/barbican-worker-756b7f868f-z7qr7" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.304307 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5b48777cdd-tqbkj" podUID="d3989a3e-45ff-45d8-80a2-4dbe57663a57" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.168:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.168:8443: connect: connection refused" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.304336 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b6c33fa-4855-4b22-a634-cbc739ec0580-combined-ca-bundle\") pod \"barbican-keystone-listener-7f87979786-hh2zw\" (UID: \"3b6c33fa-4855-4b22-a634-cbc739ec0580\") " pod="openstack/barbican-keystone-listener-7f87979786-hh2zw" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.312831 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b6c33fa-4855-4b22-a634-cbc739ec0580-config-data\") pod \"barbican-keystone-listener-7f87979786-hh2zw\" (UID: \"3b6c33fa-4855-4b22-a634-cbc739ec0580\") " pod="openstack/barbican-keystone-listener-7f87979786-hh2zw" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.338394 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xn4pq\" (UniqueName: \"kubernetes.io/projected/3b6c33fa-4855-4b22-a634-cbc739ec0580-kube-api-access-xn4pq\") pod \"barbican-keystone-listener-7f87979786-hh2zw\" (UID: \"3b6c33fa-4855-4b22-a634-cbc739ec0580\") " pod="openstack/barbican-keystone-listener-7f87979786-hh2zw" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.388066 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-575575788d-c49lv"] Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.392013 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-575575788d-c49lv" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.400929 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.402523 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1311d698-8af2-4960-9790-f470a67f8457-config-data\") pod \"barbican-worker-756b7f868f-z7qr7\" (UID: \"1311d698-8af2-4960-9790-f470a67f8457\") " pod="openstack/barbican-worker-756b7f868f-z7qr7" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.402575 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1311d698-8af2-4960-9790-f470a67f8457-combined-ca-bundle\") pod \"barbican-worker-756b7f868f-z7qr7\" (UID: \"1311d698-8af2-4960-9790-f470a67f8457\") " pod="openstack/barbican-worker-756b7f868f-z7qr7" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.402601 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1311d698-8af2-4960-9790-f470a67f8457-config-data-custom\") pod \"barbican-worker-756b7f868f-z7qr7\" (UID: \"1311d698-8af2-4960-9790-f470a67f8457\") " pod="openstack/barbican-worker-756b7f868f-z7qr7" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.402645 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rbgs\" (UniqueName: \"kubernetes.io/projected/1311d698-8af2-4960-9790-f470a67f8457-kube-api-access-8rbgs\") pod \"barbican-worker-756b7f868f-z7qr7\" (UID: \"1311d698-8af2-4960-9790-f470a67f8457\") " pod="openstack/barbican-worker-756b7f868f-z7qr7" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.402695 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1311d698-8af2-4960-9790-f470a67f8457-logs\") pod \"barbican-worker-756b7f868f-z7qr7\" (UID: \"1311d698-8af2-4960-9790-f470a67f8457\") " pod="openstack/barbican-worker-756b7f868f-z7qr7" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.403214 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1311d698-8af2-4960-9790-f470a67f8457-logs\") pod \"barbican-worker-756b7f868f-z7qr7\" (UID: \"1311d698-8af2-4960-9790-f470a67f8457\") " pod="openstack/barbican-worker-756b7f868f-z7qr7" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.411446 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1311d698-8af2-4960-9790-f470a67f8457-combined-ca-bundle\") pod \"barbican-worker-756b7f868f-z7qr7\" (UID: \"1311d698-8af2-4960-9790-f470a67f8457\") " pod="openstack/barbican-worker-756b7f868f-z7qr7" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.411864 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7f87979786-hh2zw" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.412304 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77475fb77-jzwgk"] Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.414427 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77475fb77-jzwgk" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.419835 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6cc494ff8-ggq9w" podUID="ed7353a3-748a-40cb-8865-c4badeb8e402" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.169:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.169:8443: connect: connection refused" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.424392 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1311d698-8af2-4960-9790-f470a67f8457-config-data-custom\") pod \"barbican-worker-756b7f868f-z7qr7\" (UID: \"1311d698-8af2-4960-9790-f470a67f8457\") " pod="openstack/barbican-worker-756b7f868f-z7qr7" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.425298 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1311d698-8af2-4960-9790-f470a67f8457-config-data\") pod \"barbican-worker-756b7f868f-z7qr7\" (UID: \"1311d698-8af2-4960-9790-f470a67f8457\") " pod="openstack/barbican-worker-756b7f868f-z7qr7" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.428226 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77475fb77-jzwgk"] Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.435905 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rbgs\" (UniqueName: \"kubernetes.io/projected/1311d698-8af2-4960-9790-f470a67f8457-kube-api-access-8rbgs\") pod \"barbican-worker-756b7f868f-z7qr7\" (UID: \"1311d698-8af2-4960-9790-f470a67f8457\") " pod="openstack/barbican-worker-756b7f868f-z7qr7" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.522474 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb6fd0bc-8abd-4537-9c92-f8027dde0bb8-config-data-custom\") pod \"barbican-api-575575788d-c49lv\" (UID: \"cb6fd0bc-8abd-4537-9c92-f8027dde0bb8\") " pod="openstack/barbican-api-575575788d-c49lv" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.524128 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb6fd0bc-8abd-4537-9c92-f8027dde0bb8-config-data\") pod \"barbican-api-575575788d-c49lv\" (UID: \"cb6fd0bc-8abd-4537-9c92-f8027dde0bb8\") " pod="openstack/barbican-api-575575788d-c49lv" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.524267 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c48cd283-8c2c-4825-ad5e-c014c1fbc369-ovsdbserver-nb\") pod \"dnsmasq-dns-77475fb77-jzwgk\" (UID: \"c48cd283-8c2c-4825-ad5e-c014c1fbc369\") " pod="openstack/dnsmasq-dns-77475fb77-jzwgk" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.524523 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c48cd283-8c2c-4825-ad5e-c014c1fbc369-dns-svc\") pod \"dnsmasq-dns-77475fb77-jzwgk\" (UID: \"c48cd283-8c2c-4825-ad5e-c014c1fbc369\") " pod="openstack/dnsmasq-dns-77475fb77-jzwgk" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.524562 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxgh5\" (UniqueName: \"kubernetes.io/projected/c48cd283-8c2c-4825-ad5e-c014c1fbc369-kube-api-access-pxgh5\") pod \"dnsmasq-dns-77475fb77-jzwgk\" (UID: \"c48cd283-8c2c-4825-ad5e-c014c1fbc369\") " pod="openstack/dnsmasq-dns-77475fb77-jzwgk" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.524683 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjbt9\" (UniqueName: \"kubernetes.io/projected/cb6fd0bc-8abd-4537-9c92-f8027dde0bb8-kube-api-access-tjbt9\") pod \"barbican-api-575575788d-c49lv\" (UID: \"cb6fd0bc-8abd-4537-9c92-f8027dde0bb8\") " pod="openstack/barbican-api-575575788d-c49lv" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.524716 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb6fd0bc-8abd-4537-9c92-f8027dde0bb8-logs\") pod \"barbican-api-575575788d-c49lv\" (UID: \"cb6fd0bc-8abd-4537-9c92-f8027dde0bb8\") " pod="openstack/barbican-api-575575788d-c49lv" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.525213 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb6fd0bc-8abd-4537-9c92-f8027dde0bb8-combined-ca-bundle\") pod \"barbican-api-575575788d-c49lv\" (UID: \"cb6fd0bc-8abd-4537-9c92-f8027dde0bb8\") " pod="openstack/barbican-api-575575788d-c49lv" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.525239 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-756b7f868f-z7qr7" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.525293 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c48cd283-8c2c-4825-ad5e-c014c1fbc369-dns-swift-storage-0\") pod \"dnsmasq-dns-77475fb77-jzwgk\" (UID: \"c48cd283-8c2c-4825-ad5e-c014c1fbc369\") " pod="openstack/dnsmasq-dns-77475fb77-jzwgk" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.525433 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c48cd283-8c2c-4825-ad5e-c014c1fbc369-config\") pod \"dnsmasq-dns-77475fb77-jzwgk\" (UID: \"c48cd283-8c2c-4825-ad5e-c014c1fbc369\") " pod="openstack/dnsmasq-dns-77475fb77-jzwgk" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.525489 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c48cd283-8c2c-4825-ad5e-c014c1fbc369-ovsdbserver-sb\") pod \"dnsmasq-dns-77475fb77-jzwgk\" (UID: \"c48cd283-8c2c-4825-ad5e-c014c1fbc369\") " pod="openstack/dnsmasq-dns-77475fb77-jzwgk" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.540225 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-575575788d-c49lv"] Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.576124 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-64fcf8677d-mvkvn"] Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.599389 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=12.599365943 podStartE2EDuration="12.599365943s" podCreationTimestamp="2026-02-27 17:22:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:22:18.199192152 +0000 UTC m=+1298.184504899" watchObservedRunningTime="2026-02-27 17:22:18.599365943 +0000 UTC m=+1298.584678690" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.644681 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb6fd0bc-8abd-4537-9c92-f8027dde0bb8-combined-ca-bundle\") pod \"barbican-api-575575788d-c49lv\" (UID: \"cb6fd0bc-8abd-4537-9c92-f8027dde0bb8\") " pod="openstack/barbican-api-575575788d-c49lv" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.644738 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c48cd283-8c2c-4825-ad5e-c014c1fbc369-dns-swift-storage-0\") pod \"dnsmasq-dns-77475fb77-jzwgk\" (UID: \"c48cd283-8c2c-4825-ad5e-c014c1fbc369\") " pod="openstack/dnsmasq-dns-77475fb77-jzwgk" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.645511 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c48cd283-8c2c-4825-ad5e-c014c1fbc369-dns-swift-storage-0\") pod \"dnsmasq-dns-77475fb77-jzwgk\" (UID: \"c48cd283-8c2c-4825-ad5e-c014c1fbc369\") " pod="openstack/dnsmasq-dns-77475fb77-jzwgk" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.645594 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c48cd283-8c2c-4825-ad5e-c014c1fbc369-config\") pod \"dnsmasq-dns-77475fb77-jzwgk\" (UID: \"c48cd283-8c2c-4825-ad5e-c014c1fbc369\") " pod="openstack/dnsmasq-dns-77475fb77-jzwgk" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.645617 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c48cd283-8c2c-4825-ad5e-c014c1fbc369-ovsdbserver-sb\") pod \"dnsmasq-dns-77475fb77-jzwgk\" (UID: \"c48cd283-8c2c-4825-ad5e-c014c1fbc369\") " pod="openstack/dnsmasq-dns-77475fb77-jzwgk" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.645658 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb6fd0bc-8abd-4537-9c92-f8027dde0bb8-config-data-custom\") pod \"barbican-api-575575788d-c49lv\" (UID: \"cb6fd0bc-8abd-4537-9c92-f8027dde0bb8\") " pod="openstack/barbican-api-575575788d-c49lv" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.645685 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb6fd0bc-8abd-4537-9c92-f8027dde0bb8-config-data\") pod \"barbican-api-575575788d-c49lv\" (UID: \"cb6fd0bc-8abd-4537-9c92-f8027dde0bb8\") " pod="openstack/barbican-api-575575788d-c49lv" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.645702 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c48cd283-8c2c-4825-ad5e-c014c1fbc369-ovsdbserver-nb\") pod \"dnsmasq-dns-77475fb77-jzwgk\" (UID: \"c48cd283-8c2c-4825-ad5e-c014c1fbc369\") " pod="openstack/dnsmasq-dns-77475fb77-jzwgk" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.645723 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c48cd283-8c2c-4825-ad5e-c014c1fbc369-dns-svc\") pod \"dnsmasq-dns-77475fb77-jzwgk\" (UID: \"c48cd283-8c2c-4825-ad5e-c014c1fbc369\") " pod="openstack/dnsmasq-dns-77475fb77-jzwgk" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.645741 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxgh5\" (UniqueName: \"kubernetes.io/projected/c48cd283-8c2c-4825-ad5e-c014c1fbc369-kube-api-access-pxgh5\") pod \"dnsmasq-dns-77475fb77-jzwgk\" (UID: \"c48cd283-8c2c-4825-ad5e-c014c1fbc369\") " pod="openstack/dnsmasq-dns-77475fb77-jzwgk" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.645756 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjbt9\" (UniqueName: \"kubernetes.io/projected/cb6fd0bc-8abd-4537-9c92-f8027dde0bb8-kube-api-access-tjbt9\") pod \"barbican-api-575575788d-c49lv\" (UID: \"cb6fd0bc-8abd-4537-9c92-f8027dde0bb8\") " pod="openstack/barbican-api-575575788d-c49lv" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.645773 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb6fd0bc-8abd-4537-9c92-f8027dde0bb8-logs\") pod \"barbican-api-575575788d-c49lv\" (UID: \"cb6fd0bc-8abd-4537-9c92-f8027dde0bb8\") " pod="openstack/barbican-api-575575788d-c49lv" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.646122 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb6fd0bc-8abd-4537-9c92-f8027dde0bb8-logs\") pod \"barbican-api-575575788d-c49lv\" (UID: \"cb6fd0bc-8abd-4537-9c92-f8027dde0bb8\") " pod="openstack/barbican-api-575575788d-c49lv" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.650072 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb6fd0bc-8abd-4537-9c92-f8027dde0bb8-combined-ca-bundle\") pod \"barbican-api-575575788d-c49lv\" (UID: \"cb6fd0bc-8abd-4537-9c92-f8027dde0bb8\") " pod="openstack/barbican-api-575575788d-c49lv" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.650399 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c48cd283-8c2c-4825-ad5e-c014c1fbc369-dns-svc\") pod \"dnsmasq-dns-77475fb77-jzwgk\" (UID: \"c48cd283-8c2c-4825-ad5e-c014c1fbc369\") " pod="openstack/dnsmasq-dns-77475fb77-jzwgk" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.650631 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=12.650609726 podStartE2EDuration="12.650609726s" podCreationTimestamp="2026-02-27 17:22:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:22:18.227814348 +0000 UTC m=+1298.213127095" watchObservedRunningTime="2026-02-27 17:22:18.650609726 +0000 UTC m=+1298.635922473" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.655327 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c48cd283-8c2c-4825-ad5e-c014c1fbc369-ovsdbserver-nb\") pod \"dnsmasq-dns-77475fb77-jzwgk\" (UID: \"c48cd283-8c2c-4825-ad5e-c014c1fbc369\") " pod="openstack/dnsmasq-dns-77475fb77-jzwgk" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.655437 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb6fd0bc-8abd-4537-9c92-f8027dde0bb8-config-data-custom\") pod \"barbican-api-575575788d-c49lv\" (UID: \"cb6fd0bc-8abd-4537-9c92-f8027dde0bb8\") " pod="openstack/barbican-api-575575788d-c49lv" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.655545 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c48cd283-8c2c-4825-ad5e-c014c1fbc369-ovsdbserver-sb\") pod \"dnsmasq-dns-77475fb77-jzwgk\" (UID: \"c48cd283-8c2c-4825-ad5e-c014c1fbc369\") " pod="openstack/dnsmasq-dns-77475fb77-jzwgk" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.655660 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c48cd283-8c2c-4825-ad5e-c014c1fbc369-config\") pod \"dnsmasq-dns-77475fb77-jzwgk\" (UID: \"c48cd283-8c2c-4825-ad5e-c014c1fbc369\") " pod="openstack/dnsmasq-dns-77475fb77-jzwgk" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.662751 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb6fd0bc-8abd-4537-9c92-f8027dde0bb8-config-data\") pod \"barbican-api-575575788d-c49lv\" (UID: \"cb6fd0bc-8abd-4537-9c92-f8027dde0bb8\") " pod="openstack/barbican-api-575575788d-c49lv" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.676994 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjbt9\" (UniqueName: \"kubernetes.io/projected/cb6fd0bc-8abd-4537-9c92-f8027dde0bb8-kube-api-access-tjbt9\") pod \"barbican-api-575575788d-c49lv\" (UID: \"cb6fd0bc-8abd-4537-9c92-f8027dde0bb8\") " pod="openstack/barbican-api-575575788d-c49lv" Feb 27 17:22:18 crc kubenswrapper[4700]: E0227 17:22:18.685400 4700 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Feb 27 17:22:18 crc kubenswrapper[4700]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Feb 27 17:22:18 crc kubenswrapper[4700]: > podSandboxID="0a2214348a1e10b67335e27f4e595d328ec5df3df2658744b07eac983036bf84" Feb 27 17:22:18 crc kubenswrapper[4700]: E0227 17:22:18.685682 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 17:22:18 crc kubenswrapper[4700]: container &Container{Name:dnsmasq-dns,Image:38.102.83.145:5001/podified-master-centos10/openstack-neutron-server:watcher_latest,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n65bh9bh5fch67h5dbh647h64ch56fhdhb6h554hd6h5fdhbbh598hfdh56fh69h555h5d5h5h59bh65hb8h5c9h66ch5dhf9h56dh7ch54bh576q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-swift-storage-0,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-swift-storage-0,SubPath:dns-swift-storage-0,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-nb,SubPath:ovsdbserver-nb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-sb,SubPath:ovsdbserver-sb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-97lrn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7744b9dc6c-qnsm2_openstack(5ca33b95-1b7a-4d1c-a3b8-65a5969945bc): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Feb 27 17:22:18 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 17:22:18 crc kubenswrapper[4700]: E0227 17:22:18.686759 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-7744b9dc6c-qnsm2" podUID="5ca33b95-1b7a-4d1c-a3b8-65a5969945bc" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.693966 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxgh5\" (UniqueName: \"kubernetes.io/projected/c48cd283-8c2c-4825-ad5e-c014c1fbc369-kube-api-access-pxgh5\") pod \"dnsmasq-dns-77475fb77-jzwgk\" (UID: \"c48cd283-8c2c-4825-ad5e-c014c1fbc369\") " pod="openstack/dnsmasq-dns-77475fb77-jzwgk" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.704211 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.726566 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-api-0"] Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.727653 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-api-0"] Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.729159 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.752621 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.775703 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.784314 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-watcher-internal-svc" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.785044 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-watcher-public-svc" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.785326 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-api-config-data" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.833939 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-575575788d-c49lv" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.846172 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77475fb77-jzwgk" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.856069 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bcc9554e-7836-4db9-aab8-d5b0bef440e1-logs\") pod \"watcher-api-0\" (UID: \"bcc9554e-7836-4db9-aab8-d5b0bef440e1\") " pod="openstack/watcher-api-0" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.856150 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcc9554e-7836-4db9-aab8-d5b0bef440e1-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"bcc9554e-7836-4db9-aab8-d5b0bef440e1\") " pod="openstack/watcher-api-0" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.856197 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcc9554e-7836-4db9-aab8-d5b0bef440e1-public-tls-certs\") pod \"watcher-api-0\" (UID: \"bcc9554e-7836-4db9-aab8-d5b0bef440e1\") " pod="openstack/watcher-api-0" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.856240 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcc9554e-7836-4db9-aab8-d5b0bef440e1-config-data\") pod \"watcher-api-0\" (UID: \"bcc9554e-7836-4db9-aab8-d5b0bef440e1\") " pod="openstack/watcher-api-0" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.856262 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcc9554e-7836-4db9-aab8-d5b0bef440e1-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"bcc9554e-7836-4db9-aab8-d5b0bef440e1\") " pod="openstack/watcher-api-0" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.856312 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/bcc9554e-7836-4db9-aab8-d5b0bef440e1-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"bcc9554e-7836-4db9-aab8-d5b0bef440e1\") " pod="openstack/watcher-api-0" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.856345 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tprg6\" (UniqueName: \"kubernetes.io/projected/bcc9554e-7836-4db9-aab8-d5b0bef440e1-kube-api-access-tprg6\") pod \"watcher-api-0\" (UID: \"bcc9554e-7836-4db9-aab8-d5b0bef440e1\") " pod="openstack/watcher-api-0" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.948310 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.960433 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcc9554e-7836-4db9-aab8-d5b0bef440e1-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"bcc9554e-7836-4db9-aab8-d5b0bef440e1\") " pod="openstack/watcher-api-0" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.960520 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcc9554e-7836-4db9-aab8-d5b0bef440e1-public-tls-certs\") pod \"watcher-api-0\" (UID: \"bcc9554e-7836-4db9-aab8-d5b0bef440e1\") " pod="openstack/watcher-api-0" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.960561 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcc9554e-7836-4db9-aab8-d5b0bef440e1-config-data\") pod \"watcher-api-0\" (UID: \"bcc9554e-7836-4db9-aab8-d5b0bef440e1\") " pod="openstack/watcher-api-0" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.960583 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcc9554e-7836-4db9-aab8-d5b0bef440e1-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"bcc9554e-7836-4db9-aab8-d5b0bef440e1\") " pod="openstack/watcher-api-0" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.960629 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/bcc9554e-7836-4db9-aab8-d5b0bef440e1-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"bcc9554e-7836-4db9-aab8-d5b0bef440e1\") " pod="openstack/watcher-api-0" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.960660 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tprg6\" (UniqueName: \"kubernetes.io/projected/bcc9554e-7836-4db9-aab8-d5b0bef440e1-kube-api-access-tprg6\") pod \"watcher-api-0\" (UID: \"bcc9554e-7836-4db9-aab8-d5b0bef440e1\") " pod="openstack/watcher-api-0" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.960700 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bcc9554e-7836-4db9-aab8-d5b0bef440e1-logs\") pod \"watcher-api-0\" (UID: \"bcc9554e-7836-4db9-aab8-d5b0bef440e1\") " pod="openstack/watcher-api-0" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.961201 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bcc9554e-7836-4db9-aab8-d5b0bef440e1-logs\") pod \"watcher-api-0\" (UID: \"bcc9554e-7836-4db9-aab8-d5b0bef440e1\") " pod="openstack/watcher-api-0" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.973846 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcc9554e-7836-4db9-aab8-d5b0bef440e1-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"bcc9554e-7836-4db9-aab8-d5b0bef440e1\") " pod="openstack/watcher-api-0" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.975840 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcc9554e-7836-4db9-aab8-d5b0bef440e1-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"bcc9554e-7836-4db9-aab8-d5b0bef440e1\") " pod="openstack/watcher-api-0" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.978939 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcc9554e-7836-4db9-aab8-d5b0bef440e1-public-tls-certs\") pod \"watcher-api-0\" (UID: \"bcc9554e-7836-4db9-aab8-d5b0bef440e1\") " pod="openstack/watcher-api-0" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.979822 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcc9554e-7836-4db9-aab8-d5b0bef440e1-config-data\") pod \"watcher-api-0\" (UID: \"bcc9554e-7836-4db9-aab8-d5b0bef440e1\") " pod="openstack/watcher-api-0" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.988201 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/bcc9554e-7836-4db9-aab8-d5b0bef440e1-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"bcc9554e-7836-4db9-aab8-d5b0bef440e1\") " pod="openstack/watcher-api-0" Feb 27 17:22:18 crc kubenswrapper[4700]: I0227 17:22:18.995438 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tprg6\" (UniqueName: \"kubernetes.io/projected/bcc9554e-7836-4db9-aab8-d5b0bef440e1-kube-api-access-tprg6\") pod \"watcher-api-0\" (UID: \"bcc9554e-7836-4db9-aab8-d5b0bef440e1\") " pod="openstack/watcher-api-0" Feb 27 17:22:19 crc kubenswrapper[4700]: I0227 17:22:19.050691 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4099c554-cf42-48b3-88cb-e672a66f1a05" path="/var/lib/kubelet/pods/4099c554-cf42-48b3-88cb-e672a66f1a05/volumes" Feb 27 17:22:19 crc kubenswrapper[4700]: I0227 17:22:19.051375 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9822c457-a600-4d01-9b94-b7b899ef2102" path="/var/lib/kubelet/pods/9822c457-a600-4d01-9b94-b7b899ef2102/volumes" Feb 27 17:22:19 crc kubenswrapper[4700]: I0227 17:22:19.051897 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1f9992d-c533-4184-9474-1cfe9b3e9819" path="/var/lib/kubelet/pods/d1f9992d-c533-4184-9474-1cfe9b3e9819/volumes" Feb 27 17:22:19 crc kubenswrapper[4700]: I0227 17:22:19.053164 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8e03836-cf5d-4ac9-80fe-97556117e65e" path="/var/lib/kubelet/pods/d8e03836-cf5d-4ac9-80fe-97556117e65e/volumes" Feb 27 17:22:19 crc kubenswrapper[4700]: I0227 17:22:19.078130 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Feb 27 17:22:19 crc kubenswrapper[4700]: I0227 17:22:19.091551 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7dcdff5d6d-9f49t"] Feb 27 17:22:19 crc kubenswrapper[4700]: I0227 17:22:19.224625 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7dcdff5d6d-9f49t" event={"ID":"47d9c3da-3a41-4ed3-93d9-1218f2d7a441","Type":"ContainerStarted","Data":"4d6a7613e5dbc11d372cba1c51876c1136d3ac8e117240b398bf36177c8414c3"} Feb 27 17:22:19 crc kubenswrapper[4700]: I0227 17:22:19.278767 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8946ccf97-n4k7p" event={"ID":"f0d00b91-0e33-4466-8c9f-6a1c1038d23f","Type":"ContainerStarted","Data":"c320d708dcbfe0f89eb170beac5b5ab60bc12dacba3ba4ee98246f53114f457d"} Feb 27 17:22:19 crc kubenswrapper[4700]: I0227 17:22:19.285177 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-8946ccf97-n4k7p" Feb 27 17:22:19 crc kubenswrapper[4700]: I0227 17:22:19.299095 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-756b7f868f-z7qr7"] Feb 27 17:22:19 crc kubenswrapper[4700]: I0227 17:22:19.302663 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4","Type":"ContainerStarted","Data":"7689a23a6c5601dffd3ce204470c772926997a63d771ee14f650d21e1f2d58ea"} Feb 27 17:22:19 crc kubenswrapper[4700]: I0227 17:22:19.318486 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-64fcf8677d-mvkvn" event={"ID":"df366836-b6d3-42a1-ab1d-39f7d77f6cb5","Type":"ContainerStarted","Data":"2cb5bceb568b9c7c3e8fee04bba319c843a556a14b7e1faa3c05e9e9033c865f"} Feb 27 17:22:19 crc kubenswrapper[4700]: I0227 17:22:19.326292 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7f87979786-hh2zw"] Feb 27 17:22:19 crc kubenswrapper[4700]: I0227 17:22:19.335320 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-8946ccf97-n4k7p" podStartSLOduration=8.335304236 podStartE2EDuration="8.335304236s" podCreationTimestamp="2026-02-27 17:22:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:22:19.317080185 +0000 UTC m=+1299.302392932" watchObservedRunningTime="2026-02-27 17:22:19.335304236 +0000 UTC m=+1299.320616983" Feb 27 17:22:19 crc kubenswrapper[4700]: I0227 17:22:19.343669 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-mhcf5" event={"ID":"dd3817a7-c2ab-43f0-9c74-b477ac59fa52","Type":"ContainerStarted","Data":"eba32c7c133cd06b0c577d3b9b1e8c3f433015ecdecbd878fa857cb650a8df98"} Feb 27 17:22:19 crc kubenswrapper[4700]: I0227 17:22:19.356701 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"0f1ae433-28e4-45cf-9635-dc6edbf16f6e","Type":"ContainerStarted","Data":"97555bdd71416290ebc1bd8a3f9aa7230fd3e06b7abf8dddadb69b7494e10c96"} Feb 27 17:22:19 crc kubenswrapper[4700]: I0227 17:22:19.408352 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-mhcf5" podStartSLOduration=5.554917134 podStartE2EDuration="51.408337604s" podCreationTimestamp="2026-02-27 17:21:28 +0000 UTC" firstStartedPulling="2026-02-27 17:21:30.482265272 +0000 UTC m=+1250.467578019" lastFinishedPulling="2026-02-27 17:22:16.335685742 +0000 UTC m=+1296.320998489" observedRunningTime="2026-02-27 17:22:19.385610424 +0000 UTC m=+1299.370923171" watchObservedRunningTime="2026-02-27 17:22:19.408337604 +0000 UTC m=+1299.393650351" Feb 27 17:22:19 crc kubenswrapper[4700]: I0227 17:22:19.414474 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-decision-engine-0" podStartSLOduration=3.414447875 podStartE2EDuration="3.414447875s" podCreationTimestamp="2026-02-27 17:22:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:22:19.402871259 +0000 UTC m=+1299.388184006" watchObservedRunningTime="2026-02-27 17:22:19.414447875 +0000 UTC m=+1299.399760622" Feb 27 17:22:19 crc kubenswrapper[4700]: I0227 17:22:19.525686 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-575575788d-c49lv"] Feb 27 17:22:19 crc kubenswrapper[4700]: I0227 17:22:19.544389 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77475fb77-jzwgk"] Feb 27 17:22:19 crc kubenswrapper[4700]: I0227 17:22:19.864357 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:19.997117 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7744b9dc6c-qnsm2" Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.208814 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc-ovsdbserver-sb\") pod \"5ca33b95-1b7a-4d1c-a3b8-65a5969945bc\" (UID: \"5ca33b95-1b7a-4d1c-a3b8-65a5969945bc\") " Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.208928 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97lrn\" (UniqueName: \"kubernetes.io/projected/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc-kube-api-access-97lrn\") pod \"5ca33b95-1b7a-4d1c-a3b8-65a5969945bc\" (UID: \"5ca33b95-1b7a-4d1c-a3b8-65a5969945bc\") " Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.209008 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc-config\") pod \"5ca33b95-1b7a-4d1c-a3b8-65a5969945bc\" (UID: \"5ca33b95-1b7a-4d1c-a3b8-65a5969945bc\") " Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.209073 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc-ovsdbserver-nb\") pod \"5ca33b95-1b7a-4d1c-a3b8-65a5969945bc\" (UID: \"5ca33b95-1b7a-4d1c-a3b8-65a5969945bc\") " Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.209122 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc-dns-svc\") pod \"5ca33b95-1b7a-4d1c-a3b8-65a5969945bc\" (UID: \"5ca33b95-1b7a-4d1c-a3b8-65a5969945bc\") " Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.209183 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc-dns-swift-storage-0\") pod \"5ca33b95-1b7a-4d1c-a3b8-65a5969945bc\" (UID: \"5ca33b95-1b7a-4d1c-a3b8-65a5969945bc\") " Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.280120 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc-kube-api-access-97lrn" (OuterVolumeSpecName: "kube-api-access-97lrn") pod "5ca33b95-1b7a-4d1c-a3b8-65a5969945bc" (UID: "5ca33b95-1b7a-4d1c-a3b8-65a5969945bc"). InnerVolumeSpecName "kube-api-access-97lrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.311425 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97lrn\" (UniqueName: \"kubernetes.io/projected/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc-kube-api-access-97lrn\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.377310 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-64fcf8677d-mvkvn" event={"ID":"df366836-b6d3-42a1-ab1d-39f7d77f6cb5","Type":"ContainerStarted","Data":"2c9bf816e687505a606f4fd060bda3c436925a4077ffaff3b8c6bf59ff37cda1"} Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.377377 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-64fcf8677d-mvkvn" event={"ID":"df366836-b6d3-42a1-ab1d-39f7d77f6cb5","Type":"ContainerStarted","Data":"f1e4c395d809f499ac4b00dbb1838b8f1f9a5f3addf25b4d24b7c1aa141ea5ce"} Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.377646 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-64fcf8677d-mvkvn" Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.377672 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-64fcf8677d-mvkvn" Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.379040 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"0f1ae433-28e4-45cf-9635-dc6edbf16f6e","Type":"ContainerStarted","Data":"f355943ef237971b6a0197198301ddcc9fa1e170bd952fc11766debe4edbaf6f"} Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.409581 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-64fcf8677d-mvkvn" podStartSLOduration=4.409565637 podStartE2EDuration="4.409565637s" podCreationTimestamp="2026-02-27 17:22:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:22:20.398540176 +0000 UTC m=+1300.383852923" watchObservedRunningTime="2026-02-27 17:22:20.409565637 +0000 UTC m=+1300.394878384" Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.415110 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7744b9dc6c-qnsm2" event={"ID":"5ca33b95-1b7a-4d1c-a3b8-65a5969945bc","Type":"ContainerDied","Data":"0a2214348a1e10b67335e27f4e595d328ec5df3df2658744b07eac983036bf84"} Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.415168 4700 scope.go:117] "RemoveContainer" containerID="d4f0a0eedccf69f640673b202c20c063df1d5604705e8c1a133236b1040464c7" Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.415283 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7744b9dc6c-qnsm2" Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.427711 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"bcc9554e-7836-4db9-aab8-d5b0bef440e1","Type":"ContainerStarted","Data":"47ee76e3ef58d41b1dea9e1705ee3548bd938013999ee9d4c980575d96c1fe21"} Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.436748 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77475fb77-jzwgk" event={"ID":"c48cd283-8c2c-4825-ad5e-c014c1fbc369","Type":"ContainerStarted","Data":"3f3a57688d1e2b9d41e08a9b9af3820237ddf454c09113a4ae620f716dc197b6"} Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.436797 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77475fb77-jzwgk" event={"ID":"c48cd283-8c2c-4825-ad5e-c014c1fbc369","Type":"ContainerStarted","Data":"524cd92e9cc50904b11d98e56cf6e4baa203c67867254cb99fd61ca1cf05d712"} Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.444227 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4","Type":"ContainerStarted","Data":"e1d9ecb97d7cf929a9aa95da731edfb7dc7ac87d203cef3af34b708aba7c33ca"} Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.451151 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-756b7f868f-z7qr7" event={"ID":"1311d698-8af2-4960-9790-f470a67f8457","Type":"ContainerStarted","Data":"d27bf704713aadd71fd91fb4e9681a65cb4c465cabc17305cc114f1848e41e65"} Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.492412 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc-config" (OuterVolumeSpecName: "config") pod "5ca33b95-1b7a-4d1c-a3b8-65a5969945bc" (UID: "5ca33b95-1b7a-4d1c-a3b8-65a5969945bc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.492969 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5ca33b95-1b7a-4d1c-a3b8-65a5969945bc" (UID: "5ca33b95-1b7a-4d1c-a3b8-65a5969945bc"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.499542 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5ca33b95-1b7a-4d1c-a3b8-65a5969945bc" (UID: "5ca33b95-1b7a-4d1c-a3b8-65a5969945bc"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.501644 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7dcdff5d6d-9f49t" event={"ID":"47d9c3da-3a41-4ed3-93d9-1218f2d7a441","Type":"ContainerStarted","Data":"cbba11d693179987f298d04126846971e67068aabab65554b5db7794276f54c7"} Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.503374 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-7dcdff5d6d-9f49t" Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.514860 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.514889 4700 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.514898 4700 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.517737 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5ca33b95-1b7a-4d1c-a3b8-65a5969945bc" (UID: "5ca33b95-1b7a-4d1c-a3b8-65a5969945bc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.519550 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-575575788d-c49lv" event={"ID":"cb6fd0bc-8abd-4537-9c92-f8027dde0bb8","Type":"ContainerStarted","Data":"c2395c26407b65513ecb63305725bb10b527c70bb06752aaf39a362fce871982"} Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.519594 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-575575788d-c49lv" event={"ID":"cb6fd0bc-8abd-4537-9c92-f8027dde0bb8","Type":"ContainerStarted","Data":"8aa71c56cc08c5bc921c28340baee8e30653796709e293f549308a19f1c690bd"} Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.535223 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7f87979786-hh2zw" event={"ID":"3b6c33fa-4855-4b22-a634-cbc739ec0580","Type":"ContainerStarted","Data":"54c18364b782b894ee43c93ce5d2ecae31c44dea12dc2ad0174a75610e8c5113"} Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.545744 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-applier-0" podStartSLOduration=4.545726491 podStartE2EDuration="4.545726491s" podCreationTimestamp="2026-02-27 17:22:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:22:20.492774263 +0000 UTC m=+1300.478087010" watchObservedRunningTime="2026-02-27 17:22:20.545726491 +0000 UTC m=+1300.531039228" Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.549561 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-7dcdff5d6d-9f49t" podStartSLOduration=3.549547922 podStartE2EDuration="3.549547922s" podCreationTimestamp="2026-02-27 17:22:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:22:20.52334355 +0000 UTC m=+1300.508656287" watchObservedRunningTime="2026-02-27 17:22:20.549547922 +0000 UTC m=+1300.534860669" Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.553361 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5ca33b95-1b7a-4d1c-a3b8-65a5969945bc" (UID: "5ca33b95-1b7a-4d1c-a3b8-65a5969945bc"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.616239 4700 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.616269 4700 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.798231 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7744b9dc6c-qnsm2"] Feb 27 17:22:20 crc kubenswrapper[4700]: I0227 17:22:20.821073 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7744b9dc6c-qnsm2"] Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.012093 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ca33b95-1b7a-4d1c-a3b8-65a5969945bc" path="/var/lib/kubelet/pods/5ca33b95-1b7a-4d1c-a3b8-65a5969945bc/volumes" Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.383391 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6cbb9d98-xkx97"] Feb 27 17:22:21 crc kubenswrapper[4700]: E0227 17:22:21.383745 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ca33b95-1b7a-4d1c-a3b8-65a5969945bc" containerName="init" Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.383761 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ca33b95-1b7a-4d1c-a3b8-65a5969945bc" containerName="init" Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.383971 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ca33b95-1b7a-4d1c-a3b8-65a5969945bc" containerName="init" Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.385216 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6cbb9d98-xkx97" Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.405237 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.405561 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.444088 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6cbb9d98-xkx97"] Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.536536 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a9b30341-3783-4498-95ac-0f303a599cce-config-data-custom\") pod \"barbican-api-6cbb9d98-xkx97\" (UID: \"a9b30341-3783-4498-95ac-0f303a599cce\") " pod="openstack/barbican-api-6cbb9d98-xkx97" Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.536879 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9b30341-3783-4498-95ac-0f303a599cce-public-tls-certs\") pod \"barbican-api-6cbb9d98-xkx97\" (UID: \"a9b30341-3783-4498-95ac-0f303a599cce\") " pod="openstack/barbican-api-6cbb9d98-xkx97" Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.536915 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9b30341-3783-4498-95ac-0f303a599cce-logs\") pod \"barbican-api-6cbb9d98-xkx97\" (UID: \"a9b30341-3783-4498-95ac-0f303a599cce\") " pod="openstack/barbican-api-6cbb9d98-xkx97" Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.536951 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9b30341-3783-4498-95ac-0f303a599cce-internal-tls-certs\") pod \"barbican-api-6cbb9d98-xkx97\" (UID: \"a9b30341-3783-4498-95ac-0f303a599cce\") " pod="openstack/barbican-api-6cbb9d98-xkx97" Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.536969 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zxrq\" (UniqueName: \"kubernetes.io/projected/a9b30341-3783-4498-95ac-0f303a599cce-kube-api-access-5zxrq\") pod \"barbican-api-6cbb9d98-xkx97\" (UID: \"a9b30341-3783-4498-95ac-0f303a599cce\") " pod="openstack/barbican-api-6cbb9d98-xkx97" Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.537024 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9b30341-3783-4498-95ac-0f303a599cce-config-data\") pod \"barbican-api-6cbb9d98-xkx97\" (UID: \"a9b30341-3783-4498-95ac-0f303a599cce\") " pod="openstack/barbican-api-6cbb9d98-xkx97" Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.537066 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9b30341-3783-4498-95ac-0f303a599cce-combined-ca-bundle\") pod \"barbican-api-6cbb9d98-xkx97\" (UID: \"a9b30341-3783-4498-95ac-0f303a599cce\") " pod="openstack/barbican-api-6cbb9d98-xkx97" Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.556673 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-575575788d-c49lv" event={"ID":"cb6fd0bc-8abd-4537-9c92-f8027dde0bb8","Type":"ContainerStarted","Data":"802db57c2ee3737a6305579bde290751d01c2bd74d89649757372322d6dc1bec"} Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.557694 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-575575788d-c49lv" Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.557719 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-575575788d-c49lv" Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.559905 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"bcc9554e-7836-4db9-aab8-d5b0bef440e1","Type":"ContainerStarted","Data":"4ccfe5c870e5cdb285edb2596fa13cabf7b4950cdebee135126075a3f1383f21"} Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.559980 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"bcc9554e-7836-4db9-aab8-d5b0bef440e1","Type":"ContainerStarted","Data":"c224457d145e0f2fbebb305640c1cb7af141262e5054691b554ed0b3d3c26da1"} Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.560233 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.564010 4700 generic.go:334] "Generic (PLEG): container finished" podID="c48cd283-8c2c-4825-ad5e-c014c1fbc369" containerID="3f3a57688d1e2b9d41e08a9b9af3820237ddf454c09113a4ae620f716dc197b6" exitCode=0 Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.564055 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77475fb77-jzwgk" event={"ID":"c48cd283-8c2c-4825-ad5e-c014c1fbc369","Type":"ContainerDied","Data":"3f3a57688d1e2b9d41e08a9b9af3820237ddf454c09113a4ae620f716dc197b6"} Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.564072 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77475fb77-jzwgk" event={"ID":"c48cd283-8c2c-4825-ad5e-c014c1fbc369","Type":"ContainerStarted","Data":"466f82136ad1681a5d562884d53c1245dfc13fb1c109b416395fe31d7e466329"} Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.564629 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77475fb77-jzwgk" Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.586999 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-575575788d-c49lv" podStartSLOduration=3.586983841 podStartE2EDuration="3.586983841s" podCreationTimestamp="2026-02-27 17:22:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:22:21.572393956 +0000 UTC m=+1301.557706703" watchObservedRunningTime="2026-02-27 17:22:21.586983841 +0000 UTC m=+1301.572296588" Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.628566 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77475fb77-jzwgk" podStartSLOduration=3.628528908 podStartE2EDuration="3.628528908s" podCreationTimestamp="2026-02-27 17:22:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:22:21.606690362 +0000 UTC m=+1301.592003109" watchObservedRunningTime="2026-02-27 17:22:21.628528908 +0000 UTC m=+1301.613841645" Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.632088 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-api-0" podStartSLOduration=3.632075551 podStartE2EDuration="3.632075551s" podCreationTimestamp="2026-02-27 17:22:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:22:21.623652459 +0000 UTC m=+1301.608965206" watchObservedRunningTime="2026-02-27 17:22:21.632075551 +0000 UTC m=+1301.617388308" Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.638555 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9b30341-3783-4498-95ac-0f303a599cce-config-data\") pod \"barbican-api-6cbb9d98-xkx97\" (UID: \"a9b30341-3783-4498-95ac-0f303a599cce\") " pod="openstack/barbican-api-6cbb9d98-xkx97" Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.638845 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9b30341-3783-4498-95ac-0f303a599cce-combined-ca-bundle\") pod \"barbican-api-6cbb9d98-xkx97\" (UID: \"a9b30341-3783-4498-95ac-0f303a599cce\") " pod="openstack/barbican-api-6cbb9d98-xkx97" Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.639049 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a9b30341-3783-4498-95ac-0f303a599cce-config-data-custom\") pod \"barbican-api-6cbb9d98-xkx97\" (UID: \"a9b30341-3783-4498-95ac-0f303a599cce\") " pod="openstack/barbican-api-6cbb9d98-xkx97" Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.639163 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9b30341-3783-4498-95ac-0f303a599cce-public-tls-certs\") pod \"barbican-api-6cbb9d98-xkx97\" (UID: \"a9b30341-3783-4498-95ac-0f303a599cce\") " pod="openstack/barbican-api-6cbb9d98-xkx97" Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.639341 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9b30341-3783-4498-95ac-0f303a599cce-logs\") pod \"barbican-api-6cbb9d98-xkx97\" (UID: \"a9b30341-3783-4498-95ac-0f303a599cce\") " pod="openstack/barbican-api-6cbb9d98-xkx97" Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.641517 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9b30341-3783-4498-95ac-0f303a599cce-logs\") pod \"barbican-api-6cbb9d98-xkx97\" (UID: \"a9b30341-3783-4498-95ac-0f303a599cce\") " pod="openstack/barbican-api-6cbb9d98-xkx97" Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.642001 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9b30341-3783-4498-95ac-0f303a599cce-internal-tls-certs\") pod \"barbican-api-6cbb9d98-xkx97\" (UID: \"a9b30341-3783-4498-95ac-0f303a599cce\") " pod="openstack/barbican-api-6cbb9d98-xkx97" Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.642054 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zxrq\" (UniqueName: \"kubernetes.io/projected/a9b30341-3783-4498-95ac-0f303a599cce-kube-api-access-5zxrq\") pod \"barbican-api-6cbb9d98-xkx97\" (UID: \"a9b30341-3783-4498-95ac-0f303a599cce\") " pod="openstack/barbican-api-6cbb9d98-xkx97" Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.645350 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9b30341-3783-4498-95ac-0f303a599cce-config-data\") pod \"barbican-api-6cbb9d98-xkx97\" (UID: \"a9b30341-3783-4498-95ac-0f303a599cce\") " pod="openstack/barbican-api-6cbb9d98-xkx97" Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.648341 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9b30341-3783-4498-95ac-0f303a599cce-combined-ca-bundle\") pod \"barbican-api-6cbb9d98-xkx97\" (UID: \"a9b30341-3783-4498-95ac-0f303a599cce\") " pod="openstack/barbican-api-6cbb9d98-xkx97" Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.648861 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a9b30341-3783-4498-95ac-0f303a599cce-config-data-custom\") pod \"barbican-api-6cbb9d98-xkx97\" (UID: \"a9b30341-3783-4498-95ac-0f303a599cce\") " pod="openstack/barbican-api-6cbb9d98-xkx97" Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.650405 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9b30341-3783-4498-95ac-0f303a599cce-internal-tls-certs\") pod \"barbican-api-6cbb9d98-xkx97\" (UID: \"a9b30341-3783-4498-95ac-0f303a599cce\") " pod="openstack/barbican-api-6cbb9d98-xkx97" Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.657308 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9b30341-3783-4498-95ac-0f303a599cce-public-tls-certs\") pod \"barbican-api-6cbb9d98-xkx97\" (UID: \"a9b30341-3783-4498-95ac-0f303a599cce\") " pod="openstack/barbican-api-6cbb9d98-xkx97" Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.658901 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zxrq\" (UniqueName: \"kubernetes.io/projected/a9b30341-3783-4498-95ac-0f303a599cce-kube-api-access-5zxrq\") pod \"barbican-api-6cbb9d98-xkx97\" (UID: \"a9b30341-3783-4498-95ac-0f303a599cce\") " pod="openstack/barbican-api-6cbb9d98-xkx97" Feb 27 17:22:21 crc kubenswrapper[4700]: I0227 17:22:21.728315 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6cbb9d98-xkx97" Feb 27 17:22:22 crc kubenswrapper[4700]: I0227 17:22:22.851440 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-applier-0" Feb 27 17:22:23 crc kubenswrapper[4700]: I0227 17:22:23.411446 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6cbb9d98-xkx97"] Feb 27 17:22:23 crc kubenswrapper[4700]: W0227 17:22:23.412762 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda9b30341_3783_4498_95ac_0f303a599cce.slice/crio-180142a4b779727a4d6c7c56ebe45497f0c9e92457722f1d0c9cf1dd8636e490 WatchSource:0}: Error finding container 180142a4b779727a4d6c7c56ebe45497f0c9e92457722f1d0c9cf1dd8636e490: Status 404 returned error can't find the container with id 180142a4b779727a4d6c7c56ebe45497f0c9e92457722f1d0c9cf1dd8636e490 Feb 27 17:22:23 crc kubenswrapper[4700]: I0227 17:22:23.669726 4700 generic.go:334] "Generic (PLEG): container finished" podID="0f1ae433-28e4-45cf-9635-dc6edbf16f6e" containerID="f355943ef237971b6a0197198301ddcc9fa1e170bd952fc11766debe4edbaf6f" exitCode=1 Feb 27 17:22:23 crc kubenswrapper[4700]: I0227 17:22:23.673579 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"0f1ae433-28e4-45cf-9635-dc6edbf16f6e","Type":"ContainerDied","Data":"f355943ef237971b6a0197198301ddcc9fa1e170bd952fc11766debe4edbaf6f"} Feb 27 17:22:23 crc kubenswrapper[4700]: I0227 17:22:23.674735 4700 scope.go:117] "RemoveContainer" containerID="f355943ef237971b6a0197198301ddcc9fa1e170bd952fc11766debe4edbaf6f" Feb 27 17:22:23 crc kubenswrapper[4700]: I0227 17:22:23.676076 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-756b7f868f-z7qr7" event={"ID":"1311d698-8af2-4960-9790-f470a67f8457","Type":"ContainerStarted","Data":"04e41a4c39645f34a5bda6de47cb0254480a441109dc3840fc55b70368584399"} Feb 27 17:22:23 crc kubenswrapper[4700]: I0227 17:22:23.676130 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-756b7f868f-z7qr7" event={"ID":"1311d698-8af2-4960-9790-f470a67f8457","Type":"ContainerStarted","Data":"a4561a4e7b8a2ca150bfeb8eef33af4a59355202f945fb7afe28f6a964c053f3"} Feb 27 17:22:23 crc kubenswrapper[4700]: I0227 17:22:23.682196 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6cbb9d98-xkx97" event={"ID":"a9b30341-3783-4498-95ac-0f303a599cce","Type":"ContainerStarted","Data":"dfdcad740e2e64bd366490bc2103ad3b71d897fdfb85ed75f8b517e9d3c0e8ca"} Feb 27 17:22:23 crc kubenswrapper[4700]: I0227 17:22:23.682254 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6cbb9d98-xkx97" event={"ID":"a9b30341-3783-4498-95ac-0f303a599cce","Type":"ContainerStarted","Data":"180142a4b779727a4d6c7c56ebe45497f0c9e92457722f1d0c9cf1dd8636e490"} Feb 27 17:22:23 crc kubenswrapper[4700]: I0227 17:22:23.697712 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7f87979786-hh2zw" event={"ID":"3b6c33fa-4855-4b22-a634-cbc739ec0580","Type":"ContainerStarted","Data":"b408da5ae9a02b5f560eb53471eec6435369f965efefd4c18c8e758201931e0b"} Feb 27 17:22:23 crc kubenswrapper[4700]: I0227 17:22:23.697803 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7f87979786-hh2zw" event={"ID":"3b6c33fa-4855-4b22-a634-cbc739ec0580","Type":"ContainerStarted","Data":"3593514d7ffd7f9ce13d9aa77cd564093743570dc7487c18aa4b069cec9a0060"} Feb 27 17:22:23 crc kubenswrapper[4700]: I0227 17:22:23.718393 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-7f87979786-hh2zw" podStartSLOduration=3.15764868 podStartE2EDuration="6.718370583s" podCreationTimestamp="2026-02-27 17:22:17 +0000 UTC" firstStartedPulling="2026-02-27 17:22:19.384991338 +0000 UTC m=+1299.370304085" lastFinishedPulling="2026-02-27 17:22:22.945713241 +0000 UTC m=+1302.931025988" observedRunningTime="2026-02-27 17:22:23.714051959 +0000 UTC m=+1303.699364726" watchObservedRunningTime="2026-02-27 17:22:23.718370583 +0000 UTC m=+1303.703683350" Feb 27 17:22:23 crc kubenswrapper[4700]: I0227 17:22:23.913336 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Feb 27 17:22:23 crc kubenswrapper[4700]: I0227 17:22:23.937805 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-756b7f868f-z7qr7" podStartSLOduration=3.351188627 podStartE2EDuration="6.937790673s" podCreationTimestamp="2026-02-27 17:22:17 +0000 UTC" firstStartedPulling="2026-02-27 17:22:19.357082881 +0000 UTC m=+1299.342395628" lastFinishedPulling="2026-02-27 17:22:22.943684927 +0000 UTC m=+1302.928997674" observedRunningTime="2026-02-27 17:22:23.742139959 +0000 UTC m=+1303.727452706" watchObservedRunningTime="2026-02-27 17:22:23.937790673 +0000 UTC m=+1303.923103420" Feb 27 17:22:24 crc kubenswrapper[4700]: I0227 17:22:24.079168 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Feb 27 17:22:24 crc kubenswrapper[4700]: I0227 17:22:24.708403 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6cbb9d98-xkx97" event={"ID":"a9b30341-3783-4498-95ac-0f303a599cce","Type":"ContainerStarted","Data":"fe8df4bc0a9fcc364eec5b0a3ad5852c152a8afb2410a41f1e307458559ffa42"} Feb 27 17:22:24 crc kubenswrapper[4700]: I0227 17:22:24.709507 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6cbb9d98-xkx97" Feb 27 17:22:24 crc kubenswrapper[4700]: I0227 17:22:24.709536 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6cbb9d98-xkx97" Feb 27 17:22:24 crc kubenswrapper[4700]: I0227 17:22:24.711157 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"0f1ae433-28e4-45cf-9635-dc6edbf16f6e","Type":"ContainerStarted","Data":"6c8960b7fce9f293dfe5ac76c658c4a8af1e7bbe2abfc614cc1caccc7e49b99a"} Feb 27 17:22:24 crc kubenswrapper[4700]: I0227 17:22:24.742421 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6cbb9d98-xkx97" podStartSLOduration=3.742403038 podStartE2EDuration="3.742403038s" podCreationTimestamp="2026-02-27 17:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:22:24.733163134 +0000 UTC m=+1304.718475881" watchObservedRunningTime="2026-02-27 17:22:24.742403038 +0000 UTC m=+1304.727715785" Feb 27 17:22:26 crc kubenswrapper[4700]: I0227 17:22:26.733453 4700 generic.go:334] "Generic (PLEG): container finished" podID="0f1ae433-28e4-45cf-9635-dc6edbf16f6e" containerID="6c8960b7fce9f293dfe5ac76c658c4a8af1e7bbe2abfc614cc1caccc7e49b99a" exitCode=1 Feb 27 17:22:26 crc kubenswrapper[4700]: I0227 17:22:26.733495 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"0f1ae433-28e4-45cf-9635-dc6edbf16f6e","Type":"ContainerDied","Data":"6c8960b7fce9f293dfe5ac76c658c4a8af1e7bbe2abfc614cc1caccc7e49b99a"} Feb 27 17:22:26 crc kubenswrapper[4700]: I0227 17:22:26.733908 4700 scope.go:117] "RemoveContainer" containerID="f355943ef237971b6a0197198301ddcc9fa1e170bd952fc11766debe4edbaf6f" Feb 27 17:22:26 crc kubenswrapper[4700]: I0227 17:22:26.735297 4700 scope.go:117] "RemoveContainer" containerID="6c8960b7fce9f293dfe5ac76c658c4a8af1e7bbe2abfc614cc1caccc7e49b99a" Feb 27 17:22:26 crc kubenswrapper[4700]: E0227 17:22:26.769501 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 10s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(0f1ae433-28e4-45cf-9635-dc6edbf16f6e)\"" pod="openstack/watcher-decision-engine-0" podUID="0f1ae433-28e4-45cf-9635-dc6edbf16f6e" Feb 27 17:22:26 crc kubenswrapper[4700]: I0227 17:22:26.928905 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 27 17:22:26 crc kubenswrapper[4700]: I0227 17:22:26.929086 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 27 17:22:26 crc kubenswrapper[4700]: I0227 17:22:26.980181 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 27 17:22:26 crc kubenswrapper[4700]: I0227 17:22:26.994829 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 27 17:22:27 crc kubenswrapper[4700]: I0227 17:22:27.273286 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 27 17:22:27 crc kubenswrapper[4700]: I0227 17:22:27.273566 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 27 17:22:27 crc kubenswrapper[4700]: I0227 17:22:27.324777 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 27 17:22:27 crc kubenswrapper[4700]: I0227 17:22:27.332725 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 27 17:22:27 crc kubenswrapper[4700]: I0227 17:22:27.747859 4700 generic.go:334] "Generic (PLEG): container finished" podID="dd3817a7-c2ab-43f0-9c74-b477ac59fa52" containerID="eba32c7c133cd06b0c577d3b9b1e8c3f433015ecdecbd878fa857cb650a8df98" exitCode=0 Feb 27 17:22:27 crc kubenswrapper[4700]: I0227 17:22:27.749012 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-mhcf5" event={"ID":"dd3817a7-c2ab-43f0-9c74-b477ac59fa52","Type":"ContainerDied","Data":"eba32c7c133cd06b0c577d3b9b1e8c3f433015ecdecbd878fa857cb650a8df98"} Feb 27 17:22:27 crc kubenswrapper[4700]: I0227 17:22:27.749088 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 27 17:22:27 crc kubenswrapper[4700]: I0227 17:22:27.749106 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 27 17:22:27 crc kubenswrapper[4700]: I0227 17:22:27.749119 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 27 17:22:27 crc kubenswrapper[4700]: I0227 17:22:27.749159 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 27 17:22:27 crc kubenswrapper[4700]: I0227 17:22:27.850532 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-applier-0" Feb 27 17:22:27 crc kubenswrapper[4700]: I0227 17:22:27.857400 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Feb 27 17:22:27 crc kubenswrapper[4700]: I0227 17:22:27.857438 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Feb 27 17:22:27 crc kubenswrapper[4700]: I0227 17:22:27.858155 4700 scope.go:117] "RemoveContainer" containerID="6c8960b7fce9f293dfe5ac76c658c4a8af1e7bbe2abfc614cc1caccc7e49b99a" Feb 27 17:22:27 crc kubenswrapper[4700]: E0227 17:22:27.858380 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 10s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(0f1ae433-28e4-45cf-9635-dc6edbf16f6e)\"" pod="openstack/watcher-decision-engine-0" podUID="0f1ae433-28e4-45cf-9635-dc6edbf16f6e" Feb 27 17:22:27 crc kubenswrapper[4700]: I0227 17:22:27.893712 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-applier-0" Feb 27 17:22:28 crc kubenswrapper[4700]: I0227 17:22:28.816304 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-applier-0" Feb 27 17:22:28 crc kubenswrapper[4700]: I0227 17:22:28.848883 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77475fb77-jzwgk" Feb 27 17:22:28 crc kubenswrapper[4700]: I0227 17:22:28.925038 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-787484f569-fhzcg"] Feb 27 17:22:28 crc kubenswrapper[4700]: I0227 17:22:28.925249 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-787484f569-fhzcg" podUID="7d4f9f19-62e5-4bcd-957e-345c18380eac" containerName="dnsmasq-dns" containerID="cri-o://7c8c08b9bd26fb61dd0d15098608d526d6a4041994ea5684b4d03ddc15abe370" gracePeriod=10 Feb 27 17:22:29 crc kubenswrapper[4700]: I0227 17:22:29.079479 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-api-0" Feb 27 17:22:29 crc kubenswrapper[4700]: I0227 17:22:29.103870 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-api-0" Feb 27 17:22:29 crc kubenswrapper[4700]: I0227 17:22:29.776365 4700 generic.go:334] "Generic (PLEG): container finished" podID="7d4f9f19-62e5-4bcd-957e-345c18380eac" containerID="7c8c08b9bd26fb61dd0d15098608d526d6a4041994ea5684b4d03ddc15abe370" exitCode=0 Feb 27 17:22:29 crc kubenswrapper[4700]: I0227 17:22:29.776524 4700 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 27 17:22:29 crc kubenswrapper[4700]: I0227 17:22:29.776534 4700 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 27 17:22:29 crc kubenswrapper[4700]: I0227 17:22:29.776596 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-787484f569-fhzcg" event={"ID":"7d4f9f19-62e5-4bcd-957e-345c18380eac","Type":"ContainerDied","Data":"7c8c08b9bd26fb61dd0d15098608d526d6a4041994ea5684b4d03ddc15abe370"} Feb 27 17:22:29 crc kubenswrapper[4700]: I0227 17:22:29.776716 4700 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 27 17:22:29 crc kubenswrapper[4700]: I0227 17:22:29.776737 4700 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 27 17:22:29 crc kubenswrapper[4700]: I0227 17:22:29.793423 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Feb 27 17:22:29 crc kubenswrapper[4700]: I0227 17:22:29.844056 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-787484f569-fhzcg" podUID="7d4f9f19-62e5-4bcd-957e-345c18380eac" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.165:5353: connect: connection refused" Feb 27 17:22:30 crc kubenswrapper[4700]: I0227 17:22:30.537938 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-575575788d-c49lv" Feb 27 17:22:30 crc kubenswrapper[4700]: I0227 17:22:30.612242 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-5b48777cdd-tqbkj" Feb 27 17:22:30 crc kubenswrapper[4700]: I0227 17:22:30.672057 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6cc494ff8-ggq9w" Feb 27 17:22:30 crc kubenswrapper[4700]: I0227 17:22:30.745912 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-575575788d-c49lv" Feb 27 17:22:32 crc kubenswrapper[4700]: I0227 17:22:32.038734 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-mhcf5" Feb 27 17:22:32 crc kubenswrapper[4700]: I0227 17:22:32.181594 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9nrcq\" (UniqueName: \"kubernetes.io/projected/dd3817a7-c2ab-43f0-9c74-b477ac59fa52-kube-api-access-9nrcq\") pod \"dd3817a7-c2ab-43f0-9c74-b477ac59fa52\" (UID: \"dd3817a7-c2ab-43f0-9c74-b477ac59fa52\") " Feb 27 17:22:32 crc kubenswrapper[4700]: I0227 17:22:32.181630 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dd3817a7-c2ab-43f0-9c74-b477ac59fa52-etc-machine-id\") pod \"dd3817a7-c2ab-43f0-9c74-b477ac59fa52\" (UID: \"dd3817a7-c2ab-43f0-9c74-b477ac59fa52\") " Feb 27 17:22:32 crc kubenswrapper[4700]: I0227 17:22:32.181668 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd3817a7-c2ab-43f0-9c74-b477ac59fa52-config-data\") pod \"dd3817a7-c2ab-43f0-9c74-b477ac59fa52\" (UID: \"dd3817a7-c2ab-43f0-9c74-b477ac59fa52\") " Feb 27 17:22:32 crc kubenswrapper[4700]: I0227 17:22:32.181709 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd3817a7-c2ab-43f0-9c74-b477ac59fa52-scripts\") pod \"dd3817a7-c2ab-43f0-9c74-b477ac59fa52\" (UID: \"dd3817a7-c2ab-43f0-9c74-b477ac59fa52\") " Feb 27 17:22:32 crc kubenswrapper[4700]: I0227 17:22:32.181794 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dd3817a7-c2ab-43f0-9c74-b477ac59fa52-db-sync-config-data\") pod \"dd3817a7-c2ab-43f0-9c74-b477ac59fa52\" (UID: \"dd3817a7-c2ab-43f0-9c74-b477ac59fa52\") " Feb 27 17:22:32 crc kubenswrapper[4700]: I0227 17:22:32.181839 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd3817a7-c2ab-43f0-9c74-b477ac59fa52-combined-ca-bundle\") pod \"dd3817a7-c2ab-43f0-9c74-b477ac59fa52\" (UID: \"dd3817a7-c2ab-43f0-9c74-b477ac59fa52\") " Feb 27 17:22:32 crc kubenswrapper[4700]: I0227 17:22:32.187803 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dd3817a7-c2ab-43f0-9c74-b477ac59fa52-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "dd3817a7-c2ab-43f0-9c74-b477ac59fa52" (UID: "dd3817a7-c2ab-43f0-9c74-b477ac59fa52"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:22:32 crc kubenswrapper[4700]: I0227 17:22:32.190693 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd3817a7-c2ab-43f0-9c74-b477ac59fa52-kube-api-access-9nrcq" (OuterVolumeSpecName: "kube-api-access-9nrcq") pod "dd3817a7-c2ab-43f0-9c74-b477ac59fa52" (UID: "dd3817a7-c2ab-43f0-9c74-b477ac59fa52"). InnerVolumeSpecName "kube-api-access-9nrcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:22:32 crc kubenswrapper[4700]: I0227 17:22:32.197124 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd3817a7-c2ab-43f0-9c74-b477ac59fa52-scripts" (OuterVolumeSpecName: "scripts") pod "dd3817a7-c2ab-43f0-9c74-b477ac59fa52" (UID: "dd3817a7-c2ab-43f0-9c74-b477ac59fa52"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:32 crc kubenswrapper[4700]: I0227 17:22:32.210098 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd3817a7-c2ab-43f0-9c74-b477ac59fa52-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "dd3817a7-c2ab-43f0-9c74-b477ac59fa52" (UID: "dd3817a7-c2ab-43f0-9c74-b477ac59fa52"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:32 crc kubenswrapper[4700]: I0227 17:22:32.230534 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd3817a7-c2ab-43f0-9c74-b477ac59fa52-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dd3817a7-c2ab-43f0-9c74-b477ac59fa52" (UID: "dd3817a7-c2ab-43f0-9c74-b477ac59fa52"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:32 crc kubenswrapper[4700]: I0227 17:22:32.269636 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd3817a7-c2ab-43f0-9c74-b477ac59fa52-config-data" (OuterVolumeSpecName: "config-data") pod "dd3817a7-c2ab-43f0-9c74-b477ac59fa52" (UID: "dd3817a7-c2ab-43f0-9c74-b477ac59fa52"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:32 crc kubenswrapper[4700]: I0227 17:22:32.285363 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd3817a7-c2ab-43f0-9c74-b477ac59fa52-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:32 crc kubenswrapper[4700]: I0227 17:22:32.285397 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9nrcq\" (UniqueName: \"kubernetes.io/projected/dd3817a7-c2ab-43f0-9c74-b477ac59fa52-kube-api-access-9nrcq\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:32 crc kubenswrapper[4700]: I0227 17:22:32.285411 4700 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dd3817a7-c2ab-43f0-9c74-b477ac59fa52-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:32 crc kubenswrapper[4700]: I0227 17:22:32.285420 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd3817a7-c2ab-43f0-9c74-b477ac59fa52-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:32 crc kubenswrapper[4700]: I0227 17:22:32.285429 4700 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd3817a7-c2ab-43f0-9c74-b477ac59fa52-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:32 crc kubenswrapper[4700]: I0227 17:22:32.285439 4700 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dd3817a7-c2ab-43f0-9c74-b477ac59fa52-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:32 crc kubenswrapper[4700]: I0227 17:22:32.617869 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 27 17:22:32 crc kubenswrapper[4700]: I0227 17:22:32.618197 4700 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 27 17:22:32 crc kubenswrapper[4700]: I0227 17:22:32.761022 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6cc494ff8-ggq9w" Feb 27 17:22:32 crc kubenswrapper[4700]: I0227 17:22:32.835000 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-mhcf5" event={"ID":"dd3817a7-c2ab-43f0-9c74-b477ac59fa52","Type":"ContainerDied","Data":"123d2839568692618dc1e50207e5dd1af0d3f189f8dbd11fc2e0f38e7c894202"} Feb 27 17:22:32 crc kubenswrapper[4700]: I0227 17:22:32.835039 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="123d2839568692618dc1e50207e5dd1af0d3f189f8dbd11fc2e0f38e7c894202" Feb 27 17:22:32 crc kubenswrapper[4700]: I0227 17:22:32.835094 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-mhcf5" Feb 27 17:22:32 crc kubenswrapper[4700]: I0227 17:22:32.840027 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5b48777cdd-tqbkj"] Feb 27 17:22:32 crc kubenswrapper[4700]: I0227 17:22:32.840206 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5b48777cdd-tqbkj" podUID="d3989a3e-45ff-45d8-80a2-4dbe57663a57" containerName="horizon-log" containerID="cri-o://2b6e983100e515e7297f4f273adfbef4735de5dcf00d127f6ac2101806ecb734" gracePeriod=30 Feb 27 17:22:32 crc kubenswrapper[4700]: I0227 17:22:32.840260 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5b48777cdd-tqbkj" podUID="d3989a3e-45ff-45d8-80a2-4dbe57663a57" containerName="horizon" containerID="cri-o://cf749444d7c76d9317d18a0299503d9ccecc0cfd7ec5ffd81bd3d244886124dc" gracePeriod=30 Feb 27 17:22:32 crc kubenswrapper[4700]: I0227 17:22:32.929357 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 27 17:22:32 crc kubenswrapper[4700]: I0227 17:22:32.929478 4700 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 27 17:22:32 crc kubenswrapper[4700]: I0227 17:22:32.930399 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 27 17:22:32 crc kubenswrapper[4700]: I0227 17:22:32.938663 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5b48777cdd-tqbkj" podUID="d3989a3e-45ff-45d8-80a2-4dbe57663a57" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.168:8443/dashboard/auth/login/?next=/dashboard/\": EOF" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.301884 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.329999 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 27 17:22:33 crc kubenswrapper[4700]: E0227 17:22:33.330466 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd3817a7-c2ab-43f0-9c74-b477ac59fa52" containerName="cinder-db-sync" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.330482 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd3817a7-c2ab-43f0-9c74-b477ac59fa52" containerName="cinder-db-sync" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.330672 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd3817a7-c2ab-43f0-9c74-b477ac59fa52" containerName="cinder-db-sync" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.331715 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.335366 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.335557 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.335710 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-wx827" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.338799 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.341444 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.416733 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e365b3f6-5e7d-4911-b7ee-728ca3d413dd-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e365b3f6-5e7d-4911-b7ee-728ca3d413dd\") " pod="openstack/cinder-scheduler-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.416842 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e365b3f6-5e7d-4911-b7ee-728ca3d413dd-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e365b3f6-5e7d-4911-b7ee-728ca3d413dd\") " pod="openstack/cinder-scheduler-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.416882 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e365b3f6-5e7d-4911-b7ee-728ca3d413dd-scripts\") pod \"cinder-scheduler-0\" (UID: \"e365b3f6-5e7d-4911-b7ee-728ca3d413dd\") " pod="openstack/cinder-scheduler-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.416903 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e365b3f6-5e7d-4911-b7ee-728ca3d413dd-config-data\") pod \"cinder-scheduler-0\" (UID: \"e365b3f6-5e7d-4911-b7ee-728ca3d413dd\") " pod="openstack/cinder-scheduler-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.416931 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e365b3f6-5e7d-4911-b7ee-728ca3d413dd-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e365b3f6-5e7d-4911-b7ee-728ca3d413dd\") " pod="openstack/cinder-scheduler-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.417124 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5dbk\" (UniqueName: \"kubernetes.io/projected/e365b3f6-5e7d-4911-b7ee-728ca3d413dd-kube-api-access-f5dbk\") pod \"cinder-scheduler-0\" (UID: \"e365b3f6-5e7d-4911-b7ee-728ca3d413dd\") " pod="openstack/cinder-scheduler-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.431453 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7b6dff65bc-t5czk"] Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.433294 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b6dff65bc-t5czk" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.461216 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b6dff65bc-t5czk"] Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.518703 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e365b3f6-5e7d-4911-b7ee-728ca3d413dd-scripts\") pod \"cinder-scheduler-0\" (UID: \"e365b3f6-5e7d-4911-b7ee-728ca3d413dd\") " pod="openstack/cinder-scheduler-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.519005 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e365b3f6-5e7d-4911-b7ee-728ca3d413dd-config-data\") pod \"cinder-scheduler-0\" (UID: \"e365b3f6-5e7d-4911-b7ee-728ca3d413dd\") " pod="openstack/cinder-scheduler-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.519037 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e365b3f6-5e7d-4911-b7ee-728ca3d413dd-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e365b3f6-5e7d-4911-b7ee-728ca3d413dd\") " pod="openstack/cinder-scheduler-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.519078 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kxtp\" (UniqueName: \"kubernetes.io/projected/18c2be27-ec7d-4d27-a61b-8c0c5d332ebe-kube-api-access-2kxtp\") pod \"dnsmasq-dns-7b6dff65bc-t5czk\" (UID: \"18c2be27-ec7d-4d27-a61b-8c0c5d332ebe\") " pod="openstack/dnsmasq-dns-7b6dff65bc-t5czk" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.519126 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5dbk\" (UniqueName: \"kubernetes.io/projected/e365b3f6-5e7d-4911-b7ee-728ca3d413dd-kube-api-access-f5dbk\") pod \"cinder-scheduler-0\" (UID: \"e365b3f6-5e7d-4911-b7ee-728ca3d413dd\") " pod="openstack/cinder-scheduler-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.519146 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/18c2be27-ec7d-4d27-a61b-8c0c5d332ebe-dns-svc\") pod \"dnsmasq-dns-7b6dff65bc-t5czk\" (UID: \"18c2be27-ec7d-4d27-a61b-8c0c5d332ebe\") " pod="openstack/dnsmasq-dns-7b6dff65bc-t5czk" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.519195 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18c2be27-ec7d-4d27-a61b-8c0c5d332ebe-config\") pod \"dnsmasq-dns-7b6dff65bc-t5czk\" (UID: \"18c2be27-ec7d-4d27-a61b-8c0c5d332ebe\") " pod="openstack/dnsmasq-dns-7b6dff65bc-t5czk" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.519214 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/18c2be27-ec7d-4d27-a61b-8c0c5d332ebe-ovsdbserver-sb\") pod \"dnsmasq-dns-7b6dff65bc-t5czk\" (UID: \"18c2be27-ec7d-4d27-a61b-8c0c5d332ebe\") " pod="openstack/dnsmasq-dns-7b6dff65bc-t5czk" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.519267 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/18c2be27-ec7d-4d27-a61b-8c0c5d332ebe-ovsdbserver-nb\") pod \"dnsmasq-dns-7b6dff65bc-t5czk\" (UID: \"18c2be27-ec7d-4d27-a61b-8c0c5d332ebe\") " pod="openstack/dnsmasq-dns-7b6dff65bc-t5czk" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.519290 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e365b3f6-5e7d-4911-b7ee-728ca3d413dd-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e365b3f6-5e7d-4911-b7ee-728ca3d413dd\") " pod="openstack/cinder-scheduler-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.519305 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/18c2be27-ec7d-4d27-a61b-8c0c5d332ebe-dns-swift-storage-0\") pod \"dnsmasq-dns-7b6dff65bc-t5czk\" (UID: \"18c2be27-ec7d-4d27-a61b-8c0c5d332ebe\") " pod="openstack/dnsmasq-dns-7b6dff65bc-t5czk" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.519343 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e365b3f6-5e7d-4911-b7ee-728ca3d413dd-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e365b3f6-5e7d-4911-b7ee-728ca3d413dd\") " pod="openstack/cinder-scheduler-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.520491 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e365b3f6-5e7d-4911-b7ee-728ca3d413dd-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e365b3f6-5e7d-4911-b7ee-728ca3d413dd\") " pod="openstack/cinder-scheduler-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.527272 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e365b3f6-5e7d-4911-b7ee-728ca3d413dd-scripts\") pod \"cinder-scheduler-0\" (UID: \"e365b3f6-5e7d-4911-b7ee-728ca3d413dd\") " pod="openstack/cinder-scheduler-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.527497 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e365b3f6-5e7d-4911-b7ee-728ca3d413dd-config-data\") pod \"cinder-scheduler-0\" (UID: \"e365b3f6-5e7d-4911-b7ee-728ca3d413dd\") " pod="openstack/cinder-scheduler-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.553502 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e365b3f6-5e7d-4911-b7ee-728ca3d413dd-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e365b3f6-5e7d-4911-b7ee-728ca3d413dd\") " pod="openstack/cinder-scheduler-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.563143 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5dbk\" (UniqueName: \"kubernetes.io/projected/e365b3f6-5e7d-4911-b7ee-728ca3d413dd-kube-api-access-f5dbk\") pod \"cinder-scheduler-0\" (UID: \"e365b3f6-5e7d-4911-b7ee-728ca3d413dd\") " pod="openstack/cinder-scheduler-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.576837 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.578631 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.580002 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e365b3f6-5e7d-4911-b7ee-728ca3d413dd-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e365b3f6-5e7d-4911-b7ee-728ca3d413dd\") " pod="openstack/cinder-scheduler-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.613806 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.620506 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.621517 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/18c2be27-ec7d-4d27-a61b-8c0c5d332ebe-dns-swift-storage-0\") pod \"dnsmasq-dns-7b6dff65bc-t5czk\" (UID: \"18c2be27-ec7d-4d27-a61b-8c0c5d332ebe\") " pod="openstack/dnsmasq-dns-7b6dff65bc-t5czk" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.621644 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/800523ee-3cb6-4635-9384-19650e9dc2a7-config-data\") pod \"cinder-api-0\" (UID: \"800523ee-3cb6-4635-9384-19650e9dc2a7\") " pod="openstack/cinder-api-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.621677 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/800523ee-3cb6-4635-9384-19650e9dc2a7-logs\") pod \"cinder-api-0\" (UID: \"800523ee-3cb6-4635-9384-19650e9dc2a7\") " pod="openstack/cinder-api-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.621720 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kxtp\" (UniqueName: \"kubernetes.io/projected/18c2be27-ec7d-4d27-a61b-8c0c5d332ebe-kube-api-access-2kxtp\") pod \"dnsmasq-dns-7b6dff65bc-t5czk\" (UID: \"18c2be27-ec7d-4d27-a61b-8c0c5d332ebe\") " pod="openstack/dnsmasq-dns-7b6dff65bc-t5czk" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.621746 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/800523ee-3cb6-4635-9384-19650e9dc2a7-scripts\") pod \"cinder-api-0\" (UID: \"800523ee-3cb6-4635-9384-19650e9dc2a7\") " pod="openstack/cinder-api-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.621764 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/800523ee-3cb6-4635-9384-19650e9dc2a7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"800523ee-3cb6-4635-9384-19650e9dc2a7\") " pod="openstack/cinder-api-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.621797 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/18c2be27-ec7d-4d27-a61b-8c0c5d332ebe-dns-svc\") pod \"dnsmasq-dns-7b6dff65bc-t5czk\" (UID: \"18c2be27-ec7d-4d27-a61b-8c0c5d332ebe\") " pod="openstack/dnsmasq-dns-7b6dff65bc-t5czk" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.621826 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18c2be27-ec7d-4d27-a61b-8c0c5d332ebe-config\") pod \"dnsmasq-dns-7b6dff65bc-t5czk\" (UID: \"18c2be27-ec7d-4d27-a61b-8c0c5d332ebe\") " pod="openstack/dnsmasq-dns-7b6dff65bc-t5czk" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.621843 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/18c2be27-ec7d-4d27-a61b-8c0c5d332ebe-ovsdbserver-sb\") pod \"dnsmasq-dns-7b6dff65bc-t5czk\" (UID: \"18c2be27-ec7d-4d27-a61b-8c0c5d332ebe\") " pod="openstack/dnsmasq-dns-7b6dff65bc-t5czk" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.621870 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/800523ee-3cb6-4635-9384-19650e9dc2a7-config-data-custom\") pod \"cinder-api-0\" (UID: \"800523ee-3cb6-4635-9384-19650e9dc2a7\") " pod="openstack/cinder-api-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.621888 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lt6cn\" (UniqueName: \"kubernetes.io/projected/800523ee-3cb6-4635-9384-19650e9dc2a7-kube-api-access-lt6cn\") pod \"cinder-api-0\" (UID: \"800523ee-3cb6-4635-9384-19650e9dc2a7\") " pod="openstack/cinder-api-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.621917 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/800523ee-3cb6-4635-9384-19650e9dc2a7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"800523ee-3cb6-4635-9384-19650e9dc2a7\") " pod="openstack/cinder-api-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.621942 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/18c2be27-ec7d-4d27-a61b-8c0c5d332ebe-ovsdbserver-nb\") pod \"dnsmasq-dns-7b6dff65bc-t5czk\" (UID: \"18c2be27-ec7d-4d27-a61b-8c0c5d332ebe\") " pod="openstack/dnsmasq-dns-7b6dff65bc-t5czk" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.622865 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/18c2be27-ec7d-4d27-a61b-8c0c5d332ebe-ovsdbserver-nb\") pod \"dnsmasq-dns-7b6dff65bc-t5czk\" (UID: \"18c2be27-ec7d-4d27-a61b-8c0c5d332ebe\") " pod="openstack/dnsmasq-dns-7b6dff65bc-t5czk" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.623375 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/18c2be27-ec7d-4d27-a61b-8c0c5d332ebe-dns-swift-storage-0\") pod \"dnsmasq-dns-7b6dff65bc-t5czk\" (UID: \"18c2be27-ec7d-4d27-a61b-8c0c5d332ebe\") " pod="openstack/dnsmasq-dns-7b6dff65bc-t5czk" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.626213 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18c2be27-ec7d-4d27-a61b-8c0c5d332ebe-config\") pod \"dnsmasq-dns-7b6dff65bc-t5czk\" (UID: \"18c2be27-ec7d-4d27-a61b-8c0c5d332ebe\") " pod="openstack/dnsmasq-dns-7b6dff65bc-t5czk" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.626992 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/18c2be27-ec7d-4d27-a61b-8c0c5d332ebe-ovsdbserver-sb\") pod \"dnsmasq-dns-7b6dff65bc-t5czk\" (UID: \"18c2be27-ec7d-4d27-a61b-8c0c5d332ebe\") " pod="openstack/dnsmasq-dns-7b6dff65bc-t5czk" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.631431 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/18c2be27-ec7d-4d27-a61b-8c0c5d332ebe-dns-svc\") pod \"dnsmasq-dns-7b6dff65bc-t5czk\" (UID: \"18c2be27-ec7d-4d27-a61b-8c0c5d332ebe\") " pod="openstack/dnsmasq-dns-7b6dff65bc-t5czk" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.659018 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kxtp\" (UniqueName: \"kubernetes.io/projected/18c2be27-ec7d-4d27-a61b-8c0c5d332ebe-kube-api-access-2kxtp\") pod \"dnsmasq-dns-7b6dff65bc-t5czk\" (UID: \"18c2be27-ec7d-4d27-a61b-8c0c5d332ebe\") " pod="openstack/dnsmasq-dns-7b6dff65bc-t5czk" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.687302 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.726410 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/800523ee-3cb6-4635-9384-19650e9dc2a7-config-data-custom\") pod \"cinder-api-0\" (UID: \"800523ee-3cb6-4635-9384-19650e9dc2a7\") " pod="openstack/cinder-api-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.726444 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lt6cn\" (UniqueName: \"kubernetes.io/projected/800523ee-3cb6-4635-9384-19650e9dc2a7-kube-api-access-lt6cn\") pod \"cinder-api-0\" (UID: \"800523ee-3cb6-4635-9384-19650e9dc2a7\") " pod="openstack/cinder-api-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.726487 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/800523ee-3cb6-4635-9384-19650e9dc2a7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"800523ee-3cb6-4635-9384-19650e9dc2a7\") " pod="openstack/cinder-api-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.726563 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/800523ee-3cb6-4635-9384-19650e9dc2a7-config-data\") pod \"cinder-api-0\" (UID: \"800523ee-3cb6-4635-9384-19650e9dc2a7\") " pod="openstack/cinder-api-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.726588 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/800523ee-3cb6-4635-9384-19650e9dc2a7-logs\") pod \"cinder-api-0\" (UID: \"800523ee-3cb6-4635-9384-19650e9dc2a7\") " pod="openstack/cinder-api-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.726619 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/800523ee-3cb6-4635-9384-19650e9dc2a7-scripts\") pod \"cinder-api-0\" (UID: \"800523ee-3cb6-4635-9384-19650e9dc2a7\") " pod="openstack/cinder-api-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.726635 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/800523ee-3cb6-4635-9384-19650e9dc2a7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"800523ee-3cb6-4635-9384-19650e9dc2a7\") " pod="openstack/cinder-api-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.745630 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/800523ee-3cb6-4635-9384-19650e9dc2a7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"800523ee-3cb6-4635-9384-19650e9dc2a7\") " pod="openstack/cinder-api-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.746187 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/800523ee-3cb6-4635-9384-19650e9dc2a7-logs\") pod \"cinder-api-0\" (UID: \"800523ee-3cb6-4635-9384-19650e9dc2a7\") " pod="openstack/cinder-api-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.748843 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/800523ee-3cb6-4635-9384-19650e9dc2a7-config-data\") pod \"cinder-api-0\" (UID: \"800523ee-3cb6-4635-9384-19650e9dc2a7\") " pod="openstack/cinder-api-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.759446 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b6dff65bc-t5czk" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.776083 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/800523ee-3cb6-4635-9384-19650e9dc2a7-scripts\") pod \"cinder-api-0\" (UID: \"800523ee-3cb6-4635-9384-19650e9dc2a7\") " pod="openstack/cinder-api-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.795085 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lt6cn\" (UniqueName: \"kubernetes.io/projected/800523ee-3cb6-4635-9384-19650e9dc2a7-kube-api-access-lt6cn\") pod \"cinder-api-0\" (UID: \"800523ee-3cb6-4635-9384-19650e9dc2a7\") " pod="openstack/cinder-api-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.814061 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/800523ee-3cb6-4635-9384-19650e9dc2a7-config-data-custom\") pod \"cinder-api-0\" (UID: \"800523ee-3cb6-4635-9384-19650e9dc2a7\") " pod="openstack/cinder-api-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.814720 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/800523ee-3cb6-4635-9384-19650e9dc2a7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"800523ee-3cb6-4635-9384-19650e9dc2a7\") " pod="openstack/cinder-api-0" Feb 27 17:22:33 crc kubenswrapper[4700]: I0227 17:22:33.851936 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 27 17:22:34 crc kubenswrapper[4700]: I0227 17:22:34.206319 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-787484f569-fhzcg" Feb 27 17:22:34 crc kubenswrapper[4700]: E0227 17:22:34.220246 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/ubi9/httpd-24:latest" Feb 27 17:22:34 crc kubenswrapper[4700]: E0227 17:22:34.220444 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:proxy-httpd,Image:registry.redhat.io/ubi9/httpd-24:latest,Command:[/usr/sbin/httpd],Args:[-DFOREGROUND],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:proxy-httpd,HostPort:0,ContainerPort:3000,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf/httpd.conf,SubPath:httpd.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf.d/ssl.conf,SubPath:ssl.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:run-httpd,ReadOnly:false,MountPath:/run/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:log-httpd,ReadOnly:false,MountPath:/var/log/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hm7np,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(61baccca-f7e6-4442-b612-8ab97b82036d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 27 17:22:34 crc kubenswrapper[4700]: E0227 17:22:34.221786 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"proxy-httpd\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"]" pod="openstack/ceilometer-0" podUID="61baccca-f7e6-4442-b612-8ab97b82036d" Feb 27 17:22:34 crc kubenswrapper[4700]: I0227 17:22:34.243510 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7d4f9f19-62e5-4bcd-957e-345c18380eac-dns-swift-storage-0\") pod \"7d4f9f19-62e5-4bcd-957e-345c18380eac\" (UID: \"7d4f9f19-62e5-4bcd-957e-345c18380eac\") " Feb 27 17:22:34 crc kubenswrapper[4700]: I0227 17:22:34.243563 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7d4f9f19-62e5-4bcd-957e-345c18380eac-ovsdbserver-sb\") pod \"7d4f9f19-62e5-4bcd-957e-345c18380eac\" (UID: \"7d4f9f19-62e5-4bcd-957e-345c18380eac\") " Feb 27 17:22:34 crc kubenswrapper[4700]: I0227 17:22:34.243606 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d4f9f19-62e5-4bcd-957e-345c18380eac-config\") pod \"7d4f9f19-62e5-4bcd-957e-345c18380eac\" (UID: \"7d4f9f19-62e5-4bcd-957e-345c18380eac\") " Feb 27 17:22:34 crc kubenswrapper[4700]: I0227 17:22:34.243699 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d4f9f19-62e5-4bcd-957e-345c18380eac-dns-svc\") pod \"7d4f9f19-62e5-4bcd-957e-345c18380eac\" (UID: \"7d4f9f19-62e5-4bcd-957e-345c18380eac\") " Feb 27 17:22:34 crc kubenswrapper[4700]: I0227 17:22:34.243740 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d4f9f19-62e5-4bcd-957e-345c18380eac-ovsdbserver-nb\") pod \"7d4f9f19-62e5-4bcd-957e-345c18380eac\" (UID: \"7d4f9f19-62e5-4bcd-957e-345c18380eac\") " Feb 27 17:22:34 crc kubenswrapper[4700]: I0227 17:22:34.243763 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mllb4\" (UniqueName: \"kubernetes.io/projected/7d4f9f19-62e5-4bcd-957e-345c18380eac-kube-api-access-mllb4\") pod \"7d4f9f19-62e5-4bcd-957e-345c18380eac\" (UID: \"7d4f9f19-62e5-4bcd-957e-345c18380eac\") " Feb 27 17:22:34 crc kubenswrapper[4700]: I0227 17:22:34.287878 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d4f9f19-62e5-4bcd-957e-345c18380eac-kube-api-access-mllb4" (OuterVolumeSpecName: "kube-api-access-mllb4") pod "7d4f9f19-62e5-4bcd-957e-345c18380eac" (UID: "7d4f9f19-62e5-4bcd-957e-345c18380eac"). InnerVolumeSpecName "kube-api-access-mllb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:22:34 crc kubenswrapper[4700]: I0227 17:22:34.345683 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mllb4\" (UniqueName: \"kubernetes.io/projected/7d4f9f19-62e5-4bcd-957e-345c18380eac-kube-api-access-mllb4\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:34 crc kubenswrapper[4700]: I0227 17:22:34.347105 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d4f9f19-62e5-4bcd-957e-345c18380eac-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7d4f9f19-62e5-4bcd-957e-345c18380eac" (UID: "7d4f9f19-62e5-4bcd-957e-345c18380eac"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:22:34 crc kubenswrapper[4700]: I0227 17:22:34.388636 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d4f9f19-62e5-4bcd-957e-345c18380eac-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7d4f9f19-62e5-4bcd-957e-345c18380eac" (UID: "7d4f9f19-62e5-4bcd-957e-345c18380eac"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:22:34 crc kubenswrapper[4700]: I0227 17:22:34.398904 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d4f9f19-62e5-4bcd-957e-345c18380eac-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7d4f9f19-62e5-4bcd-957e-345c18380eac" (UID: "7d4f9f19-62e5-4bcd-957e-345c18380eac"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:22:34 crc kubenswrapper[4700]: I0227 17:22:34.404832 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d4f9f19-62e5-4bcd-957e-345c18380eac-config" (OuterVolumeSpecName: "config") pod "7d4f9f19-62e5-4bcd-957e-345c18380eac" (UID: "7d4f9f19-62e5-4bcd-957e-345c18380eac"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:22:34 crc kubenswrapper[4700]: I0227 17:22:34.410869 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d4f9f19-62e5-4bcd-957e-345c18380eac-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7d4f9f19-62e5-4bcd-957e-345c18380eac" (UID: "7d4f9f19-62e5-4bcd-957e-345c18380eac"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:22:34 crc kubenswrapper[4700]: I0227 17:22:34.448699 4700 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7d4f9f19-62e5-4bcd-957e-345c18380eac-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:34 crc kubenswrapper[4700]: I0227 17:22:34.448725 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d4f9f19-62e5-4bcd-957e-345c18380eac-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:34 crc kubenswrapper[4700]: I0227 17:22:34.448736 4700 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d4f9f19-62e5-4bcd-957e-345c18380eac-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:34 crc kubenswrapper[4700]: I0227 17:22:34.448743 4700 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d4f9f19-62e5-4bcd-957e-345c18380eac-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:34 crc kubenswrapper[4700]: I0227 17:22:34.448752 4700 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7d4f9f19-62e5-4bcd-957e-345c18380eac-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:34 crc kubenswrapper[4700]: I0227 17:22:34.825964 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6cbb9d98-xkx97" Feb 27 17:22:34 crc kubenswrapper[4700]: I0227 17:22:34.873619 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-787484f569-fhzcg" Feb 27 17:22:34 crc kubenswrapper[4700]: I0227 17:22:34.874086 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-787484f569-fhzcg" event={"ID":"7d4f9f19-62e5-4bcd-957e-345c18380eac","Type":"ContainerDied","Data":"b56d1bcb9f55a4eaa8ccc45b893ff7095168b42465be0664087f40633102c8db"} Feb 27 17:22:34 crc kubenswrapper[4700]: I0227 17:22:34.874141 4700 scope.go:117] "RemoveContainer" containerID="7c8c08b9bd26fb61dd0d15098608d526d6a4041994ea5684b4d03ddc15abe370" Feb 27 17:22:34 crc kubenswrapper[4700]: I0227 17:22:34.891827 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 27 17:22:34 crc kubenswrapper[4700]: I0227 17:22:34.903689 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="61baccca-f7e6-4442-b612-8ab97b82036d" containerName="ceilometer-notification-agent" containerID="cri-o://cc73bd6dc0d5c792aec617a0df2f5575d481047b7e4f6a3c68c681a14ff1c241" gracePeriod=30 Feb 27 17:22:34 crc kubenswrapper[4700]: I0227 17:22:34.903952 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="61baccca-f7e6-4442-b612-8ab97b82036d" containerName="sg-core" containerID="cri-o://6f9ba752e86e4641d39854d1aa0c23e4bc4cc492d9d9d37d744899ec3be2bd0c" gracePeriod=30 Feb 27 17:22:34 crc kubenswrapper[4700]: I0227 17:22:34.945985 4700 scope.go:117] "RemoveContainer" containerID="93575a19a3234cea9226b813f67f83075188b0e9f1af0c913dc2b560d7450b9a" Feb 27 17:22:34 crc kubenswrapper[4700]: I0227 17:22:34.961223 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-787484f569-fhzcg"] Feb 27 17:22:34 crc kubenswrapper[4700]: I0227 17:22:34.994060 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-787484f569-fhzcg"] Feb 27 17:22:35 crc kubenswrapper[4700]: I0227 17:22:35.001138 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5b48777cdd-tqbkj" podUID="d3989a3e-45ff-45d8-80a2-4dbe57663a57" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.168:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:39386->10.217.0.168:8443: read: connection reset by peer" Feb 27 17:22:35 crc kubenswrapper[4700]: I0227 17:22:35.024688 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b6dff65bc-t5czk"] Feb 27 17:22:35 crc kubenswrapper[4700]: I0227 17:22:35.032072 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 27 17:22:35 crc kubenswrapper[4700]: I0227 17:22:35.636687 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 27 17:22:35 crc kubenswrapper[4700]: I0227 17:22:35.702423 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6cbb9d98-xkx97" Feb 27 17:22:35 crc kubenswrapper[4700]: I0227 17:22:35.776991 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-575575788d-c49lv"] Feb 27 17:22:35 crc kubenswrapper[4700]: I0227 17:22:35.777374 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-575575788d-c49lv" podUID="cb6fd0bc-8abd-4537-9c92-f8027dde0bb8" containerName="barbican-api-log" containerID="cri-o://c2395c26407b65513ecb63305725bb10b527c70bb06752aaf39a362fce871982" gracePeriod=30 Feb 27 17:22:35 crc kubenswrapper[4700]: I0227 17:22:35.778085 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-575575788d-c49lv" podUID="cb6fd0bc-8abd-4537-9c92-f8027dde0bb8" containerName="barbican-api" containerID="cri-o://802db57c2ee3737a6305579bde290751d01c2bd74d89649757372322d6dc1bec" gracePeriod=30 Feb 27 17:22:35 crc kubenswrapper[4700]: I0227 17:22:35.796755 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-575575788d-c49lv" podUID="cb6fd0bc-8abd-4537-9c92-f8027dde0bb8" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.184:9311/healthcheck\": EOF" Feb 27 17:22:35 crc kubenswrapper[4700]: I0227 17:22:35.961499 4700 generic.go:334] "Generic (PLEG): container finished" podID="d3989a3e-45ff-45d8-80a2-4dbe57663a57" containerID="cf749444d7c76d9317d18a0299503d9ccecc0cfd7ec5ffd81bd3d244886124dc" exitCode=0 Feb 27 17:22:35 crc kubenswrapper[4700]: I0227 17:22:35.961586 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5b48777cdd-tqbkj" event={"ID":"d3989a3e-45ff-45d8-80a2-4dbe57663a57","Type":"ContainerDied","Data":"cf749444d7c76d9317d18a0299503d9ccecc0cfd7ec5ffd81bd3d244886124dc"} Feb 27 17:22:35 crc kubenswrapper[4700]: I0227 17:22:35.963054 4700 generic.go:334] "Generic (PLEG): container finished" podID="18c2be27-ec7d-4d27-a61b-8c0c5d332ebe" containerID="7d103bbbca70d6c1b6a20a31114f7bfcfee1cce65fafad6dd2b8e21440fffdd9" exitCode=0 Feb 27 17:22:35 crc kubenswrapper[4700]: I0227 17:22:35.963118 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b6dff65bc-t5czk" event={"ID":"18c2be27-ec7d-4d27-a61b-8c0c5d332ebe","Type":"ContainerDied","Data":"7d103bbbca70d6c1b6a20a31114f7bfcfee1cce65fafad6dd2b8e21440fffdd9"} Feb 27 17:22:35 crc kubenswrapper[4700]: I0227 17:22:35.963142 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b6dff65bc-t5czk" event={"ID":"18c2be27-ec7d-4d27-a61b-8c0c5d332ebe","Type":"ContainerStarted","Data":"0543c218861ea5696488062d6e28b6e04cc2f07a613db947fafc831bbf8eeabb"} Feb 27 17:22:36 crc kubenswrapper[4700]: I0227 17:22:36.001943 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e365b3f6-5e7d-4911-b7ee-728ca3d413dd","Type":"ContainerStarted","Data":"9f260b7dcd7a994edc422308e2a31d1bcf762e766f8874ee204c2fa97a47bda3"} Feb 27 17:22:36 crc kubenswrapper[4700]: I0227 17:22:36.054535 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"800523ee-3cb6-4635-9384-19650e9dc2a7","Type":"ContainerStarted","Data":"fbadd113a4fa4079605535ccc1614fc6d1cb5cd4f71e4e205cd51a5e93598653"} Feb 27 17:22:36 crc kubenswrapper[4700]: I0227 17:22:36.077369 4700 generic.go:334] "Generic (PLEG): container finished" podID="61baccca-f7e6-4442-b612-8ab97b82036d" containerID="6f9ba752e86e4641d39854d1aa0c23e4bc4cc492d9d9d37d744899ec3be2bd0c" exitCode=2 Feb 27 17:22:36 crc kubenswrapper[4700]: I0227 17:22:36.077421 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"61baccca-f7e6-4442-b612-8ab97b82036d","Type":"ContainerDied","Data":"6f9ba752e86e4641d39854d1aa0c23e4bc4cc492d9d9d37d744899ec3be2bd0c"} Feb 27 17:22:37 crc kubenswrapper[4700]: I0227 17:22:37.008002 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d4f9f19-62e5-4bcd-957e-345c18380eac" path="/var/lib/kubelet/pods/7d4f9f19-62e5-4bcd-957e-345c18380eac/volumes" Feb 27 17:22:37 crc kubenswrapper[4700]: I0227 17:22:37.089763 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e365b3f6-5e7d-4911-b7ee-728ca3d413dd","Type":"ContainerStarted","Data":"803e343cfdd521fefbf86c12954e1ca55e624a13df650ab64d78ac6cd7a6c681"} Feb 27 17:22:37 crc kubenswrapper[4700]: I0227 17:22:37.093565 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"800523ee-3cb6-4635-9384-19650e9dc2a7","Type":"ContainerStarted","Data":"8d55f06a84cf1eeae52ae7a728d06f2601088c976da6ef6a11b5f994581fe92e"} Feb 27 17:22:37 crc kubenswrapper[4700]: I0227 17:22:37.097766 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b6dff65bc-t5czk" event={"ID":"18c2be27-ec7d-4d27-a61b-8c0c5d332ebe","Type":"ContainerStarted","Data":"e2a8677fd655199d727e66834fa84ee4f4ce05a81fe53e4fa222115722e3e594"} Feb 27 17:22:37 crc kubenswrapper[4700]: I0227 17:22:37.098485 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7b6dff65bc-t5czk" Feb 27 17:22:37 crc kubenswrapper[4700]: I0227 17:22:37.102411 4700 generic.go:334] "Generic (PLEG): container finished" podID="cb6fd0bc-8abd-4537-9c92-f8027dde0bb8" containerID="c2395c26407b65513ecb63305725bb10b527c70bb06752aaf39a362fce871982" exitCode=143 Feb 27 17:22:37 crc kubenswrapper[4700]: I0227 17:22:37.102454 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-575575788d-c49lv" event={"ID":"cb6fd0bc-8abd-4537-9c92-f8027dde0bb8","Type":"ContainerDied","Data":"c2395c26407b65513ecb63305725bb10b527c70bb06752aaf39a362fce871982"} Feb 27 17:22:37 crc kubenswrapper[4700]: I0227 17:22:37.116347 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7b6dff65bc-t5czk" podStartSLOduration=4.116332906 podStartE2EDuration="4.116332906s" podCreationTimestamp="2026-02-27 17:22:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:22:37.112633689 +0000 UTC m=+1317.097946436" watchObservedRunningTime="2026-02-27 17:22:37.116332906 +0000 UTC m=+1317.101645653" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.040874 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.119840 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e365b3f6-5e7d-4911-b7ee-728ca3d413dd","Type":"ContainerStarted","Data":"0a6368dff0e76b2ff761ab885586e8e7f9b63417a7e291dcdf1fce092f219d84"} Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.122657 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"800523ee-3cb6-4635-9384-19650e9dc2a7","Type":"ContainerStarted","Data":"20b68362cdfd821aa435c00deab023a702462df4a25c502bc2f07e5715e3fddf"} Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.122802 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="800523ee-3cb6-4635-9384-19650e9dc2a7" containerName="cinder-api-log" containerID="cri-o://8d55f06a84cf1eeae52ae7a728d06f2601088c976da6ef6a11b5f994581fe92e" gracePeriod=30 Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.123088 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.123132 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="800523ee-3cb6-4635-9384-19650e9dc2a7" containerName="cinder-api" containerID="cri-o://20b68362cdfd821aa435c00deab023a702462df4a25c502bc2f07e5715e3fddf" gracePeriod=30 Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.127885 4700 generic.go:334] "Generic (PLEG): container finished" podID="61baccca-f7e6-4442-b612-8ab97b82036d" containerID="cc73bd6dc0d5c792aec617a0df2f5575d481047b7e4f6a3c68c681a14ff1c241" exitCode=0 Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.130631 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.131248 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"61baccca-f7e6-4442-b612-8ab97b82036d","Type":"ContainerDied","Data":"cc73bd6dc0d5c792aec617a0df2f5575d481047b7e4f6a3c68c681a14ff1c241"} Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.131296 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"61baccca-f7e6-4442-b612-8ab97b82036d","Type":"ContainerDied","Data":"e039ff159043fe8c4a36f636e7fd3aede836587ad39537e9de0102908621e161"} Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.131315 4700 scope.go:117] "RemoveContainer" containerID="6f9ba752e86e4641d39854d1aa0c23e4bc4cc492d9d9d37d744899ec3be2bd0c" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.156484 4700 scope.go:117] "RemoveContainer" containerID="cc73bd6dc0d5c792aec617a0df2f5575d481047b7e4f6a3c68c681a14ff1c241" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.163403 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.757978741 podStartE2EDuration="5.163363529s" podCreationTimestamp="2026-02-27 17:22:33 +0000 UTC" firstStartedPulling="2026-02-27 17:22:35.04460443 +0000 UTC m=+1315.029917177" lastFinishedPulling="2026-02-27 17:22:35.449989228 +0000 UTC m=+1315.435301965" observedRunningTime="2026-02-27 17:22:38.144009388 +0000 UTC m=+1318.129322145" watchObservedRunningTime="2026-02-27 17:22:38.163363529 +0000 UTC m=+1318.148676276" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.180217 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61baccca-f7e6-4442-b612-8ab97b82036d-scripts\") pod \"61baccca-f7e6-4442-b612-8ab97b82036d\" (UID: \"61baccca-f7e6-4442-b612-8ab97b82036d\") " Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.180362 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61baccca-f7e6-4442-b612-8ab97b82036d-config-data\") pod \"61baccca-f7e6-4442-b612-8ab97b82036d\" (UID: \"61baccca-f7e6-4442-b612-8ab97b82036d\") " Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.180394 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/61baccca-f7e6-4442-b612-8ab97b82036d-log-httpd\") pod \"61baccca-f7e6-4442-b612-8ab97b82036d\" (UID: \"61baccca-f7e6-4442-b612-8ab97b82036d\") " Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.180429 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61baccca-f7e6-4442-b612-8ab97b82036d-combined-ca-bundle\") pod \"61baccca-f7e6-4442-b612-8ab97b82036d\" (UID: \"61baccca-f7e6-4442-b612-8ab97b82036d\") " Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.180479 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/61baccca-f7e6-4442-b612-8ab97b82036d-sg-core-conf-yaml\") pod \"61baccca-f7e6-4442-b612-8ab97b82036d\" (UID: \"61baccca-f7e6-4442-b612-8ab97b82036d\") " Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.180511 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hm7np\" (UniqueName: \"kubernetes.io/projected/61baccca-f7e6-4442-b612-8ab97b82036d-kube-api-access-hm7np\") pod \"61baccca-f7e6-4442-b612-8ab97b82036d\" (UID: \"61baccca-f7e6-4442-b612-8ab97b82036d\") " Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.180638 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/61baccca-f7e6-4442-b612-8ab97b82036d-run-httpd\") pod \"61baccca-f7e6-4442-b612-8ab97b82036d\" (UID: \"61baccca-f7e6-4442-b612-8ab97b82036d\") " Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.181074 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.181055036 podStartE2EDuration="5.181055036s" podCreationTimestamp="2026-02-27 17:22:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:22:38.169766828 +0000 UTC m=+1318.155079585" watchObservedRunningTime="2026-02-27 17:22:38.181055036 +0000 UTC m=+1318.166367793" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.181117 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61baccca-f7e6-4442-b612-8ab97b82036d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "61baccca-f7e6-4442-b612-8ab97b82036d" (UID: "61baccca-f7e6-4442-b612-8ab97b82036d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.181236 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61baccca-f7e6-4442-b612-8ab97b82036d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "61baccca-f7e6-4442-b612-8ab97b82036d" (UID: "61baccca-f7e6-4442-b612-8ab97b82036d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.187020 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61baccca-f7e6-4442-b612-8ab97b82036d-scripts" (OuterVolumeSpecName: "scripts") pod "61baccca-f7e6-4442-b612-8ab97b82036d" (UID: "61baccca-f7e6-4442-b612-8ab97b82036d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.190577 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61baccca-f7e6-4442-b612-8ab97b82036d-kube-api-access-hm7np" (OuterVolumeSpecName: "kube-api-access-hm7np") pod "61baccca-f7e6-4442-b612-8ab97b82036d" (UID: "61baccca-f7e6-4442-b612-8ab97b82036d"). InnerVolumeSpecName "kube-api-access-hm7np". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.192546 4700 scope.go:117] "RemoveContainer" containerID="6f9ba752e86e4641d39854d1aa0c23e4bc4cc492d9d9d37d744899ec3be2bd0c" Feb 27 17:22:38 crc kubenswrapper[4700]: E0227 17:22:38.204997 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f9ba752e86e4641d39854d1aa0c23e4bc4cc492d9d9d37d744899ec3be2bd0c\": container with ID starting with 6f9ba752e86e4641d39854d1aa0c23e4bc4cc492d9d9d37d744899ec3be2bd0c not found: ID does not exist" containerID="6f9ba752e86e4641d39854d1aa0c23e4bc4cc492d9d9d37d744899ec3be2bd0c" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.205160 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f9ba752e86e4641d39854d1aa0c23e4bc4cc492d9d9d37d744899ec3be2bd0c"} err="failed to get container status \"6f9ba752e86e4641d39854d1aa0c23e4bc4cc492d9d9d37d744899ec3be2bd0c\": rpc error: code = NotFound desc = could not find container \"6f9ba752e86e4641d39854d1aa0c23e4bc4cc492d9d9d37d744899ec3be2bd0c\": container with ID starting with 6f9ba752e86e4641d39854d1aa0c23e4bc4cc492d9d9d37d744899ec3be2bd0c not found: ID does not exist" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.205235 4700 scope.go:117] "RemoveContainer" containerID="cc73bd6dc0d5c792aec617a0df2f5575d481047b7e4f6a3c68c681a14ff1c241" Feb 27 17:22:38 crc kubenswrapper[4700]: E0227 17:22:38.205906 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc73bd6dc0d5c792aec617a0df2f5575d481047b7e4f6a3c68c681a14ff1c241\": container with ID starting with cc73bd6dc0d5c792aec617a0df2f5575d481047b7e4f6a3c68c681a14ff1c241 not found: ID does not exist" containerID="cc73bd6dc0d5c792aec617a0df2f5575d481047b7e4f6a3c68c681a14ff1c241" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.205999 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc73bd6dc0d5c792aec617a0df2f5575d481047b7e4f6a3c68c681a14ff1c241"} err="failed to get container status \"cc73bd6dc0d5c792aec617a0df2f5575d481047b7e4f6a3c68c681a14ff1c241\": rpc error: code = NotFound desc = could not find container \"cc73bd6dc0d5c792aec617a0df2f5575d481047b7e4f6a3c68c681a14ff1c241\": container with ID starting with cc73bd6dc0d5c792aec617a0df2f5575d481047b7e4f6a3c68c681a14ff1c241 not found: ID does not exist" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.210700 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61baccca-f7e6-4442-b612-8ab97b82036d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "61baccca-f7e6-4442-b612-8ab97b82036d" (UID: "61baccca-f7e6-4442-b612-8ab97b82036d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.212264 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61baccca-f7e6-4442-b612-8ab97b82036d-config-data" (OuterVolumeSpecName: "config-data") pod "61baccca-f7e6-4442-b612-8ab97b82036d" (UID: "61baccca-f7e6-4442-b612-8ab97b82036d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.212751 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61baccca-f7e6-4442-b612-8ab97b82036d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "61baccca-f7e6-4442-b612-8ab97b82036d" (UID: "61baccca-f7e6-4442-b612-8ab97b82036d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.269574 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5b48777cdd-tqbkj" podUID="d3989a3e-45ff-45d8-80a2-4dbe57663a57" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.168:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.168:8443: connect: connection refused" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.283025 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61baccca-f7e6-4442-b612-8ab97b82036d-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.283055 4700 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/61baccca-f7e6-4442-b612-8ab97b82036d-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.283064 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61baccca-f7e6-4442-b612-8ab97b82036d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.283073 4700 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/61baccca-f7e6-4442-b612-8ab97b82036d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.283083 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hm7np\" (UniqueName: \"kubernetes.io/projected/61baccca-f7e6-4442-b612-8ab97b82036d-kube-api-access-hm7np\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.283091 4700 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/61baccca-f7e6-4442-b612-8ab97b82036d-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.283099 4700 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61baccca-f7e6-4442-b612-8ab97b82036d-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.534375 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.539446 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.560185 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:22:38 crc kubenswrapper[4700]: E0227 17:22:38.560644 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61baccca-f7e6-4442-b612-8ab97b82036d" containerName="ceilometer-notification-agent" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.560655 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="61baccca-f7e6-4442-b612-8ab97b82036d" containerName="ceilometer-notification-agent" Feb 27 17:22:38 crc kubenswrapper[4700]: E0227 17:22:38.560672 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d4f9f19-62e5-4bcd-957e-345c18380eac" containerName="init" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.560678 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d4f9f19-62e5-4bcd-957e-345c18380eac" containerName="init" Feb 27 17:22:38 crc kubenswrapper[4700]: E0227 17:22:38.560701 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d4f9f19-62e5-4bcd-957e-345c18380eac" containerName="dnsmasq-dns" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.560708 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d4f9f19-62e5-4bcd-957e-345c18380eac" containerName="dnsmasq-dns" Feb 27 17:22:38 crc kubenswrapper[4700]: E0227 17:22:38.560724 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61baccca-f7e6-4442-b612-8ab97b82036d" containerName="sg-core" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.560731 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="61baccca-f7e6-4442-b612-8ab97b82036d" containerName="sg-core" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.560907 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d4f9f19-62e5-4bcd-957e-345c18380eac" containerName="dnsmasq-dns" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.560919 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="61baccca-f7e6-4442-b612-8ab97b82036d" containerName="sg-core" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.560940 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="61baccca-f7e6-4442-b612-8ab97b82036d" containerName="ceilometer-notification-agent" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.562612 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.578016 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.578710 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.578877 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.606935 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2230bd83-cd33-438b-9d9d-001e4727fb66-config-data\") pod \"ceilometer-0\" (UID: \"2230bd83-cd33-438b-9d9d-001e4727fb66\") " pod="openstack/ceilometer-0" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.607012 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2230bd83-cd33-438b-9d9d-001e4727fb66-log-httpd\") pod \"ceilometer-0\" (UID: \"2230bd83-cd33-438b-9d9d-001e4727fb66\") " pod="openstack/ceilometer-0" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.607052 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2230bd83-cd33-438b-9d9d-001e4727fb66-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2230bd83-cd33-438b-9d9d-001e4727fb66\") " pod="openstack/ceilometer-0" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.607110 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2230bd83-cd33-438b-9d9d-001e4727fb66-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2230bd83-cd33-438b-9d9d-001e4727fb66\") " pod="openstack/ceilometer-0" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.607128 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2230bd83-cd33-438b-9d9d-001e4727fb66-scripts\") pod \"ceilometer-0\" (UID: \"2230bd83-cd33-438b-9d9d-001e4727fb66\") " pod="openstack/ceilometer-0" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.607148 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87fx7\" (UniqueName: \"kubernetes.io/projected/2230bd83-cd33-438b-9d9d-001e4727fb66-kube-api-access-87fx7\") pod \"ceilometer-0\" (UID: \"2230bd83-cd33-438b-9d9d-001e4727fb66\") " pod="openstack/ceilometer-0" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.607178 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2230bd83-cd33-438b-9d9d-001e4727fb66-run-httpd\") pod \"ceilometer-0\" (UID: \"2230bd83-cd33-438b-9d9d-001e4727fb66\") " pod="openstack/ceilometer-0" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.695665 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.709158 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2230bd83-cd33-438b-9d9d-001e4727fb66-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2230bd83-cd33-438b-9d9d-001e4727fb66\") " pod="openstack/ceilometer-0" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.709216 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2230bd83-cd33-438b-9d9d-001e4727fb66-scripts\") pod \"ceilometer-0\" (UID: \"2230bd83-cd33-438b-9d9d-001e4727fb66\") " pod="openstack/ceilometer-0" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.709245 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87fx7\" (UniqueName: \"kubernetes.io/projected/2230bd83-cd33-438b-9d9d-001e4727fb66-kube-api-access-87fx7\") pod \"ceilometer-0\" (UID: \"2230bd83-cd33-438b-9d9d-001e4727fb66\") " pod="openstack/ceilometer-0" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.709278 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2230bd83-cd33-438b-9d9d-001e4727fb66-run-httpd\") pod \"ceilometer-0\" (UID: \"2230bd83-cd33-438b-9d9d-001e4727fb66\") " pod="openstack/ceilometer-0" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.709311 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2230bd83-cd33-438b-9d9d-001e4727fb66-config-data\") pod \"ceilometer-0\" (UID: \"2230bd83-cd33-438b-9d9d-001e4727fb66\") " pod="openstack/ceilometer-0" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.709365 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2230bd83-cd33-438b-9d9d-001e4727fb66-log-httpd\") pod \"ceilometer-0\" (UID: \"2230bd83-cd33-438b-9d9d-001e4727fb66\") " pod="openstack/ceilometer-0" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.709401 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2230bd83-cd33-438b-9d9d-001e4727fb66-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2230bd83-cd33-438b-9d9d-001e4727fb66\") " pod="openstack/ceilometer-0" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.710413 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2230bd83-cd33-438b-9d9d-001e4727fb66-log-httpd\") pod \"ceilometer-0\" (UID: \"2230bd83-cd33-438b-9d9d-001e4727fb66\") " pod="openstack/ceilometer-0" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.711971 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2230bd83-cd33-438b-9d9d-001e4727fb66-run-httpd\") pod \"ceilometer-0\" (UID: \"2230bd83-cd33-438b-9d9d-001e4727fb66\") " pod="openstack/ceilometer-0" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.718489 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2230bd83-cd33-438b-9d9d-001e4727fb66-scripts\") pod \"ceilometer-0\" (UID: \"2230bd83-cd33-438b-9d9d-001e4727fb66\") " pod="openstack/ceilometer-0" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.729239 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87fx7\" (UniqueName: \"kubernetes.io/projected/2230bd83-cd33-438b-9d9d-001e4727fb66-kube-api-access-87fx7\") pod \"ceilometer-0\" (UID: \"2230bd83-cd33-438b-9d9d-001e4727fb66\") " pod="openstack/ceilometer-0" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.733354 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2230bd83-cd33-438b-9d9d-001e4727fb66-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2230bd83-cd33-438b-9d9d-001e4727fb66\") " pod="openstack/ceilometer-0" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.733811 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2230bd83-cd33-438b-9d9d-001e4727fb66-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2230bd83-cd33-438b-9d9d-001e4727fb66\") " pod="openstack/ceilometer-0" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.733818 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2230bd83-cd33-438b-9d9d-001e4727fb66-config-data\") pod \"ceilometer-0\" (UID: \"2230bd83-cd33-438b-9d9d-001e4727fb66\") " pod="openstack/ceilometer-0" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.838959 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.907201 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.926773 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/800523ee-3cb6-4635-9384-19650e9dc2a7-etc-machine-id\") pod \"800523ee-3cb6-4635-9384-19650e9dc2a7\" (UID: \"800523ee-3cb6-4635-9384-19650e9dc2a7\") " Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.926846 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/800523ee-3cb6-4635-9384-19650e9dc2a7-combined-ca-bundle\") pod \"800523ee-3cb6-4635-9384-19650e9dc2a7\" (UID: \"800523ee-3cb6-4635-9384-19650e9dc2a7\") " Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.926881 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/800523ee-3cb6-4635-9384-19650e9dc2a7-logs\") pod \"800523ee-3cb6-4635-9384-19650e9dc2a7\" (UID: \"800523ee-3cb6-4635-9384-19650e9dc2a7\") " Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.926907 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/800523ee-3cb6-4635-9384-19650e9dc2a7-config-data\") pod \"800523ee-3cb6-4635-9384-19650e9dc2a7\" (UID: \"800523ee-3cb6-4635-9384-19650e9dc2a7\") " Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.926933 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/800523ee-3cb6-4635-9384-19650e9dc2a7-config-data-custom\") pod \"800523ee-3cb6-4635-9384-19650e9dc2a7\" (UID: \"800523ee-3cb6-4635-9384-19650e9dc2a7\") " Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.926958 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/800523ee-3cb6-4635-9384-19650e9dc2a7-scripts\") pod \"800523ee-3cb6-4635-9384-19650e9dc2a7\" (UID: \"800523ee-3cb6-4635-9384-19650e9dc2a7\") " Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.926978 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lt6cn\" (UniqueName: \"kubernetes.io/projected/800523ee-3cb6-4635-9384-19650e9dc2a7-kube-api-access-lt6cn\") pod \"800523ee-3cb6-4635-9384-19650e9dc2a7\" (UID: \"800523ee-3cb6-4635-9384-19650e9dc2a7\") " Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.928138 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/800523ee-3cb6-4635-9384-19650e9dc2a7-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "800523ee-3cb6-4635-9384-19650e9dc2a7" (UID: "800523ee-3cb6-4635-9384-19650e9dc2a7"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.930536 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/800523ee-3cb6-4635-9384-19650e9dc2a7-logs" (OuterVolumeSpecName: "logs") pod "800523ee-3cb6-4635-9384-19650e9dc2a7" (UID: "800523ee-3cb6-4635-9384-19650e9dc2a7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.931873 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/800523ee-3cb6-4635-9384-19650e9dc2a7-kube-api-access-lt6cn" (OuterVolumeSpecName: "kube-api-access-lt6cn") pod "800523ee-3cb6-4635-9384-19650e9dc2a7" (UID: "800523ee-3cb6-4635-9384-19650e9dc2a7"). InnerVolumeSpecName "kube-api-access-lt6cn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.940743 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/800523ee-3cb6-4635-9384-19650e9dc2a7-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "800523ee-3cb6-4635-9384-19650e9dc2a7" (UID: "800523ee-3cb6-4635-9384-19650e9dc2a7"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.945577 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/800523ee-3cb6-4635-9384-19650e9dc2a7-scripts" (OuterVolumeSpecName: "scripts") pod "800523ee-3cb6-4635-9384-19650e9dc2a7" (UID: "800523ee-3cb6-4635-9384-19650e9dc2a7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.970740 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/800523ee-3cb6-4635-9384-19650e9dc2a7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "800523ee-3cb6-4635-9384-19650e9dc2a7" (UID: "800523ee-3cb6-4635-9384-19650e9dc2a7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:38 crc kubenswrapper[4700]: I0227 17:22:38.995967 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61baccca-f7e6-4442-b612-8ab97b82036d" path="/var/lib/kubelet/pods/61baccca-f7e6-4442-b612-8ab97b82036d/volumes" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.020535 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/800523ee-3cb6-4635-9384-19650e9dc2a7-config-data" (OuterVolumeSpecName: "config-data") pod "800523ee-3cb6-4635-9384-19650e9dc2a7" (UID: "800523ee-3cb6-4635-9384-19650e9dc2a7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.029651 4700 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/800523ee-3cb6-4635-9384-19650e9dc2a7-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.029674 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/800523ee-3cb6-4635-9384-19650e9dc2a7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.029682 4700 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/800523ee-3cb6-4635-9384-19650e9dc2a7-logs\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.029692 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/800523ee-3cb6-4635-9384-19650e9dc2a7-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.029701 4700 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/800523ee-3cb6-4635-9384-19650e9dc2a7-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.029708 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lt6cn\" (UniqueName: \"kubernetes.io/projected/800523ee-3cb6-4635-9384-19650e9dc2a7-kube-api-access-lt6cn\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.029717 4700 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/800523ee-3cb6-4635-9384-19650e9dc2a7-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.098507 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-575575788d-c49lv" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.130369 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb6fd0bc-8abd-4537-9c92-f8027dde0bb8-config-data\") pod \"cb6fd0bc-8abd-4537-9c92-f8027dde0bb8\" (UID: \"cb6fd0bc-8abd-4537-9c92-f8027dde0bb8\") " Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.130414 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb6fd0bc-8abd-4537-9c92-f8027dde0bb8-combined-ca-bundle\") pod \"cb6fd0bc-8abd-4537-9c92-f8027dde0bb8\" (UID: \"cb6fd0bc-8abd-4537-9c92-f8027dde0bb8\") " Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.130533 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb6fd0bc-8abd-4537-9c92-f8027dde0bb8-config-data-custom\") pod \"cb6fd0bc-8abd-4537-9c92-f8027dde0bb8\" (UID: \"cb6fd0bc-8abd-4537-9c92-f8027dde0bb8\") " Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.130643 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb6fd0bc-8abd-4537-9c92-f8027dde0bb8-logs\") pod \"cb6fd0bc-8abd-4537-9c92-f8027dde0bb8\" (UID: \"cb6fd0bc-8abd-4537-9c92-f8027dde0bb8\") " Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.130673 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tjbt9\" (UniqueName: \"kubernetes.io/projected/cb6fd0bc-8abd-4537-9c92-f8027dde0bb8-kube-api-access-tjbt9\") pod \"cb6fd0bc-8abd-4537-9c92-f8027dde0bb8\" (UID: \"cb6fd0bc-8abd-4537-9c92-f8027dde0bb8\") " Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.141103 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb6fd0bc-8abd-4537-9c92-f8027dde0bb8-logs" (OuterVolumeSpecName: "logs") pod "cb6fd0bc-8abd-4537-9c92-f8027dde0bb8" (UID: "cb6fd0bc-8abd-4537-9c92-f8027dde0bb8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.142545 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb6fd0bc-8abd-4537-9c92-f8027dde0bb8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "cb6fd0bc-8abd-4537-9c92-f8027dde0bb8" (UID: "cb6fd0bc-8abd-4537-9c92-f8027dde0bb8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.143246 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb6fd0bc-8abd-4537-9c92-f8027dde0bb8-kube-api-access-tjbt9" (OuterVolumeSpecName: "kube-api-access-tjbt9") pod "cb6fd0bc-8abd-4537-9c92-f8027dde0bb8" (UID: "cb6fd0bc-8abd-4537-9c92-f8027dde0bb8"). InnerVolumeSpecName "kube-api-access-tjbt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.169682 4700 generic.go:334] "Generic (PLEG): container finished" podID="cb6fd0bc-8abd-4537-9c92-f8027dde0bb8" containerID="802db57c2ee3737a6305579bde290751d01c2bd74d89649757372322d6dc1bec" exitCode=0 Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.169835 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-575575788d-c49lv" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.170160 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-575575788d-c49lv" event={"ID":"cb6fd0bc-8abd-4537-9c92-f8027dde0bb8","Type":"ContainerDied","Data":"802db57c2ee3737a6305579bde290751d01c2bd74d89649757372322d6dc1bec"} Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.170211 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-575575788d-c49lv" event={"ID":"cb6fd0bc-8abd-4537-9c92-f8027dde0bb8","Type":"ContainerDied","Data":"8aa71c56cc08c5bc921c28340baee8e30653796709e293f549308a19f1c690bd"} Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.170310 4700 scope.go:117] "RemoveContainer" containerID="802db57c2ee3737a6305579bde290751d01c2bd74d89649757372322d6dc1bec" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.175856 4700 generic.go:334] "Generic (PLEG): container finished" podID="800523ee-3cb6-4635-9384-19650e9dc2a7" containerID="20b68362cdfd821aa435c00deab023a702462df4a25c502bc2f07e5715e3fddf" exitCode=0 Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.175882 4700 generic.go:334] "Generic (PLEG): container finished" podID="800523ee-3cb6-4635-9384-19650e9dc2a7" containerID="8d55f06a84cf1eeae52ae7a728d06f2601088c976da6ef6a11b5f994581fe92e" exitCode=143 Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.175898 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.175951 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"800523ee-3cb6-4635-9384-19650e9dc2a7","Type":"ContainerDied","Data":"20b68362cdfd821aa435c00deab023a702462df4a25c502bc2f07e5715e3fddf"} Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.175976 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"800523ee-3cb6-4635-9384-19650e9dc2a7","Type":"ContainerDied","Data":"8d55f06a84cf1eeae52ae7a728d06f2601088c976da6ef6a11b5f994581fe92e"} Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.175987 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"800523ee-3cb6-4635-9384-19650e9dc2a7","Type":"ContainerDied","Data":"fbadd113a4fa4079605535ccc1614fc6d1cb5cd4f71e4e205cd51a5e93598653"} Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.186039 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb6fd0bc-8abd-4537-9c92-f8027dde0bb8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cb6fd0bc-8abd-4537-9c92-f8027dde0bb8" (UID: "cb6fd0bc-8abd-4537-9c92-f8027dde0bb8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.214354 4700 scope.go:117] "RemoveContainer" containerID="c2395c26407b65513ecb63305725bb10b527c70bb06752aaf39a362fce871982" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.237773 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.244721 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb6fd0bc-8abd-4537-9c92-f8027dde0bb8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.244750 4700 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb6fd0bc-8abd-4537-9c92-f8027dde0bb8-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.244759 4700 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb6fd0bc-8abd-4537-9c92-f8027dde0bb8-logs\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.244767 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tjbt9\" (UniqueName: \"kubernetes.io/projected/cb6fd0bc-8abd-4537-9c92-f8027dde0bb8-kube-api-access-tjbt9\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.255263 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.258054 4700 scope.go:117] "RemoveContainer" containerID="802db57c2ee3737a6305579bde290751d01c2bd74d89649757372322d6dc1bec" Feb 27 17:22:39 crc kubenswrapper[4700]: E0227 17:22:39.258947 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"802db57c2ee3737a6305579bde290751d01c2bd74d89649757372322d6dc1bec\": container with ID starting with 802db57c2ee3737a6305579bde290751d01c2bd74d89649757372322d6dc1bec not found: ID does not exist" containerID="802db57c2ee3737a6305579bde290751d01c2bd74d89649757372322d6dc1bec" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.258989 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"802db57c2ee3737a6305579bde290751d01c2bd74d89649757372322d6dc1bec"} err="failed to get container status \"802db57c2ee3737a6305579bde290751d01c2bd74d89649757372322d6dc1bec\": rpc error: code = NotFound desc = could not find container \"802db57c2ee3737a6305579bde290751d01c2bd74d89649757372322d6dc1bec\": container with ID starting with 802db57c2ee3737a6305579bde290751d01c2bd74d89649757372322d6dc1bec not found: ID does not exist" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.259018 4700 scope.go:117] "RemoveContainer" containerID="c2395c26407b65513ecb63305725bb10b527c70bb06752aaf39a362fce871982" Feb 27 17:22:39 crc kubenswrapper[4700]: E0227 17:22:39.259489 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2395c26407b65513ecb63305725bb10b527c70bb06752aaf39a362fce871982\": container with ID starting with c2395c26407b65513ecb63305725bb10b527c70bb06752aaf39a362fce871982 not found: ID does not exist" containerID="c2395c26407b65513ecb63305725bb10b527c70bb06752aaf39a362fce871982" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.259519 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2395c26407b65513ecb63305725bb10b527c70bb06752aaf39a362fce871982"} err="failed to get container status \"c2395c26407b65513ecb63305725bb10b527c70bb06752aaf39a362fce871982\": rpc error: code = NotFound desc = could not find container \"c2395c26407b65513ecb63305725bb10b527c70bb06752aaf39a362fce871982\": container with ID starting with c2395c26407b65513ecb63305725bb10b527c70bb06752aaf39a362fce871982 not found: ID does not exist" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.259539 4700 scope.go:117] "RemoveContainer" containerID="20b68362cdfd821aa435c00deab023a702462df4a25c502bc2f07e5715e3fddf" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.268913 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 27 17:22:39 crc kubenswrapper[4700]: E0227 17:22:39.269343 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="800523ee-3cb6-4635-9384-19650e9dc2a7" containerName="cinder-api" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.269356 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="800523ee-3cb6-4635-9384-19650e9dc2a7" containerName="cinder-api" Feb 27 17:22:39 crc kubenswrapper[4700]: E0227 17:22:39.269373 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb6fd0bc-8abd-4537-9c92-f8027dde0bb8" containerName="barbican-api-log" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.269379 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb6fd0bc-8abd-4537-9c92-f8027dde0bb8" containerName="barbican-api-log" Feb 27 17:22:39 crc kubenswrapper[4700]: E0227 17:22:39.269392 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb6fd0bc-8abd-4537-9c92-f8027dde0bb8" containerName="barbican-api" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.269398 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb6fd0bc-8abd-4537-9c92-f8027dde0bb8" containerName="barbican-api" Feb 27 17:22:39 crc kubenswrapper[4700]: E0227 17:22:39.269422 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="800523ee-3cb6-4635-9384-19650e9dc2a7" containerName="cinder-api-log" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.269429 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="800523ee-3cb6-4635-9384-19650e9dc2a7" containerName="cinder-api-log" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.269604 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb6fd0bc-8abd-4537-9c92-f8027dde0bb8" containerName="barbican-api-log" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.269626 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="800523ee-3cb6-4635-9384-19650e9dc2a7" containerName="cinder-api-log" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.269634 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="800523ee-3cb6-4635-9384-19650e9dc2a7" containerName="cinder-api" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.269649 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb6fd0bc-8abd-4537-9c92-f8027dde0bb8" containerName="barbican-api" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.270699 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.273334 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.273504 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.273532 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.280401 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb6fd0bc-8abd-4537-9c92-f8027dde0bb8-config-data" (OuterVolumeSpecName: "config-data") pod "cb6fd0bc-8abd-4537-9c92-f8027dde0bb8" (UID: "cb6fd0bc-8abd-4537-9c92-f8027dde0bb8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.299372 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.303319 4700 scope.go:117] "RemoveContainer" containerID="8d55f06a84cf1eeae52ae7a728d06f2601088c976da6ef6a11b5f994581fe92e" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.325286 4700 scope.go:117] "RemoveContainer" containerID="20b68362cdfd821aa435c00deab023a702462df4a25c502bc2f07e5715e3fddf" Feb 27 17:22:39 crc kubenswrapper[4700]: E0227 17:22:39.326529 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20b68362cdfd821aa435c00deab023a702462df4a25c502bc2f07e5715e3fddf\": container with ID starting with 20b68362cdfd821aa435c00deab023a702462df4a25c502bc2f07e5715e3fddf not found: ID does not exist" containerID="20b68362cdfd821aa435c00deab023a702462df4a25c502bc2f07e5715e3fddf" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.326577 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20b68362cdfd821aa435c00deab023a702462df4a25c502bc2f07e5715e3fddf"} err="failed to get container status \"20b68362cdfd821aa435c00deab023a702462df4a25c502bc2f07e5715e3fddf\": rpc error: code = NotFound desc = could not find container \"20b68362cdfd821aa435c00deab023a702462df4a25c502bc2f07e5715e3fddf\": container with ID starting with 20b68362cdfd821aa435c00deab023a702462df4a25c502bc2f07e5715e3fddf not found: ID does not exist" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.326643 4700 scope.go:117] "RemoveContainer" containerID="8d55f06a84cf1eeae52ae7a728d06f2601088c976da6ef6a11b5f994581fe92e" Feb 27 17:22:39 crc kubenswrapper[4700]: E0227 17:22:39.327043 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d55f06a84cf1eeae52ae7a728d06f2601088c976da6ef6a11b5f994581fe92e\": container with ID starting with 8d55f06a84cf1eeae52ae7a728d06f2601088c976da6ef6a11b5f994581fe92e not found: ID does not exist" containerID="8d55f06a84cf1eeae52ae7a728d06f2601088c976da6ef6a11b5f994581fe92e" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.327083 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d55f06a84cf1eeae52ae7a728d06f2601088c976da6ef6a11b5f994581fe92e"} err="failed to get container status \"8d55f06a84cf1eeae52ae7a728d06f2601088c976da6ef6a11b5f994581fe92e\": rpc error: code = NotFound desc = could not find container \"8d55f06a84cf1eeae52ae7a728d06f2601088c976da6ef6a11b5f994581fe92e\": container with ID starting with 8d55f06a84cf1eeae52ae7a728d06f2601088c976da6ef6a11b5f994581fe92e not found: ID does not exist" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.327095 4700 scope.go:117] "RemoveContainer" containerID="20b68362cdfd821aa435c00deab023a702462df4a25c502bc2f07e5715e3fddf" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.327306 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20b68362cdfd821aa435c00deab023a702462df4a25c502bc2f07e5715e3fddf"} err="failed to get container status \"20b68362cdfd821aa435c00deab023a702462df4a25c502bc2f07e5715e3fddf\": rpc error: code = NotFound desc = could not find container \"20b68362cdfd821aa435c00deab023a702462df4a25c502bc2f07e5715e3fddf\": container with ID starting with 20b68362cdfd821aa435c00deab023a702462df4a25c502bc2f07e5715e3fddf not found: ID does not exist" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.327323 4700 scope.go:117] "RemoveContainer" containerID="8d55f06a84cf1eeae52ae7a728d06f2601088c976da6ef6a11b5f994581fe92e" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.327654 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d55f06a84cf1eeae52ae7a728d06f2601088c976da6ef6a11b5f994581fe92e"} err="failed to get container status \"8d55f06a84cf1eeae52ae7a728d06f2601088c976da6ef6a11b5f994581fe92e\": rpc error: code = NotFound desc = could not find container \"8d55f06a84cf1eeae52ae7a728d06f2601088c976da6ef6a11b5f994581fe92e\": container with ID starting with 8d55f06a84cf1eeae52ae7a728d06f2601088c976da6ef6a11b5f994581fe92e not found: ID does not exist" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.345879 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d9080fd5-c068-49f1-8167-e3843d34d471-config-data-custom\") pod \"cinder-api-0\" (UID: \"d9080fd5-c068-49f1-8167-e3843d34d471\") " pod="openstack/cinder-api-0" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.345917 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9080fd5-c068-49f1-8167-e3843d34d471-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d9080fd5-c068-49f1-8167-e3843d34d471\") " pod="openstack/cinder-api-0" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.345947 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9080fd5-c068-49f1-8167-e3843d34d471-public-tls-certs\") pod \"cinder-api-0\" (UID: \"d9080fd5-c068-49f1-8167-e3843d34d471\") " pod="openstack/cinder-api-0" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.346027 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9080fd5-c068-49f1-8167-e3843d34d471-scripts\") pod \"cinder-api-0\" (UID: \"d9080fd5-c068-49f1-8167-e3843d34d471\") " pod="openstack/cinder-api-0" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.346068 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9080fd5-c068-49f1-8167-e3843d34d471-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"d9080fd5-c068-49f1-8167-e3843d34d471\") " pod="openstack/cinder-api-0" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.346168 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d9080fd5-c068-49f1-8167-e3843d34d471-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d9080fd5-c068-49f1-8167-e3843d34d471\") " pod="openstack/cinder-api-0" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.346297 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmt8z\" (UniqueName: \"kubernetes.io/projected/d9080fd5-c068-49f1-8167-e3843d34d471-kube-api-access-jmt8z\") pod \"cinder-api-0\" (UID: \"d9080fd5-c068-49f1-8167-e3843d34d471\") " pod="openstack/cinder-api-0" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.346379 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9080fd5-c068-49f1-8167-e3843d34d471-config-data\") pod \"cinder-api-0\" (UID: \"d9080fd5-c068-49f1-8167-e3843d34d471\") " pod="openstack/cinder-api-0" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.346422 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9080fd5-c068-49f1-8167-e3843d34d471-logs\") pod \"cinder-api-0\" (UID: \"d9080fd5-c068-49f1-8167-e3843d34d471\") " pod="openstack/cinder-api-0" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.346530 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb6fd0bc-8abd-4537-9c92-f8027dde0bb8-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.408782 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.448708 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d9080fd5-c068-49f1-8167-e3843d34d471-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d9080fd5-c068-49f1-8167-e3843d34d471\") " pod="openstack/cinder-api-0" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.448792 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmt8z\" (UniqueName: \"kubernetes.io/projected/d9080fd5-c068-49f1-8167-e3843d34d471-kube-api-access-jmt8z\") pod \"cinder-api-0\" (UID: \"d9080fd5-c068-49f1-8167-e3843d34d471\") " pod="openstack/cinder-api-0" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.448838 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9080fd5-c068-49f1-8167-e3843d34d471-config-data\") pod \"cinder-api-0\" (UID: \"d9080fd5-c068-49f1-8167-e3843d34d471\") " pod="openstack/cinder-api-0" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.448865 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9080fd5-c068-49f1-8167-e3843d34d471-logs\") pod \"cinder-api-0\" (UID: \"d9080fd5-c068-49f1-8167-e3843d34d471\") " pod="openstack/cinder-api-0" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.448901 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9080fd5-c068-49f1-8167-e3843d34d471-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d9080fd5-c068-49f1-8167-e3843d34d471\") " pod="openstack/cinder-api-0" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.448917 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d9080fd5-c068-49f1-8167-e3843d34d471-config-data-custom\") pod \"cinder-api-0\" (UID: \"d9080fd5-c068-49f1-8167-e3843d34d471\") " pod="openstack/cinder-api-0" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.448993 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9080fd5-c068-49f1-8167-e3843d34d471-public-tls-certs\") pod \"cinder-api-0\" (UID: \"d9080fd5-c068-49f1-8167-e3843d34d471\") " pod="openstack/cinder-api-0" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.449016 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9080fd5-c068-49f1-8167-e3843d34d471-scripts\") pod \"cinder-api-0\" (UID: \"d9080fd5-c068-49f1-8167-e3843d34d471\") " pod="openstack/cinder-api-0" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.449036 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9080fd5-c068-49f1-8167-e3843d34d471-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"d9080fd5-c068-49f1-8167-e3843d34d471\") " pod="openstack/cinder-api-0" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.449709 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9080fd5-c068-49f1-8167-e3843d34d471-logs\") pod \"cinder-api-0\" (UID: \"d9080fd5-c068-49f1-8167-e3843d34d471\") " pod="openstack/cinder-api-0" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.449743 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d9080fd5-c068-49f1-8167-e3843d34d471-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d9080fd5-c068-49f1-8167-e3843d34d471\") " pod="openstack/cinder-api-0" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.453959 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d9080fd5-c068-49f1-8167-e3843d34d471-config-data-custom\") pod \"cinder-api-0\" (UID: \"d9080fd5-c068-49f1-8167-e3843d34d471\") " pod="openstack/cinder-api-0" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.454370 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9080fd5-c068-49f1-8167-e3843d34d471-public-tls-certs\") pod \"cinder-api-0\" (UID: \"d9080fd5-c068-49f1-8167-e3843d34d471\") " pod="openstack/cinder-api-0" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.455156 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9080fd5-c068-49f1-8167-e3843d34d471-config-data\") pod \"cinder-api-0\" (UID: \"d9080fd5-c068-49f1-8167-e3843d34d471\") " pod="openstack/cinder-api-0" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.455342 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9080fd5-c068-49f1-8167-e3843d34d471-scripts\") pod \"cinder-api-0\" (UID: \"d9080fd5-c068-49f1-8167-e3843d34d471\") " pod="openstack/cinder-api-0" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.456155 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9080fd5-c068-49f1-8167-e3843d34d471-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"d9080fd5-c068-49f1-8167-e3843d34d471\") " pod="openstack/cinder-api-0" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.456265 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9080fd5-c068-49f1-8167-e3843d34d471-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d9080fd5-c068-49f1-8167-e3843d34d471\") " pod="openstack/cinder-api-0" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.469823 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmt8z\" (UniqueName: \"kubernetes.io/projected/d9080fd5-c068-49f1-8167-e3843d34d471-kube-api-access-jmt8z\") pod \"cinder-api-0\" (UID: \"d9080fd5-c068-49f1-8167-e3843d34d471\") " pod="openstack/cinder-api-0" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.505845 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-575575788d-c49lv"] Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.516451 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-575575788d-c49lv"] Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.601333 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 27 17:22:39 crc kubenswrapper[4700]: I0227 17:22:39.981722 4700 scope.go:117] "RemoveContainer" containerID="6c8960b7fce9f293dfe5ac76c658c4a8af1e7bbe2abfc614cc1caccc7e49b99a" Feb 27 17:22:40 crc kubenswrapper[4700]: I0227 17:22:40.087606 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7db57cbc8-t9p8l" Feb 27 17:22:40 crc kubenswrapper[4700]: I0227 17:22:40.154968 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 27 17:22:40 crc kubenswrapper[4700]: W0227 17:22:40.156433 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9080fd5_c068_49f1_8167_e3843d34d471.slice/crio-a5f77e2b354b37e8a017712377067df4ce658d03075cf6e7a98806b5e728b287 WatchSource:0}: Error finding container a5f77e2b354b37e8a017712377067df4ce658d03075cf6e7a98806b5e728b287: Status 404 returned error can't find the container with id a5f77e2b354b37e8a017712377067df4ce658d03075cf6e7a98806b5e728b287 Feb 27 17:22:40 crc kubenswrapper[4700]: I0227 17:22:40.261285 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2230bd83-cd33-438b-9d9d-001e4727fb66","Type":"ContainerStarted","Data":"939cdbfe2666a6be220f5f5345c486c74663161256917db6f4f375d293727904"} Feb 27 17:22:40 crc kubenswrapper[4700]: I0227 17:22:40.261328 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2230bd83-cd33-438b-9d9d-001e4727fb66","Type":"ContainerStarted","Data":"69c2b00d086046db5dbe4e390eadf8c54a7f4f8e01f44f386589e10c1718169b"} Feb 27 17:22:40 crc kubenswrapper[4700]: I0227 17:22:40.272058 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d9080fd5-c068-49f1-8167-e3843d34d471","Type":"ContainerStarted","Data":"a5f77e2b354b37e8a017712377067df4ce658d03075cf6e7a98806b5e728b287"} Feb 27 17:22:40 crc kubenswrapper[4700]: I0227 17:22:40.432450 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-8946ccf97-n4k7p"] Feb 27 17:22:40 crc kubenswrapper[4700]: I0227 17:22:40.432899 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-8946ccf97-n4k7p" podUID="f0d00b91-0e33-4466-8c9f-6a1c1038d23f" containerName="neutron-api" containerID="cri-o://276f51e328821d543c3ac41196366ae8614a6c6d49c1d994e3ad93e0d3638898" gracePeriod=30 Feb 27 17:22:40 crc kubenswrapper[4700]: I0227 17:22:40.433356 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-8946ccf97-n4k7p" podUID="f0d00b91-0e33-4466-8c9f-6a1c1038d23f" containerName="neutron-httpd" containerID="cri-o://c320d708dcbfe0f89eb170beac5b5ab60bc12dacba3ba4ee98246f53114f457d" gracePeriod=30 Feb 27 17:22:40 crc kubenswrapper[4700]: I0227 17:22:40.453020 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-dc4cd4577-rvnzh"] Feb 27 17:22:40 crc kubenswrapper[4700]: I0227 17:22:40.454579 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dc4cd4577-rvnzh" Feb 27 17:22:40 crc kubenswrapper[4700]: I0227 17:22:40.477951 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dc4cd4577-rvnzh"] Feb 27 17:22:40 crc kubenswrapper[4700]: I0227 17:22:40.509848 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-8946ccf97-n4k7p" Feb 27 17:22:40 crc kubenswrapper[4700]: I0227 17:22:40.579485 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d4d08db-27c4-4d2c-8902-968271280975-combined-ca-bundle\") pod \"neutron-dc4cd4577-rvnzh\" (UID: \"8d4d08db-27c4-4d2c-8902-968271280975\") " pod="openstack/neutron-dc4cd4577-rvnzh" Feb 27 17:22:40 crc kubenswrapper[4700]: I0227 17:22:40.579533 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d4d08db-27c4-4d2c-8902-968271280975-public-tls-certs\") pod \"neutron-dc4cd4577-rvnzh\" (UID: \"8d4d08db-27c4-4d2c-8902-968271280975\") " pod="openstack/neutron-dc4cd4577-rvnzh" Feb 27 17:22:40 crc kubenswrapper[4700]: I0227 17:22:40.579589 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8d4d08db-27c4-4d2c-8902-968271280975-httpd-config\") pod \"neutron-dc4cd4577-rvnzh\" (UID: \"8d4d08db-27c4-4d2c-8902-968271280975\") " pod="openstack/neutron-dc4cd4577-rvnzh" Feb 27 17:22:40 crc kubenswrapper[4700]: I0227 17:22:40.579711 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hcv2\" (UniqueName: \"kubernetes.io/projected/8d4d08db-27c4-4d2c-8902-968271280975-kube-api-access-2hcv2\") pod \"neutron-dc4cd4577-rvnzh\" (UID: \"8d4d08db-27c4-4d2c-8902-968271280975\") " pod="openstack/neutron-dc4cd4577-rvnzh" Feb 27 17:22:40 crc kubenswrapper[4700]: I0227 17:22:40.579755 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d4d08db-27c4-4d2c-8902-968271280975-ovndb-tls-certs\") pod \"neutron-dc4cd4577-rvnzh\" (UID: \"8d4d08db-27c4-4d2c-8902-968271280975\") " pod="openstack/neutron-dc4cd4577-rvnzh" Feb 27 17:22:40 crc kubenswrapper[4700]: I0227 17:22:40.579830 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8d4d08db-27c4-4d2c-8902-968271280975-config\") pod \"neutron-dc4cd4577-rvnzh\" (UID: \"8d4d08db-27c4-4d2c-8902-968271280975\") " pod="openstack/neutron-dc4cd4577-rvnzh" Feb 27 17:22:40 crc kubenswrapper[4700]: I0227 17:22:40.579885 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d4d08db-27c4-4d2c-8902-968271280975-internal-tls-certs\") pod \"neutron-dc4cd4577-rvnzh\" (UID: \"8d4d08db-27c4-4d2c-8902-968271280975\") " pod="openstack/neutron-dc4cd4577-rvnzh" Feb 27 17:22:40 crc kubenswrapper[4700]: I0227 17:22:40.681798 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d4d08db-27c4-4d2c-8902-968271280975-internal-tls-certs\") pod \"neutron-dc4cd4577-rvnzh\" (UID: \"8d4d08db-27c4-4d2c-8902-968271280975\") " pod="openstack/neutron-dc4cd4577-rvnzh" Feb 27 17:22:40 crc kubenswrapper[4700]: I0227 17:22:40.681861 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d4d08db-27c4-4d2c-8902-968271280975-combined-ca-bundle\") pod \"neutron-dc4cd4577-rvnzh\" (UID: \"8d4d08db-27c4-4d2c-8902-968271280975\") " pod="openstack/neutron-dc4cd4577-rvnzh" Feb 27 17:22:40 crc kubenswrapper[4700]: I0227 17:22:40.681888 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d4d08db-27c4-4d2c-8902-968271280975-public-tls-certs\") pod \"neutron-dc4cd4577-rvnzh\" (UID: \"8d4d08db-27c4-4d2c-8902-968271280975\") " pod="openstack/neutron-dc4cd4577-rvnzh" Feb 27 17:22:40 crc kubenswrapper[4700]: I0227 17:22:40.681915 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8d4d08db-27c4-4d2c-8902-968271280975-httpd-config\") pod \"neutron-dc4cd4577-rvnzh\" (UID: \"8d4d08db-27c4-4d2c-8902-968271280975\") " pod="openstack/neutron-dc4cd4577-rvnzh" Feb 27 17:22:40 crc kubenswrapper[4700]: I0227 17:22:40.681981 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hcv2\" (UniqueName: \"kubernetes.io/projected/8d4d08db-27c4-4d2c-8902-968271280975-kube-api-access-2hcv2\") pod \"neutron-dc4cd4577-rvnzh\" (UID: \"8d4d08db-27c4-4d2c-8902-968271280975\") " pod="openstack/neutron-dc4cd4577-rvnzh" Feb 27 17:22:40 crc kubenswrapper[4700]: I0227 17:22:40.682002 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d4d08db-27c4-4d2c-8902-968271280975-ovndb-tls-certs\") pod \"neutron-dc4cd4577-rvnzh\" (UID: \"8d4d08db-27c4-4d2c-8902-968271280975\") " pod="openstack/neutron-dc4cd4577-rvnzh" Feb 27 17:22:40 crc kubenswrapper[4700]: I0227 17:22:40.682041 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8d4d08db-27c4-4d2c-8902-968271280975-config\") pod \"neutron-dc4cd4577-rvnzh\" (UID: \"8d4d08db-27c4-4d2c-8902-968271280975\") " pod="openstack/neutron-dc4cd4577-rvnzh" Feb 27 17:22:40 crc kubenswrapper[4700]: I0227 17:22:40.687173 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8d4d08db-27c4-4d2c-8902-968271280975-httpd-config\") pod \"neutron-dc4cd4577-rvnzh\" (UID: \"8d4d08db-27c4-4d2c-8902-968271280975\") " pod="openstack/neutron-dc4cd4577-rvnzh" Feb 27 17:22:40 crc kubenswrapper[4700]: I0227 17:22:40.689560 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d4d08db-27c4-4d2c-8902-968271280975-ovndb-tls-certs\") pod \"neutron-dc4cd4577-rvnzh\" (UID: \"8d4d08db-27c4-4d2c-8902-968271280975\") " pod="openstack/neutron-dc4cd4577-rvnzh" Feb 27 17:22:40 crc kubenswrapper[4700]: I0227 17:22:40.690356 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d4d08db-27c4-4d2c-8902-968271280975-combined-ca-bundle\") pod \"neutron-dc4cd4577-rvnzh\" (UID: \"8d4d08db-27c4-4d2c-8902-968271280975\") " pod="openstack/neutron-dc4cd4577-rvnzh" Feb 27 17:22:40 crc kubenswrapper[4700]: I0227 17:22:40.692148 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8d4d08db-27c4-4d2c-8902-968271280975-config\") pod \"neutron-dc4cd4577-rvnzh\" (UID: \"8d4d08db-27c4-4d2c-8902-968271280975\") " pod="openstack/neutron-dc4cd4577-rvnzh" Feb 27 17:22:40 crc kubenswrapper[4700]: I0227 17:22:40.694611 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d4d08db-27c4-4d2c-8902-968271280975-internal-tls-certs\") pod \"neutron-dc4cd4577-rvnzh\" (UID: \"8d4d08db-27c4-4d2c-8902-968271280975\") " pod="openstack/neutron-dc4cd4577-rvnzh" Feb 27 17:22:40 crc kubenswrapper[4700]: I0227 17:22:40.702765 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d4d08db-27c4-4d2c-8902-968271280975-public-tls-certs\") pod \"neutron-dc4cd4577-rvnzh\" (UID: \"8d4d08db-27c4-4d2c-8902-968271280975\") " pod="openstack/neutron-dc4cd4577-rvnzh" Feb 27 17:22:40 crc kubenswrapper[4700]: I0227 17:22:40.704267 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hcv2\" (UniqueName: \"kubernetes.io/projected/8d4d08db-27c4-4d2c-8902-968271280975-kube-api-access-2hcv2\") pod \"neutron-dc4cd4577-rvnzh\" (UID: \"8d4d08db-27c4-4d2c-8902-968271280975\") " pod="openstack/neutron-dc4cd4577-rvnzh" Feb 27 17:22:40 crc kubenswrapper[4700]: I0227 17:22:40.801590 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dc4cd4577-rvnzh" Feb 27 17:22:41 crc kubenswrapper[4700]: I0227 17:22:41.044515 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="800523ee-3cb6-4635-9384-19650e9dc2a7" path="/var/lib/kubelet/pods/800523ee-3cb6-4635-9384-19650e9dc2a7/volumes" Feb 27 17:22:41 crc kubenswrapper[4700]: I0227 17:22:41.046862 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb6fd0bc-8abd-4537-9c92-f8027dde0bb8" path="/var/lib/kubelet/pods/cb6fd0bc-8abd-4537-9c92-f8027dde0bb8/volumes" Feb 27 17:22:41 crc kubenswrapper[4700]: I0227 17:22:41.761445 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2230bd83-cd33-438b-9d9d-001e4727fb66","Type":"ContainerStarted","Data":"beb916ec8dfdea917abac29c24892a3eb645748ade539e9616642050bb73dace"} Feb 27 17:22:41 crc kubenswrapper[4700]: I0227 17:22:41.764485 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"0f1ae433-28e4-45cf-9635-dc6edbf16f6e","Type":"ContainerStarted","Data":"9b51d27b3971cd3f655c413b4bd270b5cc55c1c7f20422db504613285c3a5770"} Feb 27 17:22:41 crc kubenswrapper[4700]: I0227 17:22:41.791859 4700 generic.go:334] "Generic (PLEG): container finished" podID="f0d00b91-0e33-4466-8c9f-6a1c1038d23f" containerID="c320d708dcbfe0f89eb170beac5b5ab60bc12dacba3ba4ee98246f53114f457d" exitCode=0 Feb 27 17:22:41 crc kubenswrapper[4700]: I0227 17:22:41.791905 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8946ccf97-n4k7p" event={"ID":"f0d00b91-0e33-4466-8c9f-6a1c1038d23f","Type":"ContainerDied","Data":"c320d708dcbfe0f89eb170beac5b5ab60bc12dacba3ba4ee98246f53114f457d"} Feb 27 17:22:41 crc kubenswrapper[4700]: I0227 17:22:41.819618 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dc4cd4577-rvnzh"] Feb 27 17:22:42 crc kubenswrapper[4700]: I0227 17:22:42.254131 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-8946ccf97-n4k7p" podUID="f0d00b91-0e33-4466-8c9f-6a1c1038d23f" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.177:9696/\": dial tcp 10.217.0.177:9696: connect: connection refused" Feb 27 17:22:42 crc kubenswrapper[4700]: I0227 17:22:42.802211 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d9080fd5-c068-49f1-8167-e3843d34d471","Type":"ContainerStarted","Data":"f685b1ba5792ae1b792fca1628573ead459675070faee9e6339d2456a12cc03c"} Feb 27 17:22:42 crc kubenswrapper[4700]: I0227 17:22:42.802634 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d9080fd5-c068-49f1-8167-e3843d34d471","Type":"ContainerStarted","Data":"418b8bf07a10a62de2fd30e0d3b8ada8d0b0bff87e5fa89d7a3b98c015bad1ff"} Feb 27 17:22:42 crc kubenswrapper[4700]: I0227 17:22:42.802726 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 27 17:22:42 crc kubenswrapper[4700]: I0227 17:22:42.804659 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2230bd83-cd33-438b-9d9d-001e4727fb66","Type":"ContainerStarted","Data":"3ad4aca4ec61405efe35f12a970058daaf58cd3480a4fc6f5a2a641fae53761e"} Feb 27 17:22:42 crc kubenswrapper[4700]: I0227 17:22:42.806216 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dc4cd4577-rvnzh" event={"ID":"8d4d08db-27c4-4d2c-8902-968271280975","Type":"ContainerStarted","Data":"b01b8ca02725c9ef1add79ab5f0cd46d9b561b8e1d983d1c99fc78f313bbf241"} Feb 27 17:22:42 crc kubenswrapper[4700]: I0227 17:22:42.806238 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dc4cd4577-rvnzh" event={"ID":"8d4d08db-27c4-4d2c-8902-968271280975","Type":"ContainerStarted","Data":"59d9eec0cd17e133fb17e7189f16c7f8bf88832c5fb1a1e264c3afc28a3af0ee"} Feb 27 17:22:42 crc kubenswrapper[4700]: I0227 17:22:42.806248 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dc4cd4577-rvnzh" event={"ID":"8d4d08db-27c4-4d2c-8902-968271280975","Type":"ContainerStarted","Data":"1c383bb1da00c67e923fd8a3f8f23f73eec7eb05cc0c8022a8ba82f7ca4cb993"} Feb 27 17:22:42 crc kubenswrapper[4700]: I0227 17:22:42.807074 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-dc4cd4577-rvnzh" Feb 27 17:22:42 crc kubenswrapper[4700]: I0227 17:22:42.825225 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.8252086629999997 podStartE2EDuration="3.825208663s" podCreationTimestamp="2026-02-27 17:22:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:22:42.822759438 +0000 UTC m=+1322.808072185" watchObservedRunningTime="2026-02-27 17:22:42.825208663 +0000 UTC m=+1322.810521400" Feb 27 17:22:43 crc kubenswrapper[4700]: I0227 17:22:43.762577 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7b6dff65bc-t5czk" Feb 27 17:22:43 crc kubenswrapper[4700]: I0227 17:22:43.791239 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-dc4cd4577-rvnzh" podStartSLOduration=3.791221778 podStartE2EDuration="3.791221778s" podCreationTimestamp="2026-02-27 17:22:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:22:42.855970944 +0000 UTC m=+1322.841283691" watchObservedRunningTime="2026-02-27 17:22:43.791221778 +0000 UTC m=+1323.776534525" Feb 27 17:22:43 crc kubenswrapper[4700]: I0227 17:22:43.833451 4700 generic.go:334] "Generic (PLEG): container finished" podID="0f1ae433-28e4-45cf-9635-dc6edbf16f6e" containerID="9b51d27b3971cd3f655c413b4bd270b5cc55c1c7f20422db504613285c3a5770" exitCode=1 Feb 27 17:22:43 crc kubenswrapper[4700]: I0227 17:22:43.833538 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"0f1ae433-28e4-45cf-9635-dc6edbf16f6e","Type":"ContainerDied","Data":"9b51d27b3971cd3f655c413b4bd270b5cc55c1c7f20422db504613285c3a5770"} Feb 27 17:22:43 crc kubenswrapper[4700]: I0227 17:22:43.833571 4700 scope.go:117] "RemoveContainer" containerID="6c8960b7fce9f293dfe5ac76c658c4a8af1e7bbe2abfc614cc1caccc7e49b99a" Feb 27 17:22:43 crc kubenswrapper[4700]: I0227 17:22:43.839178 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2230bd83-cd33-438b-9d9d-001e4727fb66","Type":"ContainerStarted","Data":"8e03101f296c609397f67f6db25ba89c6efef855a814df305c7288fddff2ce5b"} Feb 27 17:22:43 crc kubenswrapper[4700]: I0227 17:22:43.841420 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-575575788d-c49lv" podUID="cb6fd0bc-8abd-4537-9c92-f8027dde0bb8" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.184:9311/healthcheck\": dial tcp 10.217.0.184:9311: i/o timeout (Client.Timeout exceeded while awaiting headers)" Feb 27 17:22:43 crc kubenswrapper[4700]: I0227 17:22:43.841523 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-575575788d-c49lv" podUID="cb6fd0bc-8abd-4537-9c92-f8027dde0bb8" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.184:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 27 17:22:43 crc kubenswrapper[4700]: I0227 17:22:43.842273 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77475fb77-jzwgk"] Feb 27 17:22:43 crc kubenswrapper[4700]: I0227 17:22:43.844331 4700 scope.go:117] "RemoveContainer" containerID="9b51d27b3971cd3f655c413b4bd270b5cc55c1c7f20422db504613285c3a5770" Feb 27 17:22:43 crc kubenswrapper[4700]: E0227 17:22:43.844625 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 20s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(0f1ae433-28e4-45cf-9635-dc6edbf16f6e)\"" pod="openstack/watcher-decision-engine-0" podUID="0f1ae433-28e4-45cf-9635-dc6edbf16f6e" Feb 27 17:22:43 crc kubenswrapper[4700]: I0227 17:22:43.845566 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77475fb77-jzwgk" podUID="c48cd283-8c2c-4825-ad5e-c014c1fbc369" containerName="dnsmasq-dns" containerID="cri-o://466f82136ad1681a5d562884d53c1245dfc13fb1c109b416395fe31d7e466329" gracePeriod=10 Feb 27 17:22:43 crc kubenswrapper[4700]: I0227 17:22:43.902775 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.847038604 podStartE2EDuration="5.902751741s" podCreationTimestamp="2026-02-27 17:22:38 +0000 UTC" firstStartedPulling="2026-02-27 17:22:39.462327651 +0000 UTC m=+1319.447640398" lastFinishedPulling="2026-02-27 17:22:43.518040788 +0000 UTC m=+1323.503353535" observedRunningTime="2026-02-27 17:22:43.890537019 +0000 UTC m=+1323.875849766" watchObservedRunningTime="2026-02-27 17:22:43.902751741 +0000 UTC m=+1323.888064488" Feb 27 17:22:43 crc kubenswrapper[4700]: I0227 17:22:43.918593 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 27 17:22:43 crc kubenswrapper[4700]: I0227 17:22:43.991580 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 27 17:22:44 crc kubenswrapper[4700]: I0227 17:22:44.389524 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77475fb77-jzwgk" Feb 27 17:22:44 crc kubenswrapper[4700]: I0227 17:22:44.511692 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c48cd283-8c2c-4825-ad5e-c014c1fbc369-config\") pod \"c48cd283-8c2c-4825-ad5e-c014c1fbc369\" (UID: \"c48cd283-8c2c-4825-ad5e-c014c1fbc369\") " Feb 27 17:22:44 crc kubenswrapper[4700]: I0227 17:22:44.511964 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c48cd283-8c2c-4825-ad5e-c014c1fbc369-ovsdbserver-nb\") pod \"c48cd283-8c2c-4825-ad5e-c014c1fbc369\" (UID: \"c48cd283-8c2c-4825-ad5e-c014c1fbc369\") " Feb 27 17:22:44 crc kubenswrapper[4700]: I0227 17:22:44.512064 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxgh5\" (UniqueName: \"kubernetes.io/projected/c48cd283-8c2c-4825-ad5e-c014c1fbc369-kube-api-access-pxgh5\") pod \"c48cd283-8c2c-4825-ad5e-c014c1fbc369\" (UID: \"c48cd283-8c2c-4825-ad5e-c014c1fbc369\") " Feb 27 17:22:44 crc kubenswrapper[4700]: I0227 17:22:44.512212 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c48cd283-8c2c-4825-ad5e-c014c1fbc369-ovsdbserver-sb\") pod \"c48cd283-8c2c-4825-ad5e-c014c1fbc369\" (UID: \"c48cd283-8c2c-4825-ad5e-c014c1fbc369\") " Feb 27 17:22:44 crc kubenswrapper[4700]: I0227 17:22:44.512282 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c48cd283-8c2c-4825-ad5e-c014c1fbc369-dns-swift-storage-0\") pod \"c48cd283-8c2c-4825-ad5e-c014c1fbc369\" (UID: \"c48cd283-8c2c-4825-ad5e-c014c1fbc369\") " Feb 27 17:22:44 crc kubenswrapper[4700]: I0227 17:22:44.512391 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c48cd283-8c2c-4825-ad5e-c014c1fbc369-dns-svc\") pod \"c48cd283-8c2c-4825-ad5e-c014c1fbc369\" (UID: \"c48cd283-8c2c-4825-ad5e-c014c1fbc369\") " Feb 27 17:22:44 crc kubenswrapper[4700]: I0227 17:22:44.517813 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c48cd283-8c2c-4825-ad5e-c014c1fbc369-kube-api-access-pxgh5" (OuterVolumeSpecName: "kube-api-access-pxgh5") pod "c48cd283-8c2c-4825-ad5e-c014c1fbc369" (UID: "c48cd283-8c2c-4825-ad5e-c014c1fbc369"). InnerVolumeSpecName "kube-api-access-pxgh5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:22:44 crc kubenswrapper[4700]: I0227 17:22:44.584196 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c48cd283-8c2c-4825-ad5e-c014c1fbc369-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c48cd283-8c2c-4825-ad5e-c014c1fbc369" (UID: "c48cd283-8c2c-4825-ad5e-c014c1fbc369"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:22:44 crc kubenswrapper[4700]: I0227 17:22:44.585858 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c48cd283-8c2c-4825-ad5e-c014c1fbc369-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c48cd283-8c2c-4825-ad5e-c014c1fbc369" (UID: "c48cd283-8c2c-4825-ad5e-c014c1fbc369"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:22:44 crc kubenswrapper[4700]: I0227 17:22:44.586342 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c48cd283-8c2c-4825-ad5e-c014c1fbc369-config" (OuterVolumeSpecName: "config") pod "c48cd283-8c2c-4825-ad5e-c014c1fbc369" (UID: "c48cd283-8c2c-4825-ad5e-c014c1fbc369"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:22:44 crc kubenswrapper[4700]: I0227 17:22:44.588387 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c48cd283-8c2c-4825-ad5e-c014c1fbc369-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c48cd283-8c2c-4825-ad5e-c014c1fbc369" (UID: "c48cd283-8c2c-4825-ad5e-c014c1fbc369"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:22:44 crc kubenswrapper[4700]: I0227 17:22:44.598296 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c48cd283-8c2c-4825-ad5e-c014c1fbc369-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c48cd283-8c2c-4825-ad5e-c014c1fbc369" (UID: "c48cd283-8c2c-4825-ad5e-c014c1fbc369"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:22:44 crc kubenswrapper[4700]: I0227 17:22:44.614391 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c48cd283-8c2c-4825-ad5e-c014c1fbc369-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:44 crc kubenswrapper[4700]: I0227 17:22:44.614425 4700 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c48cd283-8c2c-4825-ad5e-c014c1fbc369-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:44 crc kubenswrapper[4700]: I0227 17:22:44.614437 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxgh5\" (UniqueName: \"kubernetes.io/projected/c48cd283-8c2c-4825-ad5e-c014c1fbc369-kube-api-access-pxgh5\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:44 crc kubenswrapper[4700]: I0227 17:22:44.614445 4700 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c48cd283-8c2c-4825-ad5e-c014c1fbc369-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:44 crc kubenswrapper[4700]: I0227 17:22:44.614454 4700 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c48cd283-8c2c-4825-ad5e-c014c1fbc369-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:44 crc kubenswrapper[4700]: I0227 17:22:44.614474 4700 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c48cd283-8c2c-4825-ad5e-c014c1fbc369-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:44 crc kubenswrapper[4700]: I0227 17:22:44.850163 4700 generic.go:334] "Generic (PLEG): container finished" podID="c48cd283-8c2c-4825-ad5e-c014c1fbc369" containerID="466f82136ad1681a5d562884d53c1245dfc13fb1c109b416395fe31d7e466329" exitCode=0 Feb 27 17:22:44 crc kubenswrapper[4700]: I0227 17:22:44.850272 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77475fb77-jzwgk" Feb 27 17:22:44 crc kubenswrapper[4700]: I0227 17:22:44.852587 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77475fb77-jzwgk" event={"ID":"c48cd283-8c2c-4825-ad5e-c014c1fbc369","Type":"ContainerDied","Data":"466f82136ad1681a5d562884d53c1245dfc13fb1c109b416395fe31d7e466329"} Feb 27 17:22:44 crc kubenswrapper[4700]: I0227 17:22:44.852825 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77475fb77-jzwgk" event={"ID":"c48cd283-8c2c-4825-ad5e-c014c1fbc369","Type":"ContainerDied","Data":"524cd92e9cc50904b11d98e56cf6e4baa203c67867254cb99fd61ca1cf05d712"} Feb 27 17:22:44 crc kubenswrapper[4700]: I0227 17:22:44.852913 4700 scope.go:117] "RemoveContainer" containerID="466f82136ad1681a5d562884d53c1245dfc13fb1c109b416395fe31d7e466329" Feb 27 17:22:44 crc kubenswrapper[4700]: I0227 17:22:44.856004 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="e365b3f6-5e7d-4911-b7ee-728ca3d413dd" containerName="cinder-scheduler" containerID="cri-o://803e343cfdd521fefbf86c12954e1ca55e624a13df650ab64d78ac6cd7a6c681" gracePeriod=30 Feb 27 17:22:44 crc kubenswrapper[4700]: I0227 17:22:44.856091 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="e365b3f6-5e7d-4911-b7ee-728ca3d413dd" containerName="probe" containerID="cri-o://0a6368dff0e76b2ff761ab885586e8e7f9b63417a7e291dcdf1fce092f219d84" gracePeriod=30 Feb 27 17:22:44 crc kubenswrapper[4700]: I0227 17:22:44.856715 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 27 17:22:44 crc kubenswrapper[4700]: I0227 17:22:44.907671 4700 scope.go:117] "RemoveContainer" containerID="3f3a57688d1e2b9d41e08a9b9af3820237ddf454c09113a4ae620f716dc197b6" Feb 27 17:22:44 crc kubenswrapper[4700]: I0227 17:22:44.917112 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77475fb77-jzwgk"] Feb 27 17:22:44 crc kubenswrapper[4700]: I0227 17:22:44.926360 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77475fb77-jzwgk"] Feb 27 17:22:44 crc kubenswrapper[4700]: I0227 17:22:44.938668 4700 scope.go:117] "RemoveContainer" containerID="466f82136ad1681a5d562884d53c1245dfc13fb1c109b416395fe31d7e466329" Feb 27 17:22:44 crc kubenswrapper[4700]: E0227 17:22:44.939141 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"466f82136ad1681a5d562884d53c1245dfc13fb1c109b416395fe31d7e466329\": container with ID starting with 466f82136ad1681a5d562884d53c1245dfc13fb1c109b416395fe31d7e466329 not found: ID does not exist" containerID="466f82136ad1681a5d562884d53c1245dfc13fb1c109b416395fe31d7e466329" Feb 27 17:22:44 crc kubenswrapper[4700]: I0227 17:22:44.939173 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"466f82136ad1681a5d562884d53c1245dfc13fb1c109b416395fe31d7e466329"} err="failed to get container status \"466f82136ad1681a5d562884d53c1245dfc13fb1c109b416395fe31d7e466329\": rpc error: code = NotFound desc = could not find container \"466f82136ad1681a5d562884d53c1245dfc13fb1c109b416395fe31d7e466329\": container with ID starting with 466f82136ad1681a5d562884d53c1245dfc13fb1c109b416395fe31d7e466329 not found: ID does not exist" Feb 27 17:22:44 crc kubenswrapper[4700]: I0227 17:22:44.939196 4700 scope.go:117] "RemoveContainer" containerID="3f3a57688d1e2b9d41e08a9b9af3820237ddf454c09113a4ae620f716dc197b6" Feb 27 17:22:44 crc kubenswrapper[4700]: E0227 17:22:44.939509 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f3a57688d1e2b9d41e08a9b9af3820237ddf454c09113a4ae620f716dc197b6\": container with ID starting with 3f3a57688d1e2b9d41e08a9b9af3820237ddf454c09113a4ae620f716dc197b6 not found: ID does not exist" containerID="3f3a57688d1e2b9d41e08a9b9af3820237ddf454c09113a4ae620f716dc197b6" Feb 27 17:22:44 crc kubenswrapper[4700]: I0227 17:22:44.939544 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f3a57688d1e2b9d41e08a9b9af3820237ddf454c09113a4ae620f716dc197b6"} err="failed to get container status \"3f3a57688d1e2b9d41e08a9b9af3820237ddf454c09113a4ae620f716dc197b6\": rpc error: code = NotFound desc = could not find container \"3f3a57688d1e2b9d41e08a9b9af3820237ddf454c09113a4ae620f716dc197b6\": container with ID starting with 3f3a57688d1e2b9d41e08a9b9af3820237ddf454c09113a4ae620f716dc197b6 not found: ID does not exist" Feb 27 17:22:44 crc kubenswrapper[4700]: I0227 17:22:44.994117 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c48cd283-8c2c-4825-ad5e-c014c1fbc369" path="/var/lib/kubelet/pods/c48cd283-8c2c-4825-ad5e-c014c1fbc369/volumes" Feb 27 17:22:45 crc kubenswrapper[4700]: I0227 17:22:45.901403 4700 generic.go:334] "Generic (PLEG): container finished" podID="f0d00b91-0e33-4466-8c9f-6a1c1038d23f" containerID="276f51e328821d543c3ac41196366ae8614a6c6d49c1d994e3ad93e0d3638898" exitCode=0 Feb 27 17:22:45 crc kubenswrapper[4700]: I0227 17:22:45.901548 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8946ccf97-n4k7p" event={"ID":"f0d00b91-0e33-4466-8c9f-6a1c1038d23f","Type":"ContainerDied","Data":"276f51e328821d543c3ac41196366ae8614a6c6d49c1d994e3ad93e0d3638898"} Feb 27 17:22:45 crc kubenswrapper[4700]: I0227 17:22:45.926650 4700 generic.go:334] "Generic (PLEG): container finished" podID="e365b3f6-5e7d-4911-b7ee-728ca3d413dd" containerID="0a6368dff0e76b2ff761ab885586e8e7f9b63417a7e291dcdf1fce092f219d84" exitCode=0 Feb 27 17:22:45 crc kubenswrapper[4700]: I0227 17:22:45.926737 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e365b3f6-5e7d-4911-b7ee-728ca3d413dd","Type":"ContainerDied","Data":"0a6368dff0e76b2ff761ab885586e8e7f9b63417a7e291dcdf1fce092f219d84"} Feb 27 17:22:46 crc kubenswrapper[4700]: I0227 17:22:46.383910 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8946ccf97-n4k7p" Feb 27 17:22:46 crc kubenswrapper[4700]: I0227 17:22:46.555002 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-internal-tls-certs\") pod \"f0d00b91-0e33-4466-8c9f-6a1c1038d23f\" (UID: \"f0d00b91-0e33-4466-8c9f-6a1c1038d23f\") " Feb 27 17:22:46 crc kubenswrapper[4700]: I0227 17:22:46.555043 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2szd\" (UniqueName: \"kubernetes.io/projected/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-kube-api-access-j2szd\") pod \"f0d00b91-0e33-4466-8c9f-6a1c1038d23f\" (UID: \"f0d00b91-0e33-4466-8c9f-6a1c1038d23f\") " Feb 27 17:22:46 crc kubenswrapper[4700]: I0227 17:22:46.555158 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-config\") pod \"f0d00b91-0e33-4466-8c9f-6a1c1038d23f\" (UID: \"f0d00b91-0e33-4466-8c9f-6a1c1038d23f\") " Feb 27 17:22:46 crc kubenswrapper[4700]: I0227 17:22:46.555234 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-public-tls-certs\") pod \"f0d00b91-0e33-4466-8c9f-6a1c1038d23f\" (UID: \"f0d00b91-0e33-4466-8c9f-6a1c1038d23f\") " Feb 27 17:22:46 crc kubenswrapper[4700]: I0227 17:22:46.555270 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-httpd-config\") pod \"f0d00b91-0e33-4466-8c9f-6a1c1038d23f\" (UID: \"f0d00b91-0e33-4466-8c9f-6a1c1038d23f\") " Feb 27 17:22:46 crc kubenswrapper[4700]: I0227 17:22:46.555312 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-combined-ca-bundle\") pod \"f0d00b91-0e33-4466-8c9f-6a1c1038d23f\" (UID: \"f0d00b91-0e33-4466-8c9f-6a1c1038d23f\") " Feb 27 17:22:46 crc kubenswrapper[4700]: I0227 17:22:46.555340 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-ovndb-tls-certs\") pod \"f0d00b91-0e33-4466-8c9f-6a1c1038d23f\" (UID: \"f0d00b91-0e33-4466-8c9f-6a1c1038d23f\") " Feb 27 17:22:46 crc kubenswrapper[4700]: I0227 17:22:46.562012 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "f0d00b91-0e33-4466-8c9f-6a1c1038d23f" (UID: "f0d00b91-0e33-4466-8c9f-6a1c1038d23f"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:46 crc kubenswrapper[4700]: I0227 17:22:46.563076 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-kube-api-access-j2szd" (OuterVolumeSpecName: "kube-api-access-j2szd") pod "f0d00b91-0e33-4466-8c9f-6a1c1038d23f" (UID: "f0d00b91-0e33-4466-8c9f-6a1c1038d23f"). InnerVolumeSpecName "kube-api-access-j2szd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:22:46 crc kubenswrapper[4700]: I0227 17:22:46.614385 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f0d00b91-0e33-4466-8c9f-6a1c1038d23f" (UID: "f0d00b91-0e33-4466-8c9f-6a1c1038d23f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:46 crc kubenswrapper[4700]: I0227 17:22:46.615281 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-config" (OuterVolumeSpecName: "config") pod "f0d00b91-0e33-4466-8c9f-6a1c1038d23f" (UID: "f0d00b91-0e33-4466-8c9f-6a1c1038d23f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:46 crc kubenswrapper[4700]: I0227 17:22:46.634708 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f0d00b91-0e33-4466-8c9f-6a1c1038d23f" (UID: "f0d00b91-0e33-4466-8c9f-6a1c1038d23f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:46 crc kubenswrapper[4700]: I0227 17:22:46.634769 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f0d00b91-0e33-4466-8c9f-6a1c1038d23f" (UID: "f0d00b91-0e33-4466-8c9f-6a1c1038d23f"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:46 crc kubenswrapper[4700]: I0227 17:22:46.655900 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "f0d00b91-0e33-4466-8c9f-6a1c1038d23f" (UID: "f0d00b91-0e33-4466-8c9f-6a1c1038d23f"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:46 crc kubenswrapper[4700]: I0227 17:22:46.656707 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-ovndb-tls-certs\") pod \"f0d00b91-0e33-4466-8c9f-6a1c1038d23f\" (UID: \"f0d00b91-0e33-4466-8c9f-6a1c1038d23f\") " Feb 27 17:22:46 crc kubenswrapper[4700]: W0227 17:22:46.656879 4700 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/f0d00b91-0e33-4466-8c9f-6a1c1038d23f/volumes/kubernetes.io~secret/ovndb-tls-certs Feb 27 17:22:46 crc kubenswrapper[4700]: I0227 17:22:46.656911 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "f0d00b91-0e33-4466-8c9f-6a1c1038d23f" (UID: "f0d00b91-0e33-4466-8c9f-6a1c1038d23f"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:46 crc kubenswrapper[4700]: I0227 17:22:46.657125 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:46 crc kubenswrapper[4700]: I0227 17:22:46.657146 4700 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:46 crc kubenswrapper[4700]: I0227 17:22:46.657155 4700 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:46 crc kubenswrapper[4700]: I0227 17:22:46.657165 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:46 crc kubenswrapper[4700]: I0227 17:22:46.657173 4700 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:46 crc kubenswrapper[4700]: I0227 17:22:46.657181 4700 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:46 crc kubenswrapper[4700]: I0227 17:22:46.657190 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2szd\" (UniqueName: \"kubernetes.io/projected/f0d00b91-0e33-4466-8c9f-6a1c1038d23f-kube-api-access-j2szd\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:46 crc kubenswrapper[4700]: I0227 17:22:46.936516 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8946ccf97-n4k7p" event={"ID":"f0d00b91-0e33-4466-8c9f-6a1c1038d23f","Type":"ContainerDied","Data":"665272820cc2663fa668acd424d980adf81034d6df2f2bfb7a976ce5645487d8"} Feb 27 17:22:46 crc kubenswrapper[4700]: I0227 17:22:46.936568 4700 scope.go:117] "RemoveContainer" containerID="c320d708dcbfe0f89eb170beac5b5ab60bc12dacba3ba4ee98246f53114f457d" Feb 27 17:22:46 crc kubenswrapper[4700]: I0227 17:22:46.936695 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8946ccf97-n4k7p" Feb 27 17:22:46 crc kubenswrapper[4700]: I0227 17:22:46.974336 4700 scope.go:117] "RemoveContainer" containerID="276f51e328821d543c3ac41196366ae8614a6c6d49c1d994e3ad93e0d3638898" Feb 27 17:22:47 crc kubenswrapper[4700]: I0227 17:22:47.003436 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-8946ccf97-n4k7p"] Feb 27 17:22:47 crc kubenswrapper[4700]: I0227 17:22:47.004061 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-8946ccf97-n4k7p"] Feb 27 17:22:47 crc kubenswrapper[4700]: I0227 17:22:47.857085 4700 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/watcher-decision-engine-0" Feb 27 17:22:47 crc kubenswrapper[4700]: I0227 17:22:47.857138 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Feb 27 17:22:47 crc kubenswrapper[4700]: I0227 17:22:47.857153 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Feb 27 17:22:47 crc kubenswrapper[4700]: I0227 17:22:47.857165 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Feb 27 17:22:47 crc kubenswrapper[4700]: I0227 17:22:47.857922 4700 scope.go:117] "RemoveContainer" containerID="9b51d27b3971cd3f655c413b4bd270b5cc55c1c7f20422db504613285c3a5770" Feb 27 17:22:47 crc kubenswrapper[4700]: E0227 17:22:47.858254 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 20s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(0f1ae433-28e4-45cf-9635-dc6edbf16f6e)\"" pod="openstack/watcher-decision-engine-0" podUID="0f1ae433-28e4-45cf-9635-dc6edbf16f6e" Feb 27 17:22:48 crc kubenswrapper[4700]: I0227 17:22:48.269621 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5b48777cdd-tqbkj" podUID="d3989a3e-45ff-45d8-80a2-4dbe57663a57" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.168:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.168:8443: connect: connection refused" Feb 27 17:22:48 crc kubenswrapper[4700]: I0227 17:22:48.600354 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 27 17:22:48 crc kubenswrapper[4700]: I0227 17:22:48.696860 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e365b3f6-5e7d-4911-b7ee-728ca3d413dd-etc-machine-id\") pod \"e365b3f6-5e7d-4911-b7ee-728ca3d413dd\" (UID: \"e365b3f6-5e7d-4911-b7ee-728ca3d413dd\") " Feb 27 17:22:48 crc kubenswrapper[4700]: I0227 17:22:48.697207 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e365b3f6-5e7d-4911-b7ee-728ca3d413dd-config-data\") pod \"e365b3f6-5e7d-4911-b7ee-728ca3d413dd\" (UID: \"e365b3f6-5e7d-4911-b7ee-728ca3d413dd\") " Feb 27 17:22:48 crc kubenswrapper[4700]: I0227 17:22:48.697231 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e365b3f6-5e7d-4911-b7ee-728ca3d413dd-combined-ca-bundle\") pod \"e365b3f6-5e7d-4911-b7ee-728ca3d413dd\" (UID: \"e365b3f6-5e7d-4911-b7ee-728ca3d413dd\") " Feb 27 17:22:48 crc kubenswrapper[4700]: I0227 17:22:48.696967 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e365b3f6-5e7d-4911-b7ee-728ca3d413dd-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e365b3f6-5e7d-4911-b7ee-728ca3d413dd" (UID: "e365b3f6-5e7d-4911-b7ee-728ca3d413dd"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 17:22:48 crc kubenswrapper[4700]: I0227 17:22:48.697327 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e365b3f6-5e7d-4911-b7ee-728ca3d413dd-config-data-custom\") pod \"e365b3f6-5e7d-4911-b7ee-728ca3d413dd\" (UID: \"e365b3f6-5e7d-4911-b7ee-728ca3d413dd\") " Feb 27 17:22:48 crc kubenswrapper[4700]: I0227 17:22:48.697357 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5dbk\" (UniqueName: \"kubernetes.io/projected/e365b3f6-5e7d-4911-b7ee-728ca3d413dd-kube-api-access-f5dbk\") pod \"e365b3f6-5e7d-4911-b7ee-728ca3d413dd\" (UID: \"e365b3f6-5e7d-4911-b7ee-728ca3d413dd\") " Feb 27 17:22:48 crc kubenswrapper[4700]: I0227 17:22:48.697925 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e365b3f6-5e7d-4911-b7ee-728ca3d413dd-scripts\") pod \"e365b3f6-5e7d-4911-b7ee-728ca3d413dd\" (UID: \"e365b3f6-5e7d-4911-b7ee-728ca3d413dd\") " Feb 27 17:22:48 crc kubenswrapper[4700]: I0227 17:22:48.698352 4700 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e365b3f6-5e7d-4911-b7ee-728ca3d413dd-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:48 crc kubenswrapper[4700]: I0227 17:22:48.702430 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e365b3f6-5e7d-4911-b7ee-728ca3d413dd-kube-api-access-f5dbk" (OuterVolumeSpecName: "kube-api-access-f5dbk") pod "e365b3f6-5e7d-4911-b7ee-728ca3d413dd" (UID: "e365b3f6-5e7d-4911-b7ee-728ca3d413dd"). InnerVolumeSpecName "kube-api-access-f5dbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:22:48 crc kubenswrapper[4700]: I0227 17:22:48.702521 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e365b3f6-5e7d-4911-b7ee-728ca3d413dd-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e365b3f6-5e7d-4911-b7ee-728ca3d413dd" (UID: "e365b3f6-5e7d-4911-b7ee-728ca3d413dd"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:48 crc kubenswrapper[4700]: I0227 17:22:48.702615 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e365b3f6-5e7d-4911-b7ee-728ca3d413dd-scripts" (OuterVolumeSpecName: "scripts") pod "e365b3f6-5e7d-4911-b7ee-728ca3d413dd" (UID: "e365b3f6-5e7d-4911-b7ee-728ca3d413dd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:48 crc kubenswrapper[4700]: I0227 17:22:48.784646 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e365b3f6-5e7d-4911-b7ee-728ca3d413dd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e365b3f6-5e7d-4911-b7ee-728ca3d413dd" (UID: "e365b3f6-5e7d-4911-b7ee-728ca3d413dd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:48 crc kubenswrapper[4700]: I0227 17:22:48.799661 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e365b3f6-5e7d-4911-b7ee-728ca3d413dd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:48 crc kubenswrapper[4700]: I0227 17:22:48.799683 4700 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e365b3f6-5e7d-4911-b7ee-728ca3d413dd-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:48 crc kubenswrapper[4700]: I0227 17:22:48.799692 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5dbk\" (UniqueName: \"kubernetes.io/projected/e365b3f6-5e7d-4911-b7ee-728ca3d413dd-kube-api-access-f5dbk\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:48 crc kubenswrapper[4700]: I0227 17:22:48.799704 4700 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e365b3f6-5e7d-4911-b7ee-728ca3d413dd-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:48 crc kubenswrapper[4700]: I0227 17:22:48.894344 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e365b3f6-5e7d-4911-b7ee-728ca3d413dd-config-data" (OuterVolumeSpecName: "config-data") pod "e365b3f6-5e7d-4911-b7ee-728ca3d413dd" (UID: "e365b3f6-5e7d-4911-b7ee-728ca3d413dd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:22:48 crc kubenswrapper[4700]: I0227 17:22:48.901710 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e365b3f6-5e7d-4911-b7ee-728ca3d413dd-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:22:48 crc kubenswrapper[4700]: I0227 17:22:48.953902 4700 generic.go:334] "Generic (PLEG): container finished" podID="e365b3f6-5e7d-4911-b7ee-728ca3d413dd" containerID="803e343cfdd521fefbf86c12954e1ca55e624a13df650ab64d78ac6cd7a6c681" exitCode=0 Feb 27 17:22:48 crc kubenswrapper[4700]: I0227 17:22:48.953937 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e365b3f6-5e7d-4911-b7ee-728ca3d413dd","Type":"ContainerDied","Data":"803e343cfdd521fefbf86c12954e1ca55e624a13df650ab64d78ac6cd7a6c681"} Feb 27 17:22:48 crc kubenswrapper[4700]: I0227 17:22:48.953963 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e365b3f6-5e7d-4911-b7ee-728ca3d413dd","Type":"ContainerDied","Data":"9f260b7dcd7a994edc422308e2a31d1bcf762e766f8874ee204c2fa97a47bda3"} Feb 27 17:22:48 crc kubenswrapper[4700]: I0227 17:22:48.953981 4700 scope.go:117] "RemoveContainer" containerID="0a6368dff0e76b2ff761ab885586e8e7f9b63417a7e291dcdf1fce092f219d84" Feb 27 17:22:48 crc kubenswrapper[4700]: I0227 17:22:48.954070 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 27 17:22:48 crc kubenswrapper[4700]: I0227 17:22:48.995132 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0d00b91-0e33-4466-8c9f-6a1c1038d23f" path="/var/lib/kubelet/pods/f0d00b91-0e33-4466-8c9f-6a1c1038d23f/volumes" Feb 27 17:22:48 crc kubenswrapper[4700]: I0227 17:22:48.995721 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 27 17:22:48 crc kubenswrapper[4700]: I0227 17:22:48.998206 4700 scope.go:117] "RemoveContainer" containerID="803e343cfdd521fefbf86c12954e1ca55e624a13df650ab64d78ac6cd7a6c681" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.000348 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.021997 4700 scope.go:117] "RemoveContainer" containerID="0a6368dff0e76b2ff761ab885586e8e7f9b63417a7e291dcdf1fce092f219d84" Feb 27 17:22:49 crc kubenswrapper[4700]: E0227 17:22:49.022481 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a6368dff0e76b2ff761ab885586e8e7f9b63417a7e291dcdf1fce092f219d84\": container with ID starting with 0a6368dff0e76b2ff761ab885586e8e7f9b63417a7e291dcdf1fce092f219d84 not found: ID does not exist" containerID="0a6368dff0e76b2ff761ab885586e8e7f9b63417a7e291dcdf1fce092f219d84" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.022529 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a6368dff0e76b2ff761ab885586e8e7f9b63417a7e291dcdf1fce092f219d84"} err="failed to get container status \"0a6368dff0e76b2ff761ab885586e8e7f9b63417a7e291dcdf1fce092f219d84\": rpc error: code = NotFound desc = could not find container \"0a6368dff0e76b2ff761ab885586e8e7f9b63417a7e291dcdf1fce092f219d84\": container with ID starting with 0a6368dff0e76b2ff761ab885586e8e7f9b63417a7e291dcdf1fce092f219d84 not found: ID does not exist" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.022560 4700 scope.go:117] "RemoveContainer" containerID="803e343cfdd521fefbf86c12954e1ca55e624a13df650ab64d78ac6cd7a6c681" Feb 27 17:22:49 crc kubenswrapper[4700]: E0227 17:22:49.022910 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"803e343cfdd521fefbf86c12954e1ca55e624a13df650ab64d78ac6cd7a6c681\": container with ID starting with 803e343cfdd521fefbf86c12954e1ca55e624a13df650ab64d78ac6cd7a6c681 not found: ID does not exist" containerID="803e343cfdd521fefbf86c12954e1ca55e624a13df650ab64d78ac6cd7a6c681" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.022946 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"803e343cfdd521fefbf86c12954e1ca55e624a13df650ab64d78ac6cd7a6c681"} err="failed to get container status \"803e343cfdd521fefbf86c12954e1ca55e624a13df650ab64d78ac6cd7a6c681\": rpc error: code = NotFound desc = could not find container \"803e343cfdd521fefbf86c12954e1ca55e624a13df650ab64d78ac6cd7a6c681\": container with ID starting with 803e343cfdd521fefbf86c12954e1ca55e624a13df650ab64d78ac6cd7a6c681 not found: ID does not exist" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.023224 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 27 17:22:49 crc kubenswrapper[4700]: E0227 17:22:49.023605 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e365b3f6-5e7d-4911-b7ee-728ca3d413dd" containerName="probe" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.023622 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="e365b3f6-5e7d-4911-b7ee-728ca3d413dd" containerName="probe" Feb 27 17:22:49 crc kubenswrapper[4700]: E0227 17:22:49.023637 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e365b3f6-5e7d-4911-b7ee-728ca3d413dd" containerName="cinder-scheduler" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.023643 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="e365b3f6-5e7d-4911-b7ee-728ca3d413dd" containerName="cinder-scheduler" Feb 27 17:22:49 crc kubenswrapper[4700]: E0227 17:22:49.023654 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0d00b91-0e33-4466-8c9f-6a1c1038d23f" containerName="neutron-api" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.023662 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0d00b91-0e33-4466-8c9f-6a1c1038d23f" containerName="neutron-api" Feb 27 17:22:49 crc kubenswrapper[4700]: E0227 17:22:49.023673 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c48cd283-8c2c-4825-ad5e-c014c1fbc369" containerName="init" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.023678 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="c48cd283-8c2c-4825-ad5e-c014c1fbc369" containerName="init" Feb 27 17:22:49 crc kubenswrapper[4700]: E0227 17:22:49.023697 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c48cd283-8c2c-4825-ad5e-c014c1fbc369" containerName="dnsmasq-dns" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.023702 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="c48cd283-8c2c-4825-ad5e-c014c1fbc369" containerName="dnsmasq-dns" Feb 27 17:22:49 crc kubenswrapper[4700]: E0227 17:22:49.023714 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0d00b91-0e33-4466-8c9f-6a1c1038d23f" containerName="neutron-httpd" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.023720 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0d00b91-0e33-4466-8c9f-6a1c1038d23f" containerName="neutron-httpd" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.023884 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0d00b91-0e33-4466-8c9f-6a1c1038d23f" containerName="neutron-httpd" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.023893 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="e365b3f6-5e7d-4911-b7ee-728ca3d413dd" containerName="probe" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.023905 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="c48cd283-8c2c-4825-ad5e-c014c1fbc369" containerName="dnsmasq-dns" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.023912 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0d00b91-0e33-4466-8c9f-6a1c1038d23f" containerName="neutron-api" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.023920 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="e365b3f6-5e7d-4911-b7ee-728ca3d413dd" containerName="cinder-scheduler" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.024878 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.029822 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.039183 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.105499 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d9569ae4-5bf3-4720-bc48-0a1b735876ea-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d9569ae4-5bf3-4720-bc48-0a1b735876ea\") " pod="openstack/cinder-scheduler-0" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.105590 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d9569ae4-5bf3-4720-bc48-0a1b735876ea-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d9569ae4-5bf3-4720-bc48-0a1b735876ea\") " pod="openstack/cinder-scheduler-0" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.105630 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9569ae4-5bf3-4720-bc48-0a1b735876ea-config-data\") pod \"cinder-scheduler-0\" (UID: \"d9569ae4-5bf3-4720-bc48-0a1b735876ea\") " pod="openstack/cinder-scheduler-0" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.105664 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8c99q\" (UniqueName: \"kubernetes.io/projected/d9569ae4-5bf3-4720-bc48-0a1b735876ea-kube-api-access-8c99q\") pod \"cinder-scheduler-0\" (UID: \"d9569ae4-5bf3-4720-bc48-0a1b735876ea\") " pod="openstack/cinder-scheduler-0" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.105924 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9569ae4-5bf3-4720-bc48-0a1b735876ea-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d9569ae4-5bf3-4720-bc48-0a1b735876ea\") " pod="openstack/cinder-scheduler-0" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.105991 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9569ae4-5bf3-4720-bc48-0a1b735876ea-scripts\") pod \"cinder-scheduler-0\" (UID: \"d9569ae4-5bf3-4720-bc48-0a1b735876ea\") " pod="openstack/cinder-scheduler-0" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.209258 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d9569ae4-5bf3-4720-bc48-0a1b735876ea-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d9569ae4-5bf3-4720-bc48-0a1b735876ea\") " pod="openstack/cinder-scheduler-0" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.209612 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d9569ae4-5bf3-4720-bc48-0a1b735876ea-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d9569ae4-5bf3-4720-bc48-0a1b735876ea\") " pod="openstack/cinder-scheduler-0" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.209723 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9569ae4-5bf3-4720-bc48-0a1b735876ea-config-data\") pod \"cinder-scheduler-0\" (UID: \"d9569ae4-5bf3-4720-bc48-0a1b735876ea\") " pod="openstack/cinder-scheduler-0" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.209846 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8c99q\" (UniqueName: \"kubernetes.io/projected/d9569ae4-5bf3-4720-bc48-0a1b735876ea-kube-api-access-8c99q\") pod \"cinder-scheduler-0\" (UID: \"d9569ae4-5bf3-4720-bc48-0a1b735876ea\") " pod="openstack/cinder-scheduler-0" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.209983 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9569ae4-5bf3-4720-bc48-0a1b735876ea-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d9569ae4-5bf3-4720-bc48-0a1b735876ea\") " pod="openstack/cinder-scheduler-0" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.210088 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9569ae4-5bf3-4720-bc48-0a1b735876ea-scripts\") pod \"cinder-scheduler-0\" (UID: \"d9569ae4-5bf3-4720-bc48-0a1b735876ea\") " pod="openstack/cinder-scheduler-0" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.212558 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d9569ae4-5bf3-4720-bc48-0a1b735876ea-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d9569ae4-5bf3-4720-bc48-0a1b735876ea\") " pod="openstack/cinder-scheduler-0" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.214824 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9569ae4-5bf3-4720-bc48-0a1b735876ea-config-data\") pod \"cinder-scheduler-0\" (UID: \"d9569ae4-5bf3-4720-bc48-0a1b735876ea\") " pod="openstack/cinder-scheduler-0" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.232982 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8c99q\" (UniqueName: \"kubernetes.io/projected/d9569ae4-5bf3-4720-bc48-0a1b735876ea-kube-api-access-8c99q\") pod \"cinder-scheduler-0\" (UID: \"d9569ae4-5bf3-4720-bc48-0a1b735876ea\") " pod="openstack/cinder-scheduler-0" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.240734 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9569ae4-5bf3-4720-bc48-0a1b735876ea-scripts\") pod \"cinder-scheduler-0\" (UID: \"d9569ae4-5bf3-4720-bc48-0a1b735876ea\") " pod="openstack/cinder-scheduler-0" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.244340 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9569ae4-5bf3-4720-bc48-0a1b735876ea-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d9569ae4-5bf3-4720-bc48-0a1b735876ea\") " pod="openstack/cinder-scheduler-0" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.248991 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d9569ae4-5bf3-4720-bc48-0a1b735876ea-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d9569ae4-5bf3-4720-bc48-0a1b735876ea\") " pod="openstack/cinder-scheduler-0" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.291827 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-64fcf8677d-mvkvn" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.291876 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-64fcf8677d-mvkvn" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.387722 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.567743 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-cc78b9998-2c2dk"] Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.569782 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-cc78b9998-2c2dk" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.576655 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-cc78b9998-2c2dk"] Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.632093 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/779fb051-4fb5-4283-be4b-c46ff1602076-config-data\") pod \"placement-cc78b9998-2c2dk\" (UID: \"779fb051-4fb5-4283-be4b-c46ff1602076\") " pod="openstack/placement-cc78b9998-2c2dk" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.632403 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/779fb051-4fb5-4283-be4b-c46ff1602076-internal-tls-certs\") pod \"placement-cc78b9998-2c2dk\" (UID: \"779fb051-4fb5-4283-be4b-c46ff1602076\") " pod="openstack/placement-cc78b9998-2c2dk" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.632451 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/779fb051-4fb5-4283-be4b-c46ff1602076-combined-ca-bundle\") pod \"placement-cc78b9998-2c2dk\" (UID: \"779fb051-4fb5-4283-be4b-c46ff1602076\") " pod="openstack/placement-cc78b9998-2c2dk" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.632610 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjbz5\" (UniqueName: \"kubernetes.io/projected/779fb051-4fb5-4283-be4b-c46ff1602076-kube-api-access-bjbz5\") pod \"placement-cc78b9998-2c2dk\" (UID: \"779fb051-4fb5-4283-be4b-c46ff1602076\") " pod="openstack/placement-cc78b9998-2c2dk" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.632720 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/779fb051-4fb5-4283-be4b-c46ff1602076-logs\") pod \"placement-cc78b9998-2c2dk\" (UID: \"779fb051-4fb5-4283-be4b-c46ff1602076\") " pod="openstack/placement-cc78b9998-2c2dk" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.632760 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/779fb051-4fb5-4283-be4b-c46ff1602076-scripts\") pod \"placement-cc78b9998-2c2dk\" (UID: \"779fb051-4fb5-4283-be4b-c46ff1602076\") " pod="openstack/placement-cc78b9998-2c2dk" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.632803 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/779fb051-4fb5-4283-be4b-c46ff1602076-public-tls-certs\") pod \"placement-cc78b9998-2c2dk\" (UID: \"779fb051-4fb5-4283-be4b-c46ff1602076\") " pod="openstack/placement-cc78b9998-2c2dk" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.734174 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/779fb051-4fb5-4283-be4b-c46ff1602076-config-data\") pod \"placement-cc78b9998-2c2dk\" (UID: \"779fb051-4fb5-4283-be4b-c46ff1602076\") " pod="openstack/placement-cc78b9998-2c2dk" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.734240 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/779fb051-4fb5-4283-be4b-c46ff1602076-internal-tls-certs\") pod \"placement-cc78b9998-2c2dk\" (UID: \"779fb051-4fb5-4283-be4b-c46ff1602076\") " pod="openstack/placement-cc78b9998-2c2dk" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.734268 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/779fb051-4fb5-4283-be4b-c46ff1602076-combined-ca-bundle\") pod \"placement-cc78b9998-2c2dk\" (UID: \"779fb051-4fb5-4283-be4b-c46ff1602076\") " pod="openstack/placement-cc78b9998-2c2dk" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.734299 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjbz5\" (UniqueName: \"kubernetes.io/projected/779fb051-4fb5-4283-be4b-c46ff1602076-kube-api-access-bjbz5\") pod \"placement-cc78b9998-2c2dk\" (UID: \"779fb051-4fb5-4283-be4b-c46ff1602076\") " pod="openstack/placement-cc78b9998-2c2dk" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.734370 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/779fb051-4fb5-4283-be4b-c46ff1602076-logs\") pod \"placement-cc78b9998-2c2dk\" (UID: \"779fb051-4fb5-4283-be4b-c46ff1602076\") " pod="openstack/placement-cc78b9998-2c2dk" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.734400 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/779fb051-4fb5-4283-be4b-c46ff1602076-scripts\") pod \"placement-cc78b9998-2c2dk\" (UID: \"779fb051-4fb5-4283-be4b-c46ff1602076\") " pod="openstack/placement-cc78b9998-2c2dk" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.734421 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/779fb051-4fb5-4283-be4b-c46ff1602076-public-tls-certs\") pod \"placement-cc78b9998-2c2dk\" (UID: \"779fb051-4fb5-4283-be4b-c46ff1602076\") " pod="openstack/placement-cc78b9998-2c2dk" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.735557 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/779fb051-4fb5-4283-be4b-c46ff1602076-logs\") pod \"placement-cc78b9998-2c2dk\" (UID: \"779fb051-4fb5-4283-be4b-c46ff1602076\") " pod="openstack/placement-cc78b9998-2c2dk" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.740064 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/779fb051-4fb5-4283-be4b-c46ff1602076-public-tls-certs\") pod \"placement-cc78b9998-2c2dk\" (UID: \"779fb051-4fb5-4283-be4b-c46ff1602076\") " pod="openstack/placement-cc78b9998-2c2dk" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.740788 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/779fb051-4fb5-4283-be4b-c46ff1602076-scripts\") pod \"placement-cc78b9998-2c2dk\" (UID: \"779fb051-4fb5-4283-be4b-c46ff1602076\") " pod="openstack/placement-cc78b9998-2c2dk" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.744423 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/779fb051-4fb5-4283-be4b-c46ff1602076-combined-ca-bundle\") pod \"placement-cc78b9998-2c2dk\" (UID: \"779fb051-4fb5-4283-be4b-c46ff1602076\") " pod="openstack/placement-cc78b9998-2c2dk" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.745211 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/779fb051-4fb5-4283-be4b-c46ff1602076-config-data\") pod \"placement-cc78b9998-2c2dk\" (UID: \"779fb051-4fb5-4283-be4b-c46ff1602076\") " pod="openstack/placement-cc78b9998-2c2dk" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.745828 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/779fb051-4fb5-4283-be4b-c46ff1602076-internal-tls-certs\") pod \"placement-cc78b9998-2c2dk\" (UID: \"779fb051-4fb5-4283-be4b-c46ff1602076\") " pod="openstack/placement-cc78b9998-2c2dk" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.752908 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjbz5\" (UniqueName: \"kubernetes.io/projected/779fb051-4fb5-4283-be4b-c46ff1602076-kube-api-access-bjbz5\") pod \"placement-cc78b9998-2c2dk\" (UID: \"779fb051-4fb5-4283-be4b-c46ff1602076\") " pod="openstack/placement-cc78b9998-2c2dk" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.900300 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-cc78b9998-2c2dk" Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.917808 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 27 17:22:49 crc kubenswrapper[4700]: W0227 17:22:49.925546 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9569ae4_5bf3_4720_bc48_0a1b735876ea.slice/crio-5bbd2c531a4aedaad18cec386717765d60ee46882302c3ac7feb7badefd4e0d7 WatchSource:0}: Error finding container 5bbd2c531a4aedaad18cec386717765d60ee46882302c3ac7feb7badefd4e0d7: Status 404 returned error can't find the container with id 5bbd2c531a4aedaad18cec386717765d60ee46882302c3ac7feb7badefd4e0d7 Feb 27 17:22:49 crc kubenswrapper[4700]: I0227 17:22:49.977111 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d9569ae4-5bf3-4720-bc48-0a1b735876ea","Type":"ContainerStarted","Data":"5bbd2c531a4aedaad18cec386717765d60ee46882302c3ac7feb7badefd4e0d7"} Feb 27 17:22:50 crc kubenswrapper[4700]: I0227 17:22:50.390680 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-cc78b9998-2c2dk"] Feb 27 17:22:51 crc kubenswrapper[4700]: I0227 17:22:51.007602 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e365b3f6-5e7d-4911-b7ee-728ca3d413dd" path="/var/lib/kubelet/pods/e365b3f6-5e7d-4911-b7ee-728ca3d413dd/volumes" Feb 27 17:22:51 crc kubenswrapper[4700]: I0227 17:22:51.008695 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-cc78b9998-2c2dk" event={"ID":"779fb051-4fb5-4283-be4b-c46ff1602076","Type":"ContainerStarted","Data":"e4d16b6f6f24533cbeea7ffb10798fb921e7e26ab0af83423eb99b39442f9f23"} Feb 27 17:22:51 crc kubenswrapper[4700]: I0227 17:22:51.008723 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-cc78b9998-2c2dk" event={"ID":"779fb051-4fb5-4283-be4b-c46ff1602076","Type":"ContainerStarted","Data":"f8121212b16deab3a08eb88961cb97809288b4a69abf1032ee8672473dc8ed6f"} Feb 27 17:22:51 crc kubenswrapper[4700]: I0227 17:22:51.008736 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-cc78b9998-2c2dk" event={"ID":"779fb051-4fb5-4283-be4b-c46ff1602076","Type":"ContainerStarted","Data":"bd031f17c58fa8e2edcfa9b0cefa7ac68f11a5d35cd82e071f4478918ad77184"} Feb 27 17:22:51 crc kubenswrapper[4700]: I0227 17:22:51.008803 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-cc78b9998-2c2dk" Feb 27 17:22:51 crc kubenswrapper[4700]: I0227 17:22:51.008832 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-cc78b9998-2c2dk" Feb 27 17:22:51 crc kubenswrapper[4700]: I0227 17:22:51.011151 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d9569ae4-5bf3-4720-bc48-0a1b735876ea","Type":"ContainerStarted","Data":"d62b87162405454039d86c1377c353208b345762fd4d7b9a574f43e9994152ba"} Feb 27 17:22:51 crc kubenswrapper[4700]: I0227 17:22:51.111127 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-cc78b9998-2c2dk" podStartSLOduration=2.111103032 podStartE2EDuration="2.111103032s" podCreationTimestamp="2026-02-27 17:22:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:22:51.105826602 +0000 UTC m=+1331.091139349" watchObservedRunningTime="2026-02-27 17:22:51.111103032 +0000 UTC m=+1331.096415779" Feb 27 17:22:51 crc kubenswrapper[4700]: I0227 17:22:51.215020 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-7dcdff5d6d-9f49t" Feb 27 17:22:51 crc kubenswrapper[4700]: I0227 17:22:51.786123 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Feb 27 17:22:52 crc kubenswrapper[4700]: I0227 17:22:52.022781 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d9569ae4-5bf3-4720-bc48-0a1b735876ea","Type":"ContainerStarted","Data":"9974bc509946df337d803aaa3da2d371f9acb5c229cbe0f1f11aa7add47e2e1d"} Feb 27 17:22:52 crc kubenswrapper[4700]: I0227 17:22:52.047081 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.047064883 podStartE2EDuration="3.047064883s" podCreationTimestamp="2026-02-27 17:22:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:22:52.038197259 +0000 UTC m=+1332.023510016" watchObservedRunningTime="2026-02-27 17:22:52.047064883 +0000 UTC m=+1332.032377630" Feb 27 17:22:52 crc kubenswrapper[4700]: I0227 17:22:52.257347 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Feb 27 17:22:52 crc kubenswrapper[4700]: I0227 17:22:52.258517 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 27 17:22:52 crc kubenswrapper[4700]: I0227 17:22:52.263254 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Feb 27 17:22:52 crc kubenswrapper[4700]: I0227 17:22:52.263355 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-t68k2" Feb 27 17:22:52 crc kubenswrapper[4700]: I0227 17:22:52.263495 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Feb 27 17:22:52 crc kubenswrapper[4700]: I0227 17:22:52.268299 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 27 17:22:52 crc kubenswrapper[4700]: I0227 17:22:52.405099 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b2cd011d-8143-4800-ae5a-35c9d348892c-openstack-config\") pod \"openstackclient\" (UID: \"b2cd011d-8143-4800-ae5a-35c9d348892c\") " pod="openstack/openstackclient" Feb 27 17:22:52 crc kubenswrapper[4700]: I0227 17:22:52.405898 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b2cd011d-8143-4800-ae5a-35c9d348892c-openstack-config-secret\") pod \"openstackclient\" (UID: \"b2cd011d-8143-4800-ae5a-35c9d348892c\") " pod="openstack/openstackclient" Feb 27 17:22:52 crc kubenswrapper[4700]: I0227 17:22:52.406099 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnpwd\" (UniqueName: \"kubernetes.io/projected/b2cd011d-8143-4800-ae5a-35c9d348892c-kube-api-access-vnpwd\") pod \"openstackclient\" (UID: \"b2cd011d-8143-4800-ae5a-35c9d348892c\") " pod="openstack/openstackclient" Feb 27 17:22:52 crc kubenswrapper[4700]: I0227 17:22:52.406333 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cd011d-8143-4800-ae5a-35c9d348892c-combined-ca-bundle\") pod \"openstackclient\" (UID: \"b2cd011d-8143-4800-ae5a-35c9d348892c\") " pod="openstack/openstackclient" Feb 27 17:22:52 crc kubenswrapper[4700]: I0227 17:22:52.508215 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b2cd011d-8143-4800-ae5a-35c9d348892c-openstack-config-secret\") pod \"openstackclient\" (UID: \"b2cd011d-8143-4800-ae5a-35c9d348892c\") " pod="openstack/openstackclient" Feb 27 17:22:52 crc kubenswrapper[4700]: I0227 17:22:52.508633 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnpwd\" (UniqueName: \"kubernetes.io/projected/b2cd011d-8143-4800-ae5a-35c9d348892c-kube-api-access-vnpwd\") pod \"openstackclient\" (UID: \"b2cd011d-8143-4800-ae5a-35c9d348892c\") " pod="openstack/openstackclient" Feb 27 17:22:52 crc kubenswrapper[4700]: I0227 17:22:52.508890 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cd011d-8143-4800-ae5a-35c9d348892c-combined-ca-bundle\") pod \"openstackclient\" (UID: \"b2cd011d-8143-4800-ae5a-35c9d348892c\") " pod="openstack/openstackclient" Feb 27 17:22:52 crc kubenswrapper[4700]: I0227 17:22:52.509217 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b2cd011d-8143-4800-ae5a-35c9d348892c-openstack-config\") pod \"openstackclient\" (UID: \"b2cd011d-8143-4800-ae5a-35c9d348892c\") " pod="openstack/openstackclient" Feb 27 17:22:52 crc kubenswrapper[4700]: I0227 17:22:52.510049 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b2cd011d-8143-4800-ae5a-35c9d348892c-openstack-config\") pod \"openstackclient\" (UID: \"b2cd011d-8143-4800-ae5a-35c9d348892c\") " pod="openstack/openstackclient" Feb 27 17:22:52 crc kubenswrapper[4700]: I0227 17:22:52.517265 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2cd011d-8143-4800-ae5a-35c9d348892c-combined-ca-bundle\") pod \"openstackclient\" (UID: \"b2cd011d-8143-4800-ae5a-35c9d348892c\") " pod="openstack/openstackclient" Feb 27 17:22:52 crc kubenswrapper[4700]: I0227 17:22:52.517491 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b2cd011d-8143-4800-ae5a-35c9d348892c-openstack-config-secret\") pod \"openstackclient\" (UID: \"b2cd011d-8143-4800-ae5a-35c9d348892c\") " pod="openstack/openstackclient" Feb 27 17:22:52 crc kubenswrapper[4700]: I0227 17:22:52.527197 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnpwd\" (UniqueName: \"kubernetes.io/projected/b2cd011d-8143-4800-ae5a-35c9d348892c-kube-api-access-vnpwd\") pod \"openstackclient\" (UID: \"b2cd011d-8143-4800-ae5a-35c9d348892c\") " pod="openstack/openstackclient" Feb 27 17:22:52 crc kubenswrapper[4700]: I0227 17:22:52.575025 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 27 17:22:53 crc kubenswrapper[4700]: I0227 17:22:53.069440 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 27 17:22:54 crc kubenswrapper[4700]: I0227 17:22:54.041138 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"b2cd011d-8143-4800-ae5a-35c9d348892c","Type":"ContainerStarted","Data":"a9369ac9c26581c09825ac2826399ca8733a31dd27fac418611e9a20c41f2624"} Feb 27 17:22:54 crc kubenswrapper[4700]: I0227 17:22:54.389032 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.319624 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-xf6ff"] Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.322804 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-xf6ff" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.341019 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-xf6ff"] Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.394696 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgcv8\" (UniqueName: \"kubernetes.io/projected/b931c6bc-3ec1-493d-ad2e-21008ce1d1f5-kube-api-access-zgcv8\") pod \"nova-api-db-create-xf6ff\" (UID: \"b931c6bc-3ec1-493d-ad2e-21008ce1d1f5\") " pod="openstack/nova-api-db-create-xf6ff" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.394797 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b931c6bc-3ec1-493d-ad2e-21008ce1d1f5-operator-scripts\") pod \"nova-api-db-create-xf6ff\" (UID: \"b931c6bc-3ec1-493d-ad2e-21008ce1d1f5\") " pod="openstack/nova-api-db-create-xf6ff" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.487628 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-829r9"] Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.496743 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgcv8\" (UniqueName: \"kubernetes.io/projected/b931c6bc-3ec1-493d-ad2e-21008ce1d1f5-kube-api-access-zgcv8\") pod \"nova-api-db-create-xf6ff\" (UID: \"b931c6bc-3ec1-493d-ad2e-21008ce1d1f5\") " pod="openstack/nova-api-db-create-xf6ff" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.496821 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b931c6bc-3ec1-493d-ad2e-21008ce1d1f5-operator-scripts\") pod \"nova-api-db-create-xf6ff\" (UID: \"b931c6bc-3ec1-493d-ad2e-21008ce1d1f5\") " pod="openstack/nova-api-db-create-xf6ff" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.497444 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b931c6bc-3ec1-493d-ad2e-21008ce1d1f5-operator-scripts\") pod \"nova-api-db-create-xf6ff\" (UID: \"b931c6bc-3ec1-493d-ad2e-21008ce1d1f5\") " pod="openstack/nova-api-db-create-xf6ff" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.505994 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-829r9" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.523162 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-829r9"] Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.539385 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgcv8\" (UniqueName: \"kubernetes.io/projected/b931c6bc-3ec1-493d-ad2e-21008ce1d1f5-kube-api-access-zgcv8\") pod \"nova-api-db-create-xf6ff\" (UID: \"b931c6bc-3ec1-493d-ad2e-21008ce1d1f5\") " pod="openstack/nova-api-db-create-xf6ff" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.542945 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-6728-account-create-update-swj84"] Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.544505 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-6728-account-create-update-swj84" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.546643 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.569941 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-6728-account-create-update-swj84"] Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.598080 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n96xj\" (UniqueName: \"kubernetes.io/projected/c22fb154-217c-462c-a0a6-4b67b5bb11d2-kube-api-access-n96xj\") pod \"nova-cell0-db-create-829r9\" (UID: \"c22fb154-217c-462c-a0a6-4b67b5bb11d2\") " pod="openstack/nova-cell0-db-create-829r9" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.598153 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kb972\" (UniqueName: \"kubernetes.io/projected/26947d30-e5b4-4a9f-9a14-c41185ff5a4f-kube-api-access-kb972\") pod \"nova-api-6728-account-create-update-swj84\" (UID: \"26947d30-e5b4-4a9f-9a14-c41185ff5a4f\") " pod="openstack/nova-api-6728-account-create-update-swj84" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.598181 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/26947d30-e5b4-4a9f-9a14-c41185ff5a4f-operator-scripts\") pod \"nova-api-6728-account-create-update-swj84\" (UID: \"26947d30-e5b4-4a9f-9a14-c41185ff5a4f\") " pod="openstack/nova-api-6728-account-create-update-swj84" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.598310 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c22fb154-217c-462c-a0a6-4b67b5bb11d2-operator-scripts\") pod \"nova-cell0-db-create-829r9\" (UID: \"c22fb154-217c-462c-a0a6-4b67b5bb11d2\") " pod="openstack/nova-cell0-db-create-829r9" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.627379 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-xr75f"] Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.628633 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xr75f" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.649246 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-xr75f"] Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.650392 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-xf6ff" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.696956 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-7178-account-create-update-ttwhv"] Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.698407 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-7178-account-create-update-ttwhv" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.700120 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c22fb154-217c-462c-a0a6-4b67b5bb11d2-operator-scripts\") pod \"nova-cell0-db-create-829r9\" (UID: \"c22fb154-217c-462c-a0a6-4b67b5bb11d2\") " pod="openstack/nova-cell0-db-create-829r9" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.700168 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnkft\" (UniqueName: \"kubernetes.io/projected/05664c1a-c691-4925-a4e6-6fdc12b7998a-kube-api-access-rnkft\") pod \"nova-cell1-db-create-xr75f\" (UID: \"05664c1a-c691-4925-a4e6-6fdc12b7998a\") " pod="openstack/nova-cell1-db-create-xr75f" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.700199 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n96xj\" (UniqueName: \"kubernetes.io/projected/c22fb154-217c-462c-a0a6-4b67b5bb11d2-kube-api-access-n96xj\") pod \"nova-cell0-db-create-829r9\" (UID: \"c22fb154-217c-462c-a0a6-4b67b5bb11d2\") " pod="openstack/nova-cell0-db-create-829r9" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.700219 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kb972\" (UniqueName: \"kubernetes.io/projected/26947d30-e5b4-4a9f-9a14-c41185ff5a4f-kube-api-access-kb972\") pod \"nova-api-6728-account-create-update-swj84\" (UID: \"26947d30-e5b4-4a9f-9a14-c41185ff5a4f\") " pod="openstack/nova-api-6728-account-create-update-swj84" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.700239 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/26947d30-e5b4-4a9f-9a14-c41185ff5a4f-operator-scripts\") pod \"nova-api-6728-account-create-update-swj84\" (UID: \"26947d30-e5b4-4a9f-9a14-c41185ff5a4f\") " pod="openstack/nova-api-6728-account-create-update-swj84" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.700950 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05664c1a-c691-4925-a4e6-6fdc12b7998a-operator-scripts\") pod \"nova-cell1-db-create-xr75f\" (UID: \"05664c1a-c691-4925-a4e6-6fdc12b7998a\") " pod="openstack/nova-cell1-db-create-xr75f" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.706963 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.708614 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c22fb154-217c-462c-a0a6-4b67b5bb11d2-operator-scripts\") pod \"nova-cell0-db-create-829r9\" (UID: \"c22fb154-217c-462c-a0a6-4b67b5bb11d2\") " pod="openstack/nova-cell0-db-create-829r9" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.708651 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/26947d30-e5b4-4a9f-9a14-c41185ff5a4f-operator-scripts\") pod \"nova-api-6728-account-create-update-swj84\" (UID: \"26947d30-e5b4-4a9f-9a14-c41185ff5a4f\") " pod="openstack/nova-api-6728-account-create-update-swj84" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.726895 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kb972\" (UniqueName: \"kubernetes.io/projected/26947d30-e5b4-4a9f-9a14-c41185ff5a4f-kube-api-access-kb972\") pod \"nova-api-6728-account-create-update-swj84\" (UID: \"26947d30-e5b4-4a9f-9a14-c41185ff5a4f\") " pod="openstack/nova-api-6728-account-create-update-swj84" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.728059 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n96xj\" (UniqueName: \"kubernetes.io/projected/c22fb154-217c-462c-a0a6-4b67b5bb11d2-kube-api-access-n96xj\") pod \"nova-cell0-db-create-829r9\" (UID: \"c22fb154-217c-462c-a0a6-4b67b5bb11d2\") " pod="openstack/nova-cell0-db-create-829r9" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.737898 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-7178-account-create-update-ttwhv"] Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.808269 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04cd0229-5bc9-424e-ad73-9b3f9eb3de8d-operator-scripts\") pod \"nova-cell0-7178-account-create-update-ttwhv\" (UID: \"04cd0229-5bc9-424e-ad73-9b3f9eb3de8d\") " pod="openstack/nova-cell0-7178-account-create-update-ttwhv" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.808348 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnkft\" (UniqueName: \"kubernetes.io/projected/05664c1a-c691-4925-a4e6-6fdc12b7998a-kube-api-access-rnkft\") pod \"nova-cell1-db-create-xr75f\" (UID: \"05664c1a-c691-4925-a4e6-6fdc12b7998a\") " pod="openstack/nova-cell1-db-create-xr75f" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.808601 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhfxj\" (UniqueName: \"kubernetes.io/projected/04cd0229-5bc9-424e-ad73-9b3f9eb3de8d-kube-api-access-zhfxj\") pod \"nova-cell0-7178-account-create-update-ttwhv\" (UID: \"04cd0229-5bc9-424e-ad73-9b3f9eb3de8d\") " pod="openstack/nova-cell0-7178-account-create-update-ttwhv" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.808710 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05664c1a-c691-4925-a4e6-6fdc12b7998a-operator-scripts\") pod \"nova-cell1-db-create-xr75f\" (UID: \"05664c1a-c691-4925-a4e6-6fdc12b7998a\") " pod="openstack/nova-cell1-db-create-xr75f" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.809511 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05664c1a-c691-4925-a4e6-6fdc12b7998a-operator-scripts\") pod \"nova-cell1-db-create-xr75f\" (UID: \"05664c1a-c691-4925-a4e6-6fdc12b7998a\") " pod="openstack/nova-cell1-db-create-xr75f" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.823670 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnkft\" (UniqueName: \"kubernetes.io/projected/05664c1a-c691-4925-a4e6-6fdc12b7998a-kube-api-access-rnkft\") pod \"nova-cell1-db-create-xr75f\" (UID: \"05664c1a-c691-4925-a4e6-6fdc12b7998a\") " pod="openstack/nova-cell1-db-create-xr75f" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.894826 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-829r9" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.902946 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-4603-account-create-update-4tz7h"] Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.904312 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-6728-account-create-update-swj84" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.904740 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4603-account-create-update-4tz7h" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.907387 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.911019 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhfxj\" (UniqueName: \"kubernetes.io/projected/04cd0229-5bc9-424e-ad73-9b3f9eb3de8d-kube-api-access-zhfxj\") pod \"nova-cell0-7178-account-create-update-ttwhv\" (UID: \"04cd0229-5bc9-424e-ad73-9b3f9eb3de8d\") " pod="openstack/nova-cell0-7178-account-create-update-ttwhv" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.911256 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04cd0229-5bc9-424e-ad73-9b3f9eb3de8d-operator-scripts\") pod \"nova-cell0-7178-account-create-update-ttwhv\" (UID: \"04cd0229-5bc9-424e-ad73-9b3f9eb3de8d\") " pod="openstack/nova-cell0-7178-account-create-update-ttwhv" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.912344 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04cd0229-5bc9-424e-ad73-9b3f9eb3de8d-operator-scripts\") pod \"nova-cell0-7178-account-create-update-ttwhv\" (UID: \"04cd0229-5bc9-424e-ad73-9b3f9eb3de8d\") " pod="openstack/nova-cell0-7178-account-create-update-ttwhv" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.920018 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-4603-account-create-update-4tz7h"] Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.934875 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhfxj\" (UniqueName: \"kubernetes.io/projected/04cd0229-5bc9-424e-ad73-9b3f9eb3de8d-kube-api-access-zhfxj\") pod \"nova-cell0-7178-account-create-update-ttwhv\" (UID: \"04cd0229-5bc9-424e-ad73-9b3f9eb3de8d\") " pod="openstack/nova-cell0-7178-account-create-update-ttwhv" Feb 27 17:22:56 crc kubenswrapper[4700]: I0227 17:22:56.957132 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xr75f" Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.012966 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d646da10-fce6-4960-aedd-6aebd88f075b-operator-scripts\") pod \"nova-cell1-4603-account-create-update-4tz7h\" (UID: \"d646da10-fce6-4960-aedd-6aebd88f075b\") " pod="openstack/nova-cell1-4603-account-create-update-4tz7h" Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.013046 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsdg4\" (UniqueName: \"kubernetes.io/projected/d646da10-fce6-4960-aedd-6aebd88f075b-kube-api-access-gsdg4\") pod \"nova-cell1-4603-account-create-update-4tz7h\" (UID: \"d646da10-fce6-4960-aedd-6aebd88f075b\") " pod="openstack/nova-cell1-4603-account-create-update-4tz7h" Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.115521 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d646da10-fce6-4960-aedd-6aebd88f075b-operator-scripts\") pod \"nova-cell1-4603-account-create-update-4tz7h\" (UID: \"d646da10-fce6-4960-aedd-6aebd88f075b\") " pod="openstack/nova-cell1-4603-account-create-update-4tz7h" Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.115913 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsdg4\" (UniqueName: \"kubernetes.io/projected/d646da10-fce6-4960-aedd-6aebd88f075b-kube-api-access-gsdg4\") pod \"nova-cell1-4603-account-create-update-4tz7h\" (UID: \"d646da10-fce6-4960-aedd-6aebd88f075b\") " pod="openstack/nova-cell1-4603-account-create-update-4tz7h" Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.116324 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d646da10-fce6-4960-aedd-6aebd88f075b-operator-scripts\") pod \"nova-cell1-4603-account-create-update-4tz7h\" (UID: \"d646da10-fce6-4960-aedd-6aebd88f075b\") " pod="openstack/nova-cell1-4603-account-create-update-4tz7h" Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.118692 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-7178-account-create-update-ttwhv" Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.131760 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsdg4\" (UniqueName: \"kubernetes.io/projected/d646da10-fce6-4960-aedd-6aebd88f075b-kube-api-access-gsdg4\") pod \"nova-cell1-4603-account-create-update-4tz7h\" (UID: \"d646da10-fce6-4960-aedd-6aebd88f075b\") " pod="openstack/nova-cell1-4603-account-create-update-4tz7h" Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.232242 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4603-account-create-update-4tz7h" Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.547761 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-65457d8799-dclxm"] Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.549512 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-65457d8799-dclxm" Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.552357 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.552408 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.552613 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.563548 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-65457d8799-dclxm"] Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.625652 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e43c5d5e-63ef-45b9-af4e-627a035b376e-combined-ca-bundle\") pod \"swift-proxy-65457d8799-dclxm\" (UID: \"e43c5d5e-63ef-45b9-af4e-627a035b376e\") " pod="openstack/swift-proxy-65457d8799-dclxm" Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.625691 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e43c5d5e-63ef-45b9-af4e-627a035b376e-public-tls-certs\") pod \"swift-proxy-65457d8799-dclxm\" (UID: \"e43c5d5e-63ef-45b9-af4e-627a035b376e\") " pod="openstack/swift-proxy-65457d8799-dclxm" Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.625716 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e43c5d5e-63ef-45b9-af4e-627a035b376e-log-httpd\") pod \"swift-proxy-65457d8799-dclxm\" (UID: \"e43c5d5e-63ef-45b9-af4e-627a035b376e\") " pod="openstack/swift-proxy-65457d8799-dclxm" Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.625743 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e43c5d5e-63ef-45b9-af4e-627a035b376e-config-data\") pod \"swift-proxy-65457d8799-dclxm\" (UID: \"e43c5d5e-63ef-45b9-af4e-627a035b376e\") " pod="openstack/swift-proxy-65457d8799-dclxm" Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.625785 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e43c5d5e-63ef-45b9-af4e-627a035b376e-internal-tls-certs\") pod \"swift-proxy-65457d8799-dclxm\" (UID: \"e43c5d5e-63ef-45b9-af4e-627a035b376e\") " pod="openstack/swift-proxy-65457d8799-dclxm" Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.625817 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e43c5d5e-63ef-45b9-af4e-627a035b376e-etc-swift\") pod \"swift-proxy-65457d8799-dclxm\" (UID: \"e43c5d5e-63ef-45b9-af4e-627a035b376e\") " pod="openstack/swift-proxy-65457d8799-dclxm" Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.625851 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e43c5d5e-63ef-45b9-af4e-627a035b376e-run-httpd\") pod \"swift-proxy-65457d8799-dclxm\" (UID: \"e43c5d5e-63ef-45b9-af4e-627a035b376e\") " pod="openstack/swift-proxy-65457d8799-dclxm" Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.625887 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4p99\" (UniqueName: \"kubernetes.io/projected/e43c5d5e-63ef-45b9-af4e-627a035b376e-kube-api-access-r4p99\") pod \"swift-proxy-65457d8799-dclxm\" (UID: \"e43c5d5e-63ef-45b9-af4e-627a035b376e\") " pod="openstack/swift-proxy-65457d8799-dclxm" Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.726946 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e43c5d5e-63ef-45b9-af4e-627a035b376e-internal-tls-certs\") pod \"swift-proxy-65457d8799-dclxm\" (UID: \"e43c5d5e-63ef-45b9-af4e-627a035b376e\") " pod="openstack/swift-proxy-65457d8799-dclxm" Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.727008 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e43c5d5e-63ef-45b9-af4e-627a035b376e-etc-swift\") pod \"swift-proxy-65457d8799-dclxm\" (UID: \"e43c5d5e-63ef-45b9-af4e-627a035b376e\") " pod="openstack/swift-proxy-65457d8799-dclxm" Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.727051 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e43c5d5e-63ef-45b9-af4e-627a035b376e-run-httpd\") pod \"swift-proxy-65457d8799-dclxm\" (UID: \"e43c5d5e-63ef-45b9-af4e-627a035b376e\") " pod="openstack/swift-proxy-65457d8799-dclxm" Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.727091 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4p99\" (UniqueName: \"kubernetes.io/projected/e43c5d5e-63ef-45b9-af4e-627a035b376e-kube-api-access-r4p99\") pod \"swift-proxy-65457d8799-dclxm\" (UID: \"e43c5d5e-63ef-45b9-af4e-627a035b376e\") " pod="openstack/swift-proxy-65457d8799-dclxm" Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.727160 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e43c5d5e-63ef-45b9-af4e-627a035b376e-combined-ca-bundle\") pod \"swift-proxy-65457d8799-dclxm\" (UID: \"e43c5d5e-63ef-45b9-af4e-627a035b376e\") " pod="openstack/swift-proxy-65457d8799-dclxm" Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.727178 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e43c5d5e-63ef-45b9-af4e-627a035b376e-public-tls-certs\") pod \"swift-proxy-65457d8799-dclxm\" (UID: \"e43c5d5e-63ef-45b9-af4e-627a035b376e\") " pod="openstack/swift-proxy-65457d8799-dclxm" Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.727201 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e43c5d5e-63ef-45b9-af4e-627a035b376e-log-httpd\") pod \"swift-proxy-65457d8799-dclxm\" (UID: \"e43c5d5e-63ef-45b9-af4e-627a035b376e\") " pod="openstack/swift-proxy-65457d8799-dclxm" Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.727225 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e43c5d5e-63ef-45b9-af4e-627a035b376e-config-data\") pod \"swift-proxy-65457d8799-dclxm\" (UID: \"e43c5d5e-63ef-45b9-af4e-627a035b376e\") " pod="openstack/swift-proxy-65457d8799-dclxm" Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.727816 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e43c5d5e-63ef-45b9-af4e-627a035b376e-run-httpd\") pod \"swift-proxy-65457d8799-dclxm\" (UID: \"e43c5d5e-63ef-45b9-af4e-627a035b376e\") " pod="openstack/swift-proxy-65457d8799-dclxm" Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.728267 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e43c5d5e-63ef-45b9-af4e-627a035b376e-log-httpd\") pod \"swift-proxy-65457d8799-dclxm\" (UID: \"e43c5d5e-63ef-45b9-af4e-627a035b376e\") " pod="openstack/swift-proxy-65457d8799-dclxm" Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.732193 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e43c5d5e-63ef-45b9-af4e-627a035b376e-combined-ca-bundle\") pod \"swift-proxy-65457d8799-dclxm\" (UID: \"e43c5d5e-63ef-45b9-af4e-627a035b376e\") " pod="openstack/swift-proxy-65457d8799-dclxm" Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.732683 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e43c5d5e-63ef-45b9-af4e-627a035b376e-internal-tls-certs\") pod \"swift-proxy-65457d8799-dclxm\" (UID: \"e43c5d5e-63ef-45b9-af4e-627a035b376e\") " pod="openstack/swift-proxy-65457d8799-dclxm" Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.735690 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e43c5d5e-63ef-45b9-af4e-627a035b376e-public-tls-certs\") pod \"swift-proxy-65457d8799-dclxm\" (UID: \"e43c5d5e-63ef-45b9-af4e-627a035b376e\") " pod="openstack/swift-proxy-65457d8799-dclxm" Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.736118 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e43c5d5e-63ef-45b9-af4e-627a035b376e-etc-swift\") pod \"swift-proxy-65457d8799-dclxm\" (UID: \"e43c5d5e-63ef-45b9-af4e-627a035b376e\") " pod="openstack/swift-proxy-65457d8799-dclxm" Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.739278 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e43c5d5e-63ef-45b9-af4e-627a035b376e-config-data\") pod \"swift-proxy-65457d8799-dclxm\" (UID: \"e43c5d5e-63ef-45b9-af4e-627a035b376e\") " pod="openstack/swift-proxy-65457d8799-dclxm" Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.743704 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4p99\" (UniqueName: \"kubernetes.io/projected/e43c5d5e-63ef-45b9-af4e-627a035b376e-kube-api-access-r4p99\") pod \"swift-proxy-65457d8799-dclxm\" (UID: \"e43c5d5e-63ef-45b9-af4e-627a035b376e\") " pod="openstack/swift-proxy-65457d8799-dclxm" Feb 27 17:22:57 crc kubenswrapper[4700]: I0227 17:22:57.896726 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-65457d8799-dclxm" Feb 27 17:22:58 crc kubenswrapper[4700]: I0227 17:22:58.269243 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5b48777cdd-tqbkj" podUID="d3989a3e-45ff-45d8-80a2-4dbe57663a57" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.168:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.168:8443: connect: connection refused" Feb 27 17:22:58 crc kubenswrapper[4700]: I0227 17:22:58.709113 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:22:58 crc kubenswrapper[4700]: I0227 17:22:58.709425 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2230bd83-cd33-438b-9d9d-001e4727fb66" containerName="ceilometer-central-agent" containerID="cri-o://939cdbfe2666a6be220f5f5345c486c74663161256917db6f4f375d293727904" gracePeriod=30 Feb 27 17:22:58 crc kubenswrapper[4700]: I0227 17:22:58.709589 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2230bd83-cd33-438b-9d9d-001e4727fb66" containerName="proxy-httpd" containerID="cri-o://8e03101f296c609397f67f6db25ba89c6efef855a814df305c7288fddff2ce5b" gracePeriod=30 Feb 27 17:22:58 crc kubenswrapper[4700]: I0227 17:22:58.709642 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2230bd83-cd33-438b-9d9d-001e4727fb66" containerName="sg-core" containerID="cri-o://3ad4aca4ec61405efe35f12a970058daaf58cd3480a4fc6f5a2a641fae53761e" gracePeriod=30 Feb 27 17:22:58 crc kubenswrapper[4700]: I0227 17:22:58.709684 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2230bd83-cd33-438b-9d9d-001e4727fb66" containerName="ceilometer-notification-agent" containerID="cri-o://beb916ec8dfdea917abac29c24892a3eb645748ade539e9616642050bb73dace" gracePeriod=30 Feb 27 17:22:58 crc kubenswrapper[4700]: I0227 17:22:58.716756 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="2230bd83-cd33-438b-9d9d-001e4727fb66" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.191:3000/\": EOF" Feb 27 17:22:59 crc kubenswrapper[4700]: I0227 17:22:59.106765 4700 generic.go:334] "Generic (PLEG): container finished" podID="2230bd83-cd33-438b-9d9d-001e4727fb66" containerID="8e03101f296c609397f67f6db25ba89c6efef855a814df305c7288fddff2ce5b" exitCode=0 Feb 27 17:22:59 crc kubenswrapper[4700]: I0227 17:22:59.106797 4700 generic.go:334] "Generic (PLEG): container finished" podID="2230bd83-cd33-438b-9d9d-001e4727fb66" containerID="3ad4aca4ec61405efe35f12a970058daaf58cd3480a4fc6f5a2a641fae53761e" exitCode=2 Feb 27 17:22:59 crc kubenswrapper[4700]: I0227 17:22:59.106816 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2230bd83-cd33-438b-9d9d-001e4727fb66","Type":"ContainerDied","Data":"8e03101f296c609397f67f6db25ba89c6efef855a814df305c7288fddff2ce5b"} Feb 27 17:22:59 crc kubenswrapper[4700]: I0227 17:22:59.106841 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2230bd83-cd33-438b-9d9d-001e4727fb66","Type":"ContainerDied","Data":"3ad4aca4ec61405efe35f12a970058daaf58cd3480a4fc6f5a2a641fae53761e"} Feb 27 17:22:59 crc kubenswrapper[4700]: I0227 17:22:59.530781 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 27 17:22:59 crc kubenswrapper[4700]: I0227 17:22:59.981484 4700 scope.go:117] "RemoveContainer" containerID="9b51d27b3971cd3f655c413b4bd270b5cc55c1c7f20422db504613285c3a5770" Feb 27 17:22:59 crc kubenswrapper[4700]: E0227 17:22:59.981840 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 20s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(0f1ae433-28e4-45cf-9635-dc6edbf16f6e)\"" pod="openstack/watcher-decision-engine-0" podUID="0f1ae433-28e4-45cf-9635-dc6edbf16f6e" Feb 27 17:23:00 crc kubenswrapper[4700]: I0227 17:23:00.139528 4700 generic.go:334] "Generic (PLEG): container finished" podID="2230bd83-cd33-438b-9d9d-001e4727fb66" containerID="939cdbfe2666a6be220f5f5345c486c74663161256917db6f4f375d293727904" exitCode=0 Feb 27 17:23:00 crc kubenswrapper[4700]: I0227 17:23:00.139583 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2230bd83-cd33-438b-9d9d-001e4727fb66","Type":"ContainerDied","Data":"939cdbfe2666a6be220f5f5345c486c74663161256917db6f4f375d293727904"} Feb 27 17:23:01 crc kubenswrapper[4700]: I0227 17:23:01.152248 4700 generic.go:334] "Generic (PLEG): container finished" podID="2230bd83-cd33-438b-9d9d-001e4727fb66" containerID="beb916ec8dfdea917abac29c24892a3eb645748ade539e9616642050bb73dace" exitCode=0 Feb 27 17:23:01 crc kubenswrapper[4700]: I0227 17:23:01.152322 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2230bd83-cd33-438b-9d9d-001e4727fb66","Type":"ContainerDied","Data":"beb916ec8dfdea917abac29c24892a3eb645748ade539e9616642050bb73dace"} Feb 27 17:23:02 crc kubenswrapper[4700]: I0227 17:23:02.480765 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 17:23:02 crc kubenswrapper[4700]: I0227 17:23:02.530191 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2230bd83-cd33-438b-9d9d-001e4727fb66-scripts\") pod \"2230bd83-cd33-438b-9d9d-001e4727fb66\" (UID: \"2230bd83-cd33-438b-9d9d-001e4727fb66\") " Feb 27 17:23:02 crc kubenswrapper[4700]: I0227 17:23:02.530279 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2230bd83-cd33-438b-9d9d-001e4727fb66-config-data\") pod \"2230bd83-cd33-438b-9d9d-001e4727fb66\" (UID: \"2230bd83-cd33-438b-9d9d-001e4727fb66\") " Feb 27 17:23:02 crc kubenswrapper[4700]: I0227 17:23:02.530305 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2230bd83-cd33-438b-9d9d-001e4727fb66-combined-ca-bundle\") pod \"2230bd83-cd33-438b-9d9d-001e4727fb66\" (UID: \"2230bd83-cd33-438b-9d9d-001e4727fb66\") " Feb 27 17:23:02 crc kubenswrapper[4700]: I0227 17:23:02.530386 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2230bd83-cd33-438b-9d9d-001e4727fb66-run-httpd\") pod \"2230bd83-cd33-438b-9d9d-001e4727fb66\" (UID: \"2230bd83-cd33-438b-9d9d-001e4727fb66\") " Feb 27 17:23:02 crc kubenswrapper[4700]: I0227 17:23:02.530454 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87fx7\" (UniqueName: \"kubernetes.io/projected/2230bd83-cd33-438b-9d9d-001e4727fb66-kube-api-access-87fx7\") pod \"2230bd83-cd33-438b-9d9d-001e4727fb66\" (UID: \"2230bd83-cd33-438b-9d9d-001e4727fb66\") " Feb 27 17:23:02 crc kubenswrapper[4700]: I0227 17:23:02.530564 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2230bd83-cd33-438b-9d9d-001e4727fb66-log-httpd\") pod \"2230bd83-cd33-438b-9d9d-001e4727fb66\" (UID: \"2230bd83-cd33-438b-9d9d-001e4727fb66\") " Feb 27 17:23:02 crc kubenswrapper[4700]: I0227 17:23:02.531052 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2230bd83-cd33-438b-9d9d-001e4727fb66-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2230bd83-cd33-438b-9d9d-001e4727fb66" (UID: "2230bd83-cd33-438b-9d9d-001e4727fb66"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:23:02 crc kubenswrapper[4700]: I0227 17:23:02.531094 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2230bd83-cd33-438b-9d9d-001e4727fb66-sg-core-conf-yaml\") pod \"2230bd83-cd33-438b-9d9d-001e4727fb66\" (UID: \"2230bd83-cd33-438b-9d9d-001e4727fb66\") " Feb 27 17:23:02 crc kubenswrapper[4700]: I0227 17:23:02.531281 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2230bd83-cd33-438b-9d9d-001e4727fb66-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2230bd83-cd33-438b-9d9d-001e4727fb66" (UID: "2230bd83-cd33-438b-9d9d-001e4727fb66"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:23:02 crc kubenswrapper[4700]: I0227 17:23:02.531670 4700 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2230bd83-cd33-438b-9d9d-001e4727fb66-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:02 crc kubenswrapper[4700]: I0227 17:23:02.531693 4700 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2230bd83-cd33-438b-9d9d-001e4727fb66-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:02 crc kubenswrapper[4700]: I0227 17:23:02.536914 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2230bd83-cd33-438b-9d9d-001e4727fb66-scripts" (OuterVolumeSpecName: "scripts") pod "2230bd83-cd33-438b-9d9d-001e4727fb66" (UID: "2230bd83-cd33-438b-9d9d-001e4727fb66"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:02 crc kubenswrapper[4700]: I0227 17:23:02.537030 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2230bd83-cd33-438b-9d9d-001e4727fb66-kube-api-access-87fx7" (OuterVolumeSpecName: "kube-api-access-87fx7") pod "2230bd83-cd33-438b-9d9d-001e4727fb66" (UID: "2230bd83-cd33-438b-9d9d-001e4727fb66"). InnerVolumeSpecName "kube-api-access-87fx7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:23:02 crc kubenswrapper[4700]: I0227 17:23:02.558412 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2230bd83-cd33-438b-9d9d-001e4727fb66-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2230bd83-cd33-438b-9d9d-001e4727fb66" (UID: "2230bd83-cd33-438b-9d9d-001e4727fb66"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:02 crc kubenswrapper[4700]: I0227 17:23:02.633090 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87fx7\" (UniqueName: \"kubernetes.io/projected/2230bd83-cd33-438b-9d9d-001e4727fb66-kube-api-access-87fx7\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:02 crc kubenswrapper[4700]: I0227 17:23:02.633119 4700 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2230bd83-cd33-438b-9d9d-001e4727fb66-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:02 crc kubenswrapper[4700]: I0227 17:23:02.633129 4700 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2230bd83-cd33-438b-9d9d-001e4727fb66-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:02 crc kubenswrapper[4700]: I0227 17:23:02.634318 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2230bd83-cd33-438b-9d9d-001e4727fb66-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2230bd83-cd33-438b-9d9d-001e4727fb66" (UID: "2230bd83-cd33-438b-9d9d-001e4727fb66"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:02 crc kubenswrapper[4700]: I0227 17:23:02.657489 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2230bd83-cd33-438b-9d9d-001e4727fb66-config-data" (OuterVolumeSpecName: "config-data") pod "2230bd83-cd33-438b-9d9d-001e4727fb66" (UID: "2230bd83-cd33-438b-9d9d-001e4727fb66"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:02 crc kubenswrapper[4700]: W0227 17:23:02.720234 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb931c6bc_3ec1_493d_ad2e_21008ce1d1f5.slice/crio-2aeb1338713d8e6fcf3349bbc870e9b10da835c06270be994cdd2ab511354693 WatchSource:0}: Error finding container 2aeb1338713d8e6fcf3349bbc870e9b10da835c06270be994cdd2ab511354693: Status 404 returned error can't find the container with id 2aeb1338713d8e6fcf3349bbc870e9b10da835c06270be994cdd2ab511354693 Feb 27 17:23:02 crc kubenswrapper[4700]: I0227 17:23:02.726947 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-xf6ff"] Feb 27 17:23:02 crc kubenswrapper[4700]: I0227 17:23:02.734671 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2230bd83-cd33-438b-9d9d-001e4727fb66-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:02 crc kubenswrapper[4700]: I0227 17:23:02.734696 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2230bd83-cd33-438b-9d9d-001e4727fb66-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.109904 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-4603-account-create-update-4tz7h"] Feb 27 17:23:03 crc kubenswrapper[4700]: W0227 17:23:03.121198 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd646da10_fce6_4960_aedd_6aebd88f075b.slice/crio-d3dcf9566f17769d5cd80f56f3ff673c6512e91fc252efcc4a1ebe6058a7f0dc WatchSource:0}: Error finding container d3dcf9566f17769d5cd80f56f3ff673c6512e91fc252efcc4a1ebe6058a7f0dc: Status 404 returned error can't find the container with id d3dcf9566f17769d5cd80f56f3ff673c6512e91fc252efcc4a1ebe6058a7f0dc Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.131641 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-xr75f"] Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.140276 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-7178-account-create-update-ttwhv"] Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.152129 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-6728-account-create-update-swj84"] Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.164000 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-829r9"] Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.180437 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"b2cd011d-8143-4800-ae5a-35c9d348892c","Type":"ContainerStarted","Data":"2274fef6a90438c26771daec1fc6c770ec0c100647734c92e26e8c3504c2dbd0"} Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.181752 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-4603-account-create-update-4tz7h" event={"ID":"d646da10-fce6-4960-aedd-6aebd88f075b","Type":"ContainerStarted","Data":"d3dcf9566f17769d5cd80f56f3ff673c6512e91fc252efcc4a1ebe6058a7f0dc"} Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.184802 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-xf6ff" event={"ID":"b931c6bc-3ec1-493d-ad2e-21008ce1d1f5","Type":"ContainerStarted","Data":"6e0eeaaf9cba985e7619f89de8cc8165762e556f8d16817b3a9a8b3b42c601cc"} Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.184938 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-xf6ff" event={"ID":"b931c6bc-3ec1-493d-ad2e-21008ce1d1f5","Type":"ContainerStarted","Data":"2aeb1338713d8e6fcf3349bbc870e9b10da835c06270be994cdd2ab511354693"} Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.197763 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-7178-account-create-update-ttwhv" event={"ID":"04cd0229-5bc9-424e-ad73-9b3f9eb3de8d","Type":"ContainerStarted","Data":"2d413e6c2e41188a4e9db0b480e6944a1febff0b632f2d23ab71108169b0d85c"} Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.202122 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.011173498 podStartE2EDuration="11.202094372s" podCreationTimestamp="2026-02-27 17:22:52 +0000 UTC" firstStartedPulling="2026-02-27 17:22:53.06775476 +0000 UTC m=+1333.053067527" lastFinishedPulling="2026-02-27 17:23:02.258675654 +0000 UTC m=+1342.243988401" observedRunningTime="2026-02-27 17:23:03.19557663 +0000 UTC m=+1343.180889377" watchObservedRunningTime="2026-02-27 17:23:03.202094372 +0000 UTC m=+1343.187407129" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.241507 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.241555 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2230bd83-cd33-438b-9d9d-001e4727fb66","Type":"ContainerDied","Data":"69c2b00d086046db5dbe4e390eadf8c54a7f4f8e01f44f386589e10c1718169b"} Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.241598 4700 scope.go:117] "RemoveContainer" containerID="8e03101f296c609397f67f6db25ba89c6efef855a814df305c7288fddff2ce5b" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.253835 4700 generic.go:334] "Generic (PLEG): container finished" podID="d3989a3e-45ff-45d8-80a2-4dbe57663a57" containerID="2b6e983100e515e7297f4f273adfbef4735de5dcf00d127f6ac2101806ecb734" exitCode=137 Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.253900 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5b48777cdd-tqbkj" event={"ID":"d3989a3e-45ff-45d8-80a2-4dbe57663a57","Type":"ContainerDied","Data":"2b6e983100e515e7297f4f273adfbef4735de5dcf00d127f6ac2101806ecb734"} Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.254967 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-xf6ff" podStartSLOduration=7.254935306 podStartE2EDuration="7.254935306s" podCreationTimestamp="2026-02-27 17:22:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:23:03.210920025 +0000 UTC m=+1343.196232762" watchObservedRunningTime="2026-02-27 17:23:03.254935306 +0000 UTC m=+1343.240248053" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.255380 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-xr75f" event={"ID":"05664c1a-c691-4925-a4e6-6fdc12b7998a","Type":"ContainerStarted","Data":"a723a43445b624ef352056f5bb28830b3c2cfbd2f1797faf11295c1e2f9962e9"} Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.266705 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-65457d8799-dclxm"] Feb 27 17:23:03 crc kubenswrapper[4700]: W0227 17:23:03.307090 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode43c5d5e_63ef_45b9_af4e_627a035b376e.slice/crio-6eb775ef013a039dbd8cc73d746d4a93d349f52d780f1c607d748350db34d12e WatchSource:0}: Error finding container 6eb775ef013a039dbd8cc73d746d4a93d349f52d780f1c607d748350db34d12e: Status 404 returned error can't find the container with id 6eb775ef013a039dbd8cc73d746d4a93d349f52d780f1c607d748350db34d12e Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.353634 4700 scope.go:117] "RemoveContainer" containerID="3ad4aca4ec61405efe35f12a970058daaf58cd3480a4fc6f5a2a641fae53761e" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.470680 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5b48777cdd-tqbkj" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.493118 4700 scope.go:117] "RemoveContainer" containerID="beb916ec8dfdea917abac29c24892a3eb645748ade539e9616642050bb73dace" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.537521 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.547779 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.558904 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3989a3e-45ff-45d8-80a2-4dbe57663a57-combined-ca-bundle\") pod \"d3989a3e-45ff-45d8-80a2-4dbe57663a57\" (UID: \"d3989a3e-45ff-45d8-80a2-4dbe57663a57\") " Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.558941 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3989a3e-45ff-45d8-80a2-4dbe57663a57-horizon-tls-certs\") pod \"d3989a3e-45ff-45d8-80a2-4dbe57663a57\" (UID: \"d3989a3e-45ff-45d8-80a2-4dbe57663a57\") " Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.558991 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8v9ht\" (UniqueName: \"kubernetes.io/projected/d3989a3e-45ff-45d8-80a2-4dbe57663a57-kube-api-access-8v9ht\") pod \"d3989a3e-45ff-45d8-80a2-4dbe57663a57\" (UID: \"d3989a3e-45ff-45d8-80a2-4dbe57663a57\") " Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.559046 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d3989a3e-45ff-45d8-80a2-4dbe57663a57-horizon-secret-key\") pod \"d3989a3e-45ff-45d8-80a2-4dbe57663a57\" (UID: \"d3989a3e-45ff-45d8-80a2-4dbe57663a57\") " Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.559105 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3989a3e-45ff-45d8-80a2-4dbe57663a57-logs\") pod \"d3989a3e-45ff-45d8-80a2-4dbe57663a57\" (UID: \"d3989a3e-45ff-45d8-80a2-4dbe57663a57\") " Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.559209 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d3989a3e-45ff-45d8-80a2-4dbe57663a57-config-data\") pod \"d3989a3e-45ff-45d8-80a2-4dbe57663a57\" (UID: \"d3989a3e-45ff-45d8-80a2-4dbe57663a57\") " Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.559246 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d3989a3e-45ff-45d8-80a2-4dbe57663a57-scripts\") pod \"d3989a3e-45ff-45d8-80a2-4dbe57663a57\" (UID: \"d3989a3e-45ff-45d8-80a2-4dbe57663a57\") " Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.560305 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3989a3e-45ff-45d8-80a2-4dbe57663a57-logs" (OuterVolumeSpecName: "logs") pod "d3989a3e-45ff-45d8-80a2-4dbe57663a57" (UID: "d3989a3e-45ff-45d8-80a2-4dbe57663a57"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.563057 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:23:03 crc kubenswrapper[4700]: E0227 17:23:03.563576 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2230bd83-cd33-438b-9d9d-001e4727fb66" containerName="proxy-httpd" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.563676 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="2230bd83-cd33-438b-9d9d-001e4727fb66" containerName="proxy-httpd" Feb 27 17:23:03 crc kubenswrapper[4700]: E0227 17:23:03.563696 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2230bd83-cd33-438b-9d9d-001e4727fb66" containerName="ceilometer-notification-agent" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.563704 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="2230bd83-cd33-438b-9d9d-001e4727fb66" containerName="ceilometer-notification-agent" Feb 27 17:23:03 crc kubenswrapper[4700]: E0227 17:23:03.563718 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3989a3e-45ff-45d8-80a2-4dbe57663a57" containerName="horizon-log" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.563725 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3989a3e-45ff-45d8-80a2-4dbe57663a57" containerName="horizon-log" Feb 27 17:23:03 crc kubenswrapper[4700]: E0227 17:23:03.563742 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3989a3e-45ff-45d8-80a2-4dbe57663a57" containerName="horizon" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.563749 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3989a3e-45ff-45d8-80a2-4dbe57663a57" containerName="horizon" Feb 27 17:23:03 crc kubenswrapper[4700]: E0227 17:23:03.563767 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2230bd83-cd33-438b-9d9d-001e4727fb66" containerName="ceilometer-central-agent" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.563774 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="2230bd83-cd33-438b-9d9d-001e4727fb66" containerName="ceilometer-central-agent" Feb 27 17:23:03 crc kubenswrapper[4700]: E0227 17:23:03.563801 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2230bd83-cd33-438b-9d9d-001e4727fb66" containerName="sg-core" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.563808 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="2230bd83-cd33-438b-9d9d-001e4727fb66" containerName="sg-core" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.564076 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="2230bd83-cd33-438b-9d9d-001e4727fb66" containerName="ceilometer-notification-agent" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.564100 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="2230bd83-cd33-438b-9d9d-001e4727fb66" containerName="sg-core" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.564112 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3989a3e-45ff-45d8-80a2-4dbe57663a57" containerName="horizon-log" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.564130 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="2230bd83-cd33-438b-9d9d-001e4727fb66" containerName="ceilometer-central-agent" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.564144 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="2230bd83-cd33-438b-9d9d-001e4727fb66" containerName="proxy-httpd" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.564163 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3989a3e-45ff-45d8-80a2-4dbe57663a57" containerName="horizon" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.565243 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3989a3e-45ff-45d8-80a2-4dbe57663a57-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "d3989a3e-45ff-45d8-80a2-4dbe57663a57" (UID: "d3989a3e-45ff-45d8-80a2-4dbe57663a57"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.572250 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.578883 4700 scope.go:117] "RemoveContainer" containerID="939cdbfe2666a6be220f5f5345c486c74663161256917db6f4f375d293727904" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.580033 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.580153 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3989a3e-45ff-45d8-80a2-4dbe57663a57-kube-api-access-8v9ht" (OuterVolumeSpecName: "kube-api-access-8v9ht") pod "d3989a3e-45ff-45d8-80a2-4dbe57663a57" (UID: "d3989a3e-45ff-45d8-80a2-4dbe57663a57"). InnerVolumeSpecName "kube-api-access-8v9ht". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.580309 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.588073 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.653881 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3989a3e-45ff-45d8-80a2-4dbe57663a57-scripts" (OuterVolumeSpecName: "scripts") pod "d3989a3e-45ff-45d8-80a2-4dbe57663a57" (UID: "d3989a3e-45ff-45d8-80a2-4dbe57663a57"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.663181 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82104760-b734-4a3b-89c7-c6774f2ace27-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"82104760-b734-4a3b-89c7-c6774f2ace27\") " pod="openstack/ceilometer-0" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.663296 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/82104760-b734-4a3b-89c7-c6774f2ace27-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"82104760-b734-4a3b-89c7-c6774f2ace27\") " pod="openstack/ceilometer-0" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.663434 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82104760-b734-4a3b-89c7-c6774f2ace27-log-httpd\") pod \"ceilometer-0\" (UID: \"82104760-b734-4a3b-89c7-c6774f2ace27\") " pod="openstack/ceilometer-0" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.663533 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bn9d8\" (UniqueName: \"kubernetes.io/projected/82104760-b734-4a3b-89c7-c6774f2ace27-kube-api-access-bn9d8\") pod \"ceilometer-0\" (UID: \"82104760-b734-4a3b-89c7-c6774f2ace27\") " pod="openstack/ceilometer-0" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.663559 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82104760-b734-4a3b-89c7-c6774f2ace27-run-httpd\") pod \"ceilometer-0\" (UID: \"82104760-b734-4a3b-89c7-c6774f2ace27\") " pod="openstack/ceilometer-0" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.663582 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82104760-b734-4a3b-89c7-c6774f2ace27-scripts\") pod \"ceilometer-0\" (UID: \"82104760-b734-4a3b-89c7-c6774f2ace27\") " pod="openstack/ceilometer-0" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.663663 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82104760-b734-4a3b-89c7-c6774f2ace27-config-data\") pod \"ceilometer-0\" (UID: \"82104760-b734-4a3b-89c7-c6774f2ace27\") " pod="openstack/ceilometer-0" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.663710 4700 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d3989a3e-45ff-45d8-80a2-4dbe57663a57-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.663721 4700 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3989a3e-45ff-45d8-80a2-4dbe57663a57-logs\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.663730 4700 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d3989a3e-45ff-45d8-80a2-4dbe57663a57-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.663738 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8v9ht\" (UniqueName: \"kubernetes.io/projected/d3989a3e-45ff-45d8-80a2-4dbe57663a57-kube-api-access-8v9ht\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.676357 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3989a3e-45ff-45d8-80a2-4dbe57663a57-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d3989a3e-45ff-45d8-80a2-4dbe57663a57" (UID: "d3989a3e-45ff-45d8-80a2-4dbe57663a57"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.685854 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3989a3e-45ff-45d8-80a2-4dbe57663a57-config-data" (OuterVolumeSpecName: "config-data") pod "d3989a3e-45ff-45d8-80a2-4dbe57663a57" (UID: "d3989a3e-45ff-45d8-80a2-4dbe57663a57"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.766042 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bn9d8\" (UniqueName: \"kubernetes.io/projected/82104760-b734-4a3b-89c7-c6774f2ace27-kube-api-access-bn9d8\") pod \"ceilometer-0\" (UID: \"82104760-b734-4a3b-89c7-c6774f2ace27\") " pod="openstack/ceilometer-0" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.766101 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82104760-b734-4a3b-89c7-c6774f2ace27-run-httpd\") pod \"ceilometer-0\" (UID: \"82104760-b734-4a3b-89c7-c6774f2ace27\") " pod="openstack/ceilometer-0" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.766131 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82104760-b734-4a3b-89c7-c6774f2ace27-scripts\") pod \"ceilometer-0\" (UID: \"82104760-b734-4a3b-89c7-c6774f2ace27\") " pod="openstack/ceilometer-0" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.766221 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82104760-b734-4a3b-89c7-c6774f2ace27-config-data\") pod \"ceilometer-0\" (UID: \"82104760-b734-4a3b-89c7-c6774f2ace27\") " pod="openstack/ceilometer-0" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.766262 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82104760-b734-4a3b-89c7-c6774f2ace27-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"82104760-b734-4a3b-89c7-c6774f2ace27\") " pod="openstack/ceilometer-0" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.766335 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/82104760-b734-4a3b-89c7-c6774f2ace27-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"82104760-b734-4a3b-89c7-c6774f2ace27\") " pod="openstack/ceilometer-0" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.766365 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82104760-b734-4a3b-89c7-c6774f2ace27-log-httpd\") pod \"ceilometer-0\" (UID: \"82104760-b734-4a3b-89c7-c6774f2ace27\") " pod="openstack/ceilometer-0" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.766479 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d3989a3e-45ff-45d8-80a2-4dbe57663a57-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.766500 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3989a3e-45ff-45d8-80a2-4dbe57663a57-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.767009 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82104760-b734-4a3b-89c7-c6774f2ace27-log-httpd\") pod \"ceilometer-0\" (UID: \"82104760-b734-4a3b-89c7-c6774f2ace27\") " pod="openstack/ceilometer-0" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.767671 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82104760-b734-4a3b-89c7-c6774f2ace27-run-httpd\") pod \"ceilometer-0\" (UID: \"82104760-b734-4a3b-89c7-c6774f2ace27\") " pod="openstack/ceilometer-0" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.775716 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82104760-b734-4a3b-89c7-c6774f2ace27-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"82104760-b734-4a3b-89c7-c6774f2ace27\") " pod="openstack/ceilometer-0" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.776363 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82104760-b734-4a3b-89c7-c6774f2ace27-config-data\") pod \"ceilometer-0\" (UID: \"82104760-b734-4a3b-89c7-c6774f2ace27\") " pod="openstack/ceilometer-0" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.776759 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/82104760-b734-4a3b-89c7-c6774f2ace27-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"82104760-b734-4a3b-89c7-c6774f2ace27\") " pod="openstack/ceilometer-0" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.776862 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82104760-b734-4a3b-89c7-c6774f2ace27-scripts\") pod \"ceilometer-0\" (UID: \"82104760-b734-4a3b-89c7-c6774f2ace27\") " pod="openstack/ceilometer-0" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.788236 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bn9d8\" (UniqueName: \"kubernetes.io/projected/82104760-b734-4a3b-89c7-c6774f2ace27-kube-api-access-bn9d8\") pod \"ceilometer-0\" (UID: \"82104760-b734-4a3b-89c7-c6774f2ace27\") " pod="openstack/ceilometer-0" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.811741 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3989a3e-45ff-45d8-80a2-4dbe57663a57-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "d3989a3e-45ff-45d8-80a2-4dbe57663a57" (UID: "d3989a3e-45ff-45d8-80a2-4dbe57663a57"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.868051 4700 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3989a3e-45ff-45d8-80a2-4dbe57663a57-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:03 crc kubenswrapper[4700]: I0227 17:23:03.929197 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 17:23:04 crc kubenswrapper[4700]: I0227 17:23:04.020740 4700 scope.go:117] "RemoveContainer" containerID="cb8af2e94530ddd5327d653df36b6c92e24ad0db68ed6c6a9b9d9f95eb6a5e71" Feb 27 17:23:04 crc kubenswrapper[4700]: I0227 17:23:04.276323 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-829r9" event={"ID":"c22fb154-217c-462c-a0a6-4b67b5bb11d2","Type":"ContainerStarted","Data":"9e561378214aa9078b85a2a26e03e0b80ebcd6e117cee4356fbc8010ac8b4944"} Feb 27 17:23:04 crc kubenswrapper[4700]: I0227 17:23:04.276684 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-829r9" event={"ID":"c22fb154-217c-462c-a0a6-4b67b5bb11d2","Type":"ContainerStarted","Data":"fa0af98683a27f81bcb5ad6743717785e53ad3e43580e9339e8011d8959c4cf5"} Feb 27 17:23:04 crc kubenswrapper[4700]: I0227 17:23:04.284682 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-4603-account-create-update-4tz7h" event={"ID":"d646da10-fce6-4960-aedd-6aebd88f075b","Type":"ContainerStarted","Data":"428a86c77496a1ef355760ae818554a4470338024f8e15f6a0eecae75188b776"} Feb 27 17:23:04 crc kubenswrapper[4700]: I0227 17:23:04.306273 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5b48777cdd-tqbkj" event={"ID":"d3989a3e-45ff-45d8-80a2-4dbe57663a57","Type":"ContainerDied","Data":"1aa8f44a853b1bb812ffcf4b2d862218885ccb5e6483fcdaa92691b146310aab"} Feb 27 17:23:04 crc kubenswrapper[4700]: I0227 17:23:04.306316 4700 scope.go:117] "RemoveContainer" containerID="cf749444d7c76d9317d18a0299503d9ccecc0cfd7ec5ffd81bd3d244886124dc" Feb 27 17:23:04 crc kubenswrapper[4700]: I0227 17:23:04.306408 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5b48777cdd-tqbkj" Feb 27 17:23:04 crc kubenswrapper[4700]: I0227 17:23:04.311837 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-6728-account-create-update-swj84" event={"ID":"26947d30-e5b4-4a9f-9a14-c41185ff5a4f","Type":"ContainerStarted","Data":"7275f5c66342d314054d0067e1ee2df3fa76dc8fe893971d7e1ed2db1f2a95c5"} Feb 27 17:23:04 crc kubenswrapper[4700]: I0227 17:23:04.311869 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-6728-account-create-update-swj84" event={"ID":"26947d30-e5b4-4a9f-9a14-c41185ff5a4f","Type":"ContainerStarted","Data":"f126bd793958f92c91977a255ebeff5772276dae6306dbe730ed3c348e5c6a6e"} Feb 27 17:23:04 crc kubenswrapper[4700]: I0227 17:23:04.323224 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-829r9" podStartSLOduration=8.32318198 podStartE2EDuration="8.32318198s" podCreationTimestamp="2026-02-27 17:22:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:23:04.298025566 +0000 UTC m=+1344.283338313" watchObservedRunningTime="2026-02-27 17:23:04.32318198 +0000 UTC m=+1344.308494727" Feb 27 17:23:04 crc kubenswrapper[4700]: I0227 17:23:04.323328 4700 generic.go:334] "Generic (PLEG): container finished" podID="b931c6bc-3ec1-493d-ad2e-21008ce1d1f5" containerID="6e0eeaaf9cba985e7619f89de8cc8165762e556f8d16817b3a9a8b3b42c601cc" exitCode=0 Feb 27 17:23:04 crc kubenswrapper[4700]: I0227 17:23:04.323434 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-xf6ff" event={"ID":"b931c6bc-3ec1-493d-ad2e-21008ce1d1f5","Type":"ContainerDied","Data":"6e0eeaaf9cba985e7619f89de8cc8165762e556f8d16817b3a9a8b3b42c601cc"} Feb 27 17:23:04 crc kubenswrapper[4700]: I0227 17:23:04.326557 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-7178-account-create-update-ttwhv" event={"ID":"04cd0229-5bc9-424e-ad73-9b3f9eb3de8d","Type":"ContainerStarted","Data":"04297c3017efd2ccd0fcdc65482cf6fd54aa9e25a3275b3de77fd7e7e227baa9"} Feb 27 17:23:04 crc kubenswrapper[4700]: I0227 17:23:04.362936 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-4603-account-create-update-4tz7h" podStartSLOduration=8.362908648 podStartE2EDuration="8.362908648s" podCreationTimestamp="2026-02-27 17:22:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:23:04.318579718 +0000 UTC m=+1344.303892475" watchObservedRunningTime="2026-02-27 17:23:04.362908648 +0000 UTC m=+1344.348221395" Feb 27 17:23:04 crc kubenswrapper[4700]: I0227 17:23:04.372299 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-xr75f" event={"ID":"05664c1a-c691-4925-a4e6-6fdc12b7998a","Type":"ContainerStarted","Data":"f72590160ae6e9739a856399d7858272e98b055deb042e5960a406b9535a918b"} Feb 27 17:23:04 crc kubenswrapper[4700]: I0227 17:23:04.384166 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-65457d8799-dclxm" event={"ID":"e43c5d5e-63ef-45b9-af4e-627a035b376e","Type":"ContainerStarted","Data":"17473fb6b636b7f20d02f4e0becdf2dbfe648a3e50de5eb07a50e272727e25a1"} Feb 27 17:23:04 crc kubenswrapper[4700]: I0227 17:23:04.384210 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-65457d8799-dclxm" Feb 27 17:23:04 crc kubenswrapper[4700]: I0227 17:23:04.384221 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-65457d8799-dclxm" event={"ID":"e43c5d5e-63ef-45b9-af4e-627a035b376e","Type":"ContainerStarted","Data":"6eb775ef013a039dbd8cc73d746d4a93d349f52d780f1c607d748350db34d12e"} Feb 27 17:23:04 crc kubenswrapper[4700]: I0227 17:23:04.384234 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-65457d8799-dclxm" Feb 27 17:23:04 crc kubenswrapper[4700]: I0227 17:23:04.402665 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-7178-account-create-update-ttwhv" podStartSLOduration=8.402646777 podStartE2EDuration="8.402646777s" podCreationTimestamp="2026-02-27 17:22:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:23:04.369482751 +0000 UTC m=+1344.354795498" watchObservedRunningTime="2026-02-27 17:23:04.402646777 +0000 UTC m=+1344.387959524" Feb 27 17:23:04 crc kubenswrapper[4700]: I0227 17:23:04.442957 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5b48777cdd-tqbkj"] Feb 27 17:23:04 crc kubenswrapper[4700]: I0227 17:23:04.450510 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5b48777cdd-tqbkj"] Feb 27 17:23:04 crc kubenswrapper[4700]: I0227 17:23:04.452572 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-65457d8799-dclxm" podStartSLOduration=7.452555654 podStartE2EDuration="7.452555654s" podCreationTimestamp="2026-02-27 17:22:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:23:04.418049173 +0000 UTC m=+1344.403361920" watchObservedRunningTime="2026-02-27 17:23:04.452555654 +0000 UTC m=+1344.437868401" Feb 27 17:23:04 crc kubenswrapper[4700]: I0227 17:23:04.463264 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-xr75f" podStartSLOduration=8.463248996 podStartE2EDuration="8.463248996s" podCreationTimestamp="2026-02-27 17:22:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:23:04.437842916 +0000 UTC m=+1344.423155663" watchObservedRunningTime="2026-02-27 17:23:04.463248996 +0000 UTC m=+1344.448561743" Feb 27 17:23:04 crc kubenswrapper[4700]: I0227 17:23:04.471589 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:23:04 crc kubenswrapper[4700]: I0227 17:23:04.564386 4700 scope.go:117] "RemoveContainer" containerID="2b6e983100e515e7297f4f273adfbef4735de5dcf00d127f6ac2101806ecb734" Feb 27 17:23:04 crc kubenswrapper[4700]: I0227 17:23:04.993276 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2230bd83-cd33-438b-9d9d-001e4727fb66" path="/var/lib/kubelet/pods/2230bd83-cd33-438b-9d9d-001e4727fb66/volumes" Feb 27 17:23:04 crc kubenswrapper[4700]: I0227 17:23:04.994607 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3989a3e-45ff-45d8-80a2-4dbe57663a57" path="/var/lib/kubelet/pods/d3989a3e-45ff-45d8-80a2-4dbe57663a57/volumes" Feb 27 17:23:05 crc kubenswrapper[4700]: I0227 17:23:05.390566 4700 generic.go:334] "Generic (PLEG): container finished" podID="04cd0229-5bc9-424e-ad73-9b3f9eb3de8d" containerID="04297c3017efd2ccd0fcdc65482cf6fd54aa9e25a3275b3de77fd7e7e227baa9" exitCode=0 Feb 27 17:23:05 crc kubenswrapper[4700]: I0227 17:23:05.390659 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-7178-account-create-update-ttwhv" event={"ID":"04cd0229-5bc9-424e-ad73-9b3f9eb3de8d","Type":"ContainerDied","Data":"04297c3017efd2ccd0fcdc65482cf6fd54aa9e25a3275b3de77fd7e7e227baa9"} Feb 27 17:23:05 crc kubenswrapper[4700]: I0227 17:23:05.393011 4700 generic.go:334] "Generic (PLEG): container finished" podID="d646da10-fce6-4960-aedd-6aebd88f075b" containerID="428a86c77496a1ef355760ae818554a4470338024f8e15f6a0eecae75188b776" exitCode=0 Feb 27 17:23:05 crc kubenswrapper[4700]: I0227 17:23:05.393094 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-4603-account-create-update-4tz7h" event={"ID":"d646da10-fce6-4960-aedd-6aebd88f075b","Type":"ContainerDied","Data":"428a86c77496a1ef355760ae818554a4470338024f8e15f6a0eecae75188b776"} Feb 27 17:23:05 crc kubenswrapper[4700]: I0227 17:23:05.394981 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-65457d8799-dclxm" event={"ID":"e43c5d5e-63ef-45b9-af4e-627a035b376e","Type":"ContainerStarted","Data":"d46f75ab848bb1604e70d1140aab42b491208887e2d61da61ed26bed80958a90"} Feb 27 17:23:05 crc kubenswrapper[4700]: I0227 17:23:05.396853 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82104760-b734-4a3b-89c7-c6774f2ace27","Type":"ContainerStarted","Data":"f98afe6cc5de5a8cc7c6c4b99e4d6445bf8fc3572b8e74beeffc813c2c415e76"} Feb 27 17:23:05 crc kubenswrapper[4700]: I0227 17:23:05.396883 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82104760-b734-4a3b-89c7-c6774f2ace27","Type":"ContainerStarted","Data":"2fa25eca5355fa202d4ac28fe3ddfa48510e0c2e220d76a114db47de5a2e77c5"} Feb 27 17:23:05 crc kubenswrapper[4700]: I0227 17:23:05.396901 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82104760-b734-4a3b-89c7-c6774f2ace27","Type":"ContainerStarted","Data":"c3f3c402d90fb56137e21ddac03eb3d3df93ebb104f63e6c1b858a3153a046ac"} Feb 27 17:23:05 crc kubenswrapper[4700]: I0227 17:23:05.398284 4700 generic.go:334] "Generic (PLEG): container finished" podID="26947d30-e5b4-4a9f-9a14-c41185ff5a4f" containerID="7275f5c66342d314054d0067e1ee2df3fa76dc8fe893971d7e1ed2db1f2a95c5" exitCode=0 Feb 27 17:23:05 crc kubenswrapper[4700]: I0227 17:23:05.398367 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-6728-account-create-update-swj84" event={"ID":"26947d30-e5b4-4a9f-9a14-c41185ff5a4f","Type":"ContainerDied","Data":"7275f5c66342d314054d0067e1ee2df3fa76dc8fe893971d7e1ed2db1f2a95c5"} Feb 27 17:23:05 crc kubenswrapper[4700]: I0227 17:23:05.401245 4700 generic.go:334] "Generic (PLEG): container finished" podID="05664c1a-c691-4925-a4e6-6fdc12b7998a" containerID="f72590160ae6e9739a856399d7858272e98b055deb042e5960a406b9535a918b" exitCode=0 Feb 27 17:23:05 crc kubenswrapper[4700]: I0227 17:23:05.401312 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-xr75f" event={"ID":"05664c1a-c691-4925-a4e6-6fdc12b7998a","Type":"ContainerDied","Data":"f72590160ae6e9739a856399d7858272e98b055deb042e5960a406b9535a918b"} Feb 27 17:23:05 crc kubenswrapper[4700]: I0227 17:23:05.412583 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-829r9" event={"ID":"c22fb154-217c-462c-a0a6-4b67b5bb11d2","Type":"ContainerDied","Data":"9e561378214aa9078b85a2a26e03e0b80ebcd6e117cee4356fbc8010ac8b4944"} Feb 27 17:23:05 crc kubenswrapper[4700]: I0227 17:23:05.412438 4700 generic.go:334] "Generic (PLEG): container finished" podID="c22fb154-217c-462c-a0a6-4b67b5bb11d2" containerID="9e561378214aa9078b85a2a26e03e0b80ebcd6e117cee4356fbc8010ac8b4944" exitCode=0 Feb 27 17:23:05 crc kubenswrapper[4700]: I0227 17:23:05.911620 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-6728-account-create-update-swj84" Feb 27 17:23:05 crc kubenswrapper[4700]: I0227 17:23:05.920770 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-xf6ff" Feb 27 17:23:06 crc kubenswrapper[4700]: I0227 17:23:06.014703 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/26947d30-e5b4-4a9f-9a14-c41185ff5a4f-operator-scripts\") pod \"26947d30-e5b4-4a9f-9a14-c41185ff5a4f\" (UID: \"26947d30-e5b4-4a9f-9a14-c41185ff5a4f\") " Feb 27 17:23:06 crc kubenswrapper[4700]: I0227 17:23:06.014849 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgcv8\" (UniqueName: \"kubernetes.io/projected/b931c6bc-3ec1-493d-ad2e-21008ce1d1f5-kube-api-access-zgcv8\") pod \"b931c6bc-3ec1-493d-ad2e-21008ce1d1f5\" (UID: \"b931c6bc-3ec1-493d-ad2e-21008ce1d1f5\") " Feb 27 17:23:06 crc kubenswrapper[4700]: I0227 17:23:06.014958 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kb972\" (UniqueName: \"kubernetes.io/projected/26947d30-e5b4-4a9f-9a14-c41185ff5a4f-kube-api-access-kb972\") pod \"26947d30-e5b4-4a9f-9a14-c41185ff5a4f\" (UID: \"26947d30-e5b4-4a9f-9a14-c41185ff5a4f\") " Feb 27 17:23:06 crc kubenswrapper[4700]: I0227 17:23:06.014979 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b931c6bc-3ec1-493d-ad2e-21008ce1d1f5-operator-scripts\") pod \"b931c6bc-3ec1-493d-ad2e-21008ce1d1f5\" (UID: \"b931c6bc-3ec1-493d-ad2e-21008ce1d1f5\") " Feb 27 17:23:06 crc kubenswrapper[4700]: I0227 17:23:06.016747 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26947d30-e5b4-4a9f-9a14-c41185ff5a4f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "26947d30-e5b4-4a9f-9a14-c41185ff5a4f" (UID: "26947d30-e5b4-4a9f-9a14-c41185ff5a4f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:23:06 crc kubenswrapper[4700]: I0227 17:23:06.017886 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b931c6bc-3ec1-493d-ad2e-21008ce1d1f5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b931c6bc-3ec1-493d-ad2e-21008ce1d1f5" (UID: "b931c6bc-3ec1-493d-ad2e-21008ce1d1f5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:23:06 crc kubenswrapper[4700]: I0227 17:23:06.022166 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26947d30-e5b4-4a9f-9a14-c41185ff5a4f-kube-api-access-kb972" (OuterVolumeSpecName: "kube-api-access-kb972") pod "26947d30-e5b4-4a9f-9a14-c41185ff5a4f" (UID: "26947d30-e5b4-4a9f-9a14-c41185ff5a4f"). InnerVolumeSpecName "kube-api-access-kb972". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:23:06 crc kubenswrapper[4700]: I0227 17:23:06.023659 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b931c6bc-3ec1-493d-ad2e-21008ce1d1f5-kube-api-access-zgcv8" (OuterVolumeSpecName: "kube-api-access-zgcv8") pod "b931c6bc-3ec1-493d-ad2e-21008ce1d1f5" (UID: "b931c6bc-3ec1-493d-ad2e-21008ce1d1f5"). InnerVolumeSpecName "kube-api-access-zgcv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:23:06 crc kubenswrapper[4700]: I0227 17:23:06.117756 4700 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/26947d30-e5b4-4a9f-9a14-c41185ff5a4f-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:06 crc kubenswrapper[4700]: I0227 17:23:06.117798 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgcv8\" (UniqueName: \"kubernetes.io/projected/b931c6bc-3ec1-493d-ad2e-21008ce1d1f5-kube-api-access-zgcv8\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:06 crc kubenswrapper[4700]: I0227 17:23:06.117816 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kb972\" (UniqueName: \"kubernetes.io/projected/26947d30-e5b4-4a9f-9a14-c41185ff5a4f-kube-api-access-kb972\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:06 crc kubenswrapper[4700]: I0227 17:23:06.117832 4700 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b931c6bc-3ec1-493d-ad2e-21008ce1d1f5-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:06 crc kubenswrapper[4700]: I0227 17:23:06.423083 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82104760-b734-4a3b-89c7-c6774f2ace27","Type":"ContainerStarted","Data":"3c5f4b118989a58fc0d031ddfb3e2d6e548520a4718c3f9320a3df03d5564b40"} Feb 27 17:23:06 crc kubenswrapper[4700]: I0227 17:23:06.424399 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-6728-account-create-update-swj84" event={"ID":"26947d30-e5b4-4a9f-9a14-c41185ff5a4f","Type":"ContainerDied","Data":"f126bd793958f92c91977a255ebeff5772276dae6306dbe730ed3c348e5c6a6e"} Feb 27 17:23:06 crc kubenswrapper[4700]: I0227 17:23:06.424432 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-6728-account-create-update-swj84" Feb 27 17:23:06 crc kubenswrapper[4700]: I0227 17:23:06.424431 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f126bd793958f92c91977a255ebeff5772276dae6306dbe730ed3c348e5c6a6e" Feb 27 17:23:06 crc kubenswrapper[4700]: I0227 17:23:06.425808 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-xf6ff" event={"ID":"b931c6bc-3ec1-493d-ad2e-21008ce1d1f5","Type":"ContainerDied","Data":"2aeb1338713d8e6fcf3349bbc870e9b10da835c06270be994cdd2ab511354693"} Feb 27 17:23:06 crc kubenswrapper[4700]: I0227 17:23:06.425861 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2aeb1338713d8e6fcf3349bbc870e9b10da835c06270be994cdd2ab511354693" Feb 27 17:23:06 crc kubenswrapper[4700]: I0227 17:23:06.425870 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-xf6ff" Feb 27 17:23:06 crc kubenswrapper[4700]: I0227 17:23:06.869807 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xr75f" Feb 27 17:23:06 crc kubenswrapper[4700]: I0227 17:23:06.939588 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05664c1a-c691-4925-a4e6-6fdc12b7998a-operator-scripts\") pod \"05664c1a-c691-4925-a4e6-6fdc12b7998a\" (UID: \"05664c1a-c691-4925-a4e6-6fdc12b7998a\") " Feb 27 17:23:06 crc kubenswrapper[4700]: I0227 17:23:06.939715 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnkft\" (UniqueName: \"kubernetes.io/projected/05664c1a-c691-4925-a4e6-6fdc12b7998a-kube-api-access-rnkft\") pod \"05664c1a-c691-4925-a4e6-6fdc12b7998a\" (UID: \"05664c1a-c691-4925-a4e6-6fdc12b7998a\") " Feb 27 17:23:06 crc kubenswrapper[4700]: I0227 17:23:06.941390 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05664c1a-c691-4925-a4e6-6fdc12b7998a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "05664c1a-c691-4925-a4e6-6fdc12b7998a" (UID: "05664c1a-c691-4925-a4e6-6fdc12b7998a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:23:06 crc kubenswrapper[4700]: I0227 17:23:06.947284 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05664c1a-c691-4925-a4e6-6fdc12b7998a-kube-api-access-rnkft" (OuterVolumeSpecName: "kube-api-access-rnkft") pod "05664c1a-c691-4925-a4e6-6fdc12b7998a" (UID: "05664c1a-c691-4925-a4e6-6fdc12b7998a"). InnerVolumeSpecName "kube-api-access-rnkft". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:23:07 crc kubenswrapper[4700]: I0227 17:23:07.043423 4700 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05664c1a-c691-4925-a4e6-6fdc12b7998a-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:07 crc kubenswrapper[4700]: I0227 17:23:07.043589 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnkft\" (UniqueName: \"kubernetes.io/projected/05664c1a-c691-4925-a4e6-6fdc12b7998a-kube-api-access-rnkft\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:07 crc kubenswrapper[4700]: I0227 17:23:07.117754 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:23:07 crc kubenswrapper[4700]: I0227 17:23:07.202533 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-7178-account-create-update-ttwhv" Feb 27 17:23:07 crc kubenswrapper[4700]: I0227 17:23:07.207899 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-829r9" Feb 27 17:23:07 crc kubenswrapper[4700]: I0227 17:23:07.215601 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4603-account-create-update-4tz7h" Feb 27 17:23:07 crc kubenswrapper[4700]: I0227 17:23:07.355374 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsdg4\" (UniqueName: \"kubernetes.io/projected/d646da10-fce6-4960-aedd-6aebd88f075b-kube-api-access-gsdg4\") pod \"d646da10-fce6-4960-aedd-6aebd88f075b\" (UID: \"d646da10-fce6-4960-aedd-6aebd88f075b\") " Feb 27 17:23:07 crc kubenswrapper[4700]: I0227 17:23:07.355451 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04cd0229-5bc9-424e-ad73-9b3f9eb3de8d-operator-scripts\") pod \"04cd0229-5bc9-424e-ad73-9b3f9eb3de8d\" (UID: \"04cd0229-5bc9-424e-ad73-9b3f9eb3de8d\") " Feb 27 17:23:07 crc kubenswrapper[4700]: I0227 17:23:07.355574 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c22fb154-217c-462c-a0a6-4b67b5bb11d2-operator-scripts\") pod \"c22fb154-217c-462c-a0a6-4b67b5bb11d2\" (UID: \"c22fb154-217c-462c-a0a6-4b67b5bb11d2\") " Feb 27 17:23:07 crc kubenswrapper[4700]: I0227 17:23:07.355652 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n96xj\" (UniqueName: \"kubernetes.io/projected/c22fb154-217c-462c-a0a6-4b67b5bb11d2-kube-api-access-n96xj\") pod \"c22fb154-217c-462c-a0a6-4b67b5bb11d2\" (UID: \"c22fb154-217c-462c-a0a6-4b67b5bb11d2\") " Feb 27 17:23:07 crc kubenswrapper[4700]: I0227 17:23:07.355700 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d646da10-fce6-4960-aedd-6aebd88f075b-operator-scripts\") pod \"d646da10-fce6-4960-aedd-6aebd88f075b\" (UID: \"d646da10-fce6-4960-aedd-6aebd88f075b\") " Feb 27 17:23:07 crc kubenswrapper[4700]: I0227 17:23:07.355718 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhfxj\" (UniqueName: \"kubernetes.io/projected/04cd0229-5bc9-424e-ad73-9b3f9eb3de8d-kube-api-access-zhfxj\") pod \"04cd0229-5bc9-424e-ad73-9b3f9eb3de8d\" (UID: \"04cd0229-5bc9-424e-ad73-9b3f9eb3de8d\") " Feb 27 17:23:07 crc kubenswrapper[4700]: I0227 17:23:07.355978 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04cd0229-5bc9-424e-ad73-9b3f9eb3de8d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "04cd0229-5bc9-424e-ad73-9b3f9eb3de8d" (UID: "04cd0229-5bc9-424e-ad73-9b3f9eb3de8d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:23:07 crc kubenswrapper[4700]: I0227 17:23:07.356324 4700 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04cd0229-5bc9-424e-ad73-9b3f9eb3de8d-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:07 crc kubenswrapper[4700]: I0227 17:23:07.357089 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c22fb154-217c-462c-a0a6-4b67b5bb11d2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c22fb154-217c-462c-a0a6-4b67b5bb11d2" (UID: "c22fb154-217c-462c-a0a6-4b67b5bb11d2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:23:07 crc kubenswrapper[4700]: I0227 17:23:07.357447 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d646da10-fce6-4960-aedd-6aebd88f075b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d646da10-fce6-4960-aedd-6aebd88f075b" (UID: "d646da10-fce6-4960-aedd-6aebd88f075b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:23:07 crc kubenswrapper[4700]: I0227 17:23:07.361727 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c22fb154-217c-462c-a0a6-4b67b5bb11d2-kube-api-access-n96xj" (OuterVolumeSpecName: "kube-api-access-n96xj") pod "c22fb154-217c-462c-a0a6-4b67b5bb11d2" (UID: "c22fb154-217c-462c-a0a6-4b67b5bb11d2"). InnerVolumeSpecName "kube-api-access-n96xj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:23:07 crc kubenswrapper[4700]: I0227 17:23:07.362632 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d646da10-fce6-4960-aedd-6aebd88f075b-kube-api-access-gsdg4" (OuterVolumeSpecName: "kube-api-access-gsdg4") pod "d646da10-fce6-4960-aedd-6aebd88f075b" (UID: "d646da10-fce6-4960-aedd-6aebd88f075b"). InnerVolumeSpecName "kube-api-access-gsdg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:23:07 crc kubenswrapper[4700]: I0227 17:23:07.365637 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04cd0229-5bc9-424e-ad73-9b3f9eb3de8d-kube-api-access-zhfxj" (OuterVolumeSpecName: "kube-api-access-zhfxj") pod "04cd0229-5bc9-424e-ad73-9b3f9eb3de8d" (UID: "04cd0229-5bc9-424e-ad73-9b3f9eb3de8d"). InnerVolumeSpecName "kube-api-access-zhfxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:23:07 crc kubenswrapper[4700]: I0227 17:23:07.434973 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-829r9" event={"ID":"c22fb154-217c-462c-a0a6-4b67b5bb11d2","Type":"ContainerDied","Data":"fa0af98683a27f81bcb5ad6743717785e53ad3e43580e9339e8011d8959c4cf5"} Feb 27 17:23:07 crc kubenswrapper[4700]: I0227 17:23:07.435284 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-829r9" Feb 27 17:23:07 crc kubenswrapper[4700]: I0227 17:23:07.435294 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa0af98683a27f81bcb5ad6743717785e53ad3e43580e9339e8011d8959c4cf5" Feb 27 17:23:07 crc kubenswrapper[4700]: I0227 17:23:07.436590 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-4603-account-create-update-4tz7h" event={"ID":"d646da10-fce6-4960-aedd-6aebd88f075b","Type":"ContainerDied","Data":"d3dcf9566f17769d5cd80f56f3ff673c6512e91fc252efcc4a1ebe6058a7f0dc"} Feb 27 17:23:07 crc kubenswrapper[4700]: I0227 17:23:07.436633 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3dcf9566f17769d5cd80f56f3ff673c6512e91fc252efcc4a1ebe6058a7f0dc" Feb 27 17:23:07 crc kubenswrapper[4700]: I0227 17:23:07.436630 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4603-account-create-update-4tz7h" Feb 27 17:23:07 crc kubenswrapper[4700]: I0227 17:23:07.438088 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-xr75f" event={"ID":"05664c1a-c691-4925-a4e6-6fdc12b7998a","Type":"ContainerDied","Data":"a723a43445b624ef352056f5bb28830b3c2cfbd2f1797faf11295c1e2f9962e9"} Feb 27 17:23:07 crc kubenswrapper[4700]: I0227 17:23:07.438111 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a723a43445b624ef352056f5bb28830b3c2cfbd2f1797faf11295c1e2f9962e9" Feb 27 17:23:07 crc kubenswrapper[4700]: I0227 17:23:07.438197 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xr75f" Feb 27 17:23:07 crc kubenswrapper[4700]: I0227 17:23:07.439953 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-7178-account-create-update-ttwhv" event={"ID":"04cd0229-5bc9-424e-ad73-9b3f9eb3de8d","Type":"ContainerDied","Data":"2d413e6c2e41188a4e9db0b480e6944a1febff0b632f2d23ab71108169b0d85c"} Feb 27 17:23:07 crc kubenswrapper[4700]: I0227 17:23:07.439991 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d413e6c2e41188a4e9db0b480e6944a1febff0b632f2d23ab71108169b0d85c" Feb 27 17:23:07 crc kubenswrapper[4700]: I0227 17:23:07.440040 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-7178-account-create-update-ttwhv" Feb 27 17:23:07 crc kubenswrapper[4700]: I0227 17:23:07.460289 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n96xj\" (UniqueName: \"kubernetes.io/projected/c22fb154-217c-462c-a0a6-4b67b5bb11d2-kube-api-access-n96xj\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:07 crc kubenswrapper[4700]: I0227 17:23:07.460352 4700 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d646da10-fce6-4960-aedd-6aebd88f075b-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:07 crc kubenswrapper[4700]: I0227 17:23:07.460368 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhfxj\" (UniqueName: \"kubernetes.io/projected/04cd0229-5bc9-424e-ad73-9b3f9eb3de8d-kube-api-access-zhfxj\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:07 crc kubenswrapper[4700]: I0227 17:23:07.460380 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsdg4\" (UniqueName: \"kubernetes.io/projected/d646da10-fce6-4960-aedd-6aebd88f075b-kube-api-access-gsdg4\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:07 crc kubenswrapper[4700]: I0227 17:23:07.460391 4700 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c22fb154-217c-462c-a0a6-4b67b5bb11d2-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:08 crc kubenswrapper[4700]: I0227 17:23:08.452832 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82104760-b734-4a3b-89c7-c6774f2ace27","Type":"ContainerStarted","Data":"9955ee7b957831bf50a5646e6ef0a6dd7561229e7e807d415c2c1aa5735544de"} Feb 27 17:23:08 crc kubenswrapper[4700]: I0227 17:23:08.453007 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="82104760-b734-4a3b-89c7-c6774f2ace27" containerName="ceilometer-central-agent" containerID="cri-o://2fa25eca5355fa202d4ac28fe3ddfa48510e0c2e220d76a114db47de5a2e77c5" gracePeriod=30 Feb 27 17:23:08 crc kubenswrapper[4700]: I0227 17:23:08.453298 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 27 17:23:08 crc kubenswrapper[4700]: I0227 17:23:08.453602 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="82104760-b734-4a3b-89c7-c6774f2ace27" containerName="proxy-httpd" containerID="cri-o://9955ee7b957831bf50a5646e6ef0a6dd7561229e7e807d415c2c1aa5735544de" gracePeriod=30 Feb 27 17:23:08 crc kubenswrapper[4700]: I0227 17:23:08.453659 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="82104760-b734-4a3b-89c7-c6774f2ace27" containerName="sg-core" containerID="cri-o://3c5f4b118989a58fc0d031ddfb3e2d6e548520a4718c3f9320a3df03d5564b40" gracePeriod=30 Feb 27 17:23:08 crc kubenswrapper[4700]: I0227 17:23:08.453701 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="82104760-b734-4a3b-89c7-c6774f2ace27" containerName="ceilometer-notification-agent" containerID="cri-o://f98afe6cc5de5a8cc7c6c4b99e4d6445bf8fc3572b8e74beeffc813c2c415e76" gracePeriod=30 Feb 27 17:23:08 crc kubenswrapper[4700]: I0227 17:23:08.489481 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.039580746 podStartE2EDuration="5.489450224s" podCreationTimestamp="2026-02-27 17:23:03 +0000 UTC" firstStartedPulling="2026-02-27 17:23:04.586133949 +0000 UTC m=+1344.571446696" lastFinishedPulling="2026-02-27 17:23:08.036003427 +0000 UTC m=+1348.021316174" observedRunningTime="2026-02-27 17:23:08.481159695 +0000 UTC m=+1348.466472502" watchObservedRunningTime="2026-02-27 17:23:08.489450224 +0000 UTC m=+1348.474762971" Feb 27 17:23:09 crc kubenswrapper[4700]: I0227 17:23:09.465326 4700 generic.go:334] "Generic (PLEG): container finished" podID="82104760-b734-4a3b-89c7-c6774f2ace27" containerID="3c5f4b118989a58fc0d031ddfb3e2d6e548520a4718c3f9320a3df03d5564b40" exitCode=2 Feb 27 17:23:09 crc kubenswrapper[4700]: I0227 17:23:09.465753 4700 generic.go:334] "Generic (PLEG): container finished" podID="82104760-b734-4a3b-89c7-c6774f2ace27" containerID="f98afe6cc5de5a8cc7c6c4b99e4d6445bf8fc3572b8e74beeffc813c2c415e76" exitCode=0 Feb 27 17:23:09 crc kubenswrapper[4700]: I0227 17:23:09.465376 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82104760-b734-4a3b-89c7-c6774f2ace27","Type":"ContainerDied","Data":"3c5f4b118989a58fc0d031ddfb3e2d6e548520a4718c3f9320a3df03d5564b40"} Feb 27 17:23:09 crc kubenswrapper[4700]: I0227 17:23:09.465845 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82104760-b734-4a3b-89c7-c6774f2ace27","Type":"ContainerDied","Data":"f98afe6cc5de5a8cc7c6c4b99e4d6445bf8fc3572b8e74beeffc813c2c415e76"} Feb 27 17:23:10 crc kubenswrapper[4700]: I0227 17:23:10.824599 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-dc4cd4577-rvnzh" Feb 27 17:23:10 crc kubenswrapper[4700]: I0227 17:23:10.977934 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7db57cbc8-t9p8l"] Feb 27 17:23:10 crc kubenswrapper[4700]: I0227 17:23:10.978147 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7db57cbc8-t9p8l" podUID="24cf75b6-81b2-43e8-b75e-c41171313a0f" containerName="neutron-api" containerID="cri-o://3675602a5b830e0a8101ab8de56cf97a66e7e05593be9748eed8b289659a40a0" gracePeriod=30 Feb 27 17:23:10 crc kubenswrapper[4700]: I0227 17:23:10.978564 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7db57cbc8-t9p8l" podUID="24cf75b6-81b2-43e8-b75e-c41171313a0f" containerName="neutron-httpd" containerID="cri-o://994e8654f68b07a71d26635805e774bb44d66badb496942fec0b2817d51b9f27" gracePeriod=30 Feb 27 17:23:10 crc kubenswrapper[4700]: I0227 17:23:10.989045 4700 scope.go:117] "RemoveContainer" containerID="9b51d27b3971cd3f655c413b4bd270b5cc55c1c7f20422db504613285c3a5770" Feb 27 17:23:11 crc kubenswrapper[4700]: I0227 17:23:11.489778 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"0f1ae433-28e4-45cf-9635-dc6edbf16f6e","Type":"ContainerStarted","Data":"2d0da9f450ceed9ed3cba22dade279b2f7018efcb70068e35e804ea9219036d5"} Feb 27 17:23:11 crc kubenswrapper[4700]: I0227 17:23:11.492162 4700 generic.go:334] "Generic (PLEG): container finished" podID="24cf75b6-81b2-43e8-b75e-c41171313a0f" containerID="994e8654f68b07a71d26635805e774bb44d66badb496942fec0b2817d51b9f27" exitCode=0 Feb 27 17:23:11 crc kubenswrapper[4700]: I0227 17:23:11.492242 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7db57cbc8-t9p8l" event={"ID":"24cf75b6-81b2-43e8-b75e-c41171313a0f","Type":"ContainerDied","Data":"994e8654f68b07a71d26635805e774bb44d66badb496942fec0b2817d51b9f27"} Feb 27 17:23:11 crc kubenswrapper[4700]: I0227 17:23:11.500311 4700 generic.go:334] "Generic (PLEG): container finished" podID="82104760-b734-4a3b-89c7-c6774f2ace27" containerID="2fa25eca5355fa202d4ac28fe3ddfa48510e0c2e220d76a114db47de5a2e77c5" exitCode=0 Feb 27 17:23:11 crc kubenswrapper[4700]: I0227 17:23:11.500341 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82104760-b734-4a3b-89c7-c6774f2ace27","Type":"ContainerDied","Data":"2fa25eca5355fa202d4ac28fe3ddfa48510e0c2e220d76a114db47de5a2e77c5"} Feb 27 17:23:12 crc kubenswrapper[4700]: I0227 17:23:12.033921 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-lf45g"] Feb 27 17:23:12 crc kubenswrapper[4700]: E0227 17:23:12.034283 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b931c6bc-3ec1-493d-ad2e-21008ce1d1f5" containerName="mariadb-database-create" Feb 27 17:23:12 crc kubenswrapper[4700]: I0227 17:23:12.034296 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="b931c6bc-3ec1-493d-ad2e-21008ce1d1f5" containerName="mariadb-database-create" Feb 27 17:23:12 crc kubenswrapper[4700]: E0227 17:23:12.034309 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26947d30-e5b4-4a9f-9a14-c41185ff5a4f" containerName="mariadb-account-create-update" Feb 27 17:23:12 crc kubenswrapper[4700]: I0227 17:23:12.034316 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="26947d30-e5b4-4a9f-9a14-c41185ff5a4f" containerName="mariadb-account-create-update" Feb 27 17:23:12 crc kubenswrapper[4700]: E0227 17:23:12.034330 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05664c1a-c691-4925-a4e6-6fdc12b7998a" containerName="mariadb-database-create" Feb 27 17:23:12 crc kubenswrapper[4700]: I0227 17:23:12.034335 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="05664c1a-c691-4925-a4e6-6fdc12b7998a" containerName="mariadb-database-create" Feb 27 17:23:12 crc kubenswrapper[4700]: E0227 17:23:12.034350 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c22fb154-217c-462c-a0a6-4b67b5bb11d2" containerName="mariadb-database-create" Feb 27 17:23:12 crc kubenswrapper[4700]: I0227 17:23:12.034356 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="c22fb154-217c-462c-a0a6-4b67b5bb11d2" containerName="mariadb-database-create" Feb 27 17:23:12 crc kubenswrapper[4700]: E0227 17:23:12.034366 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04cd0229-5bc9-424e-ad73-9b3f9eb3de8d" containerName="mariadb-account-create-update" Feb 27 17:23:12 crc kubenswrapper[4700]: I0227 17:23:12.034372 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="04cd0229-5bc9-424e-ad73-9b3f9eb3de8d" containerName="mariadb-account-create-update" Feb 27 17:23:12 crc kubenswrapper[4700]: E0227 17:23:12.034387 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d646da10-fce6-4960-aedd-6aebd88f075b" containerName="mariadb-account-create-update" Feb 27 17:23:12 crc kubenswrapper[4700]: I0227 17:23:12.034392 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="d646da10-fce6-4960-aedd-6aebd88f075b" containerName="mariadb-account-create-update" Feb 27 17:23:12 crc kubenswrapper[4700]: I0227 17:23:12.034566 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="05664c1a-c691-4925-a4e6-6fdc12b7998a" containerName="mariadb-database-create" Feb 27 17:23:12 crc kubenswrapper[4700]: I0227 17:23:12.034581 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="d646da10-fce6-4960-aedd-6aebd88f075b" containerName="mariadb-account-create-update" Feb 27 17:23:12 crc kubenswrapper[4700]: I0227 17:23:12.034595 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="26947d30-e5b4-4a9f-9a14-c41185ff5a4f" containerName="mariadb-account-create-update" Feb 27 17:23:12 crc kubenswrapper[4700]: I0227 17:23:12.034604 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="04cd0229-5bc9-424e-ad73-9b3f9eb3de8d" containerName="mariadb-account-create-update" Feb 27 17:23:12 crc kubenswrapper[4700]: I0227 17:23:12.034614 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="b931c6bc-3ec1-493d-ad2e-21008ce1d1f5" containerName="mariadb-database-create" Feb 27 17:23:12 crc kubenswrapper[4700]: I0227 17:23:12.034627 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="c22fb154-217c-462c-a0a6-4b67b5bb11d2" containerName="mariadb-database-create" Feb 27 17:23:12 crc kubenswrapper[4700]: I0227 17:23:12.035226 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-lf45g" Feb 27 17:23:12 crc kubenswrapper[4700]: I0227 17:23:12.036891 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-ctcbq" Feb 27 17:23:12 crc kubenswrapper[4700]: I0227 17:23:12.037367 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Feb 27 17:23:12 crc kubenswrapper[4700]: I0227 17:23:12.037869 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 27 17:23:12 crc kubenswrapper[4700]: I0227 17:23:12.048977 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-lf45g"] Feb 27 17:23:12 crc kubenswrapper[4700]: I0227 17:23:12.193856 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cb8c359-c338-4350-acb2-b302786f082a-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-lf45g\" (UID: \"3cb8c359-c338-4350-acb2-b302786f082a\") " pod="openstack/nova-cell0-conductor-db-sync-lf45g" Feb 27 17:23:12 crc kubenswrapper[4700]: I0227 17:23:12.194211 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zp8v4\" (UniqueName: \"kubernetes.io/projected/3cb8c359-c338-4350-acb2-b302786f082a-kube-api-access-zp8v4\") pod \"nova-cell0-conductor-db-sync-lf45g\" (UID: \"3cb8c359-c338-4350-acb2-b302786f082a\") " pod="openstack/nova-cell0-conductor-db-sync-lf45g" Feb 27 17:23:12 crc kubenswrapper[4700]: I0227 17:23:12.194236 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cb8c359-c338-4350-acb2-b302786f082a-config-data\") pod \"nova-cell0-conductor-db-sync-lf45g\" (UID: \"3cb8c359-c338-4350-acb2-b302786f082a\") " pod="openstack/nova-cell0-conductor-db-sync-lf45g" Feb 27 17:23:12 crc kubenswrapper[4700]: I0227 17:23:12.194287 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cb8c359-c338-4350-acb2-b302786f082a-scripts\") pod \"nova-cell0-conductor-db-sync-lf45g\" (UID: \"3cb8c359-c338-4350-acb2-b302786f082a\") " pod="openstack/nova-cell0-conductor-db-sync-lf45g" Feb 27 17:23:12 crc kubenswrapper[4700]: I0227 17:23:12.296667 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cb8c359-c338-4350-acb2-b302786f082a-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-lf45g\" (UID: \"3cb8c359-c338-4350-acb2-b302786f082a\") " pod="openstack/nova-cell0-conductor-db-sync-lf45g" Feb 27 17:23:12 crc kubenswrapper[4700]: I0227 17:23:12.296726 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zp8v4\" (UniqueName: \"kubernetes.io/projected/3cb8c359-c338-4350-acb2-b302786f082a-kube-api-access-zp8v4\") pod \"nova-cell0-conductor-db-sync-lf45g\" (UID: \"3cb8c359-c338-4350-acb2-b302786f082a\") " pod="openstack/nova-cell0-conductor-db-sync-lf45g" Feb 27 17:23:12 crc kubenswrapper[4700]: I0227 17:23:12.296761 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cb8c359-c338-4350-acb2-b302786f082a-config-data\") pod \"nova-cell0-conductor-db-sync-lf45g\" (UID: \"3cb8c359-c338-4350-acb2-b302786f082a\") " pod="openstack/nova-cell0-conductor-db-sync-lf45g" Feb 27 17:23:12 crc kubenswrapper[4700]: I0227 17:23:12.296840 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cb8c359-c338-4350-acb2-b302786f082a-scripts\") pod \"nova-cell0-conductor-db-sync-lf45g\" (UID: \"3cb8c359-c338-4350-acb2-b302786f082a\") " pod="openstack/nova-cell0-conductor-db-sync-lf45g" Feb 27 17:23:12 crc kubenswrapper[4700]: I0227 17:23:12.303577 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cb8c359-c338-4350-acb2-b302786f082a-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-lf45g\" (UID: \"3cb8c359-c338-4350-acb2-b302786f082a\") " pod="openstack/nova-cell0-conductor-db-sync-lf45g" Feb 27 17:23:12 crc kubenswrapper[4700]: I0227 17:23:12.303800 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cb8c359-c338-4350-acb2-b302786f082a-scripts\") pod \"nova-cell0-conductor-db-sync-lf45g\" (UID: \"3cb8c359-c338-4350-acb2-b302786f082a\") " pod="openstack/nova-cell0-conductor-db-sync-lf45g" Feb 27 17:23:12 crc kubenswrapper[4700]: I0227 17:23:12.312336 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cb8c359-c338-4350-acb2-b302786f082a-config-data\") pod \"nova-cell0-conductor-db-sync-lf45g\" (UID: \"3cb8c359-c338-4350-acb2-b302786f082a\") " pod="openstack/nova-cell0-conductor-db-sync-lf45g" Feb 27 17:23:12 crc kubenswrapper[4700]: I0227 17:23:12.317111 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zp8v4\" (UniqueName: \"kubernetes.io/projected/3cb8c359-c338-4350-acb2-b302786f082a-kube-api-access-zp8v4\") pod \"nova-cell0-conductor-db-sync-lf45g\" (UID: \"3cb8c359-c338-4350-acb2-b302786f082a\") " pod="openstack/nova-cell0-conductor-db-sync-lf45g" Feb 27 17:23:12 crc kubenswrapper[4700]: I0227 17:23:12.399829 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-lf45g" Feb 27 17:23:12 crc kubenswrapper[4700]: I0227 17:23:12.876181 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-lf45g"] Feb 27 17:23:12 crc kubenswrapper[4700]: W0227 17:23:12.880107 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3cb8c359_c338_4350_acb2_b302786f082a.slice/crio-b892a4e5b88cec8f104a9fdf9c9a43f0327efb4ebbe73c0d70e9743328a44dd5 WatchSource:0}: Error finding container b892a4e5b88cec8f104a9fdf9c9a43f0327efb4ebbe73c0d70e9743328a44dd5: Status 404 returned error can't find the container with id b892a4e5b88cec8f104a9fdf9c9a43f0327efb4ebbe73c0d70e9743328a44dd5 Feb 27 17:23:12 crc kubenswrapper[4700]: I0227 17:23:12.907295 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-65457d8799-dclxm" Feb 27 17:23:12 crc kubenswrapper[4700]: I0227 17:23:12.911724 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-65457d8799-dclxm" Feb 27 17:23:13 crc kubenswrapper[4700]: I0227 17:23:13.537904 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-lf45g" event={"ID":"3cb8c359-c338-4350-acb2-b302786f082a","Type":"ContainerStarted","Data":"b892a4e5b88cec8f104a9fdf9c9a43f0327efb4ebbe73c0d70e9743328a44dd5"} Feb 27 17:23:14 crc kubenswrapper[4700]: I0227 17:23:14.553698 4700 generic.go:334] "Generic (PLEG): container finished" podID="0f1ae433-28e4-45cf-9635-dc6edbf16f6e" containerID="2d0da9f450ceed9ed3cba22dade279b2f7018efcb70068e35e804ea9219036d5" exitCode=1 Feb 27 17:23:14 crc kubenswrapper[4700]: I0227 17:23:14.553817 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"0f1ae433-28e4-45cf-9635-dc6edbf16f6e","Type":"ContainerDied","Data":"2d0da9f450ceed9ed3cba22dade279b2f7018efcb70068e35e804ea9219036d5"} Feb 27 17:23:14 crc kubenswrapper[4700]: I0227 17:23:14.554064 4700 scope.go:117] "RemoveContainer" containerID="9b51d27b3971cd3f655c413b4bd270b5cc55c1c7f20422db504613285c3a5770" Feb 27 17:23:14 crc kubenswrapper[4700]: I0227 17:23:14.555163 4700 scope.go:117] "RemoveContainer" containerID="2d0da9f450ceed9ed3cba22dade279b2f7018efcb70068e35e804ea9219036d5" Feb 27 17:23:14 crc kubenswrapper[4700]: E0227 17:23:14.555788 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 40s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(0f1ae433-28e4-45cf-9635-dc6edbf16f6e)\"" pod="openstack/watcher-decision-engine-0" podUID="0f1ae433-28e4-45cf-9635-dc6edbf16f6e" Feb 27 17:23:16 crc kubenswrapper[4700]: I0227 17:23:16.594078 4700 generic.go:334] "Generic (PLEG): container finished" podID="24cf75b6-81b2-43e8-b75e-c41171313a0f" containerID="3675602a5b830e0a8101ab8de56cf97a66e7e05593be9748eed8b289659a40a0" exitCode=0 Feb 27 17:23:16 crc kubenswrapper[4700]: I0227 17:23:16.594366 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7db57cbc8-t9p8l" event={"ID":"24cf75b6-81b2-43e8-b75e-c41171313a0f","Type":"ContainerDied","Data":"3675602a5b830e0a8101ab8de56cf97a66e7e05593be9748eed8b289659a40a0"} Feb 27 17:23:17 crc kubenswrapper[4700]: I0227 17:23:17.856953 4700 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/watcher-decision-engine-0" Feb 27 17:23:17 crc kubenswrapper[4700]: I0227 17:23:17.857382 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Feb 27 17:23:17 crc kubenswrapper[4700]: I0227 17:23:17.857397 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Feb 27 17:23:17 crc kubenswrapper[4700]: I0227 17:23:17.857407 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Feb 27 17:23:17 crc kubenswrapper[4700]: I0227 17:23:17.858069 4700 scope.go:117] "RemoveContainer" containerID="2d0da9f450ceed9ed3cba22dade279b2f7018efcb70068e35e804ea9219036d5" Feb 27 17:23:17 crc kubenswrapper[4700]: E0227 17:23:17.858375 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 40s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(0f1ae433-28e4-45cf-9635-dc6edbf16f6e)\"" pod="openstack/watcher-decision-engine-0" podUID="0f1ae433-28e4-45cf-9635-dc6edbf16f6e" Feb 27 17:23:20 crc kubenswrapper[4700]: I0227 17:23:20.967338 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-cc78b9998-2c2dk" Feb 27 17:23:21 crc kubenswrapper[4700]: I0227 17:23:21.057483 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-cc78b9998-2c2dk" Feb 27 17:23:21 crc kubenswrapper[4700]: I0227 17:23:21.150409 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-64fcf8677d-mvkvn"] Feb 27 17:23:21 crc kubenswrapper[4700]: I0227 17:23:21.150746 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-64fcf8677d-mvkvn" podUID="df366836-b6d3-42a1-ab1d-39f7d77f6cb5" containerName="placement-log" containerID="cri-o://f1e4c395d809f499ac4b00dbb1838b8f1f9a5f3addf25b4d24b7c1aa141ea5ce" gracePeriod=30 Feb 27 17:23:21 crc kubenswrapper[4700]: I0227 17:23:21.151287 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-64fcf8677d-mvkvn" podUID="df366836-b6d3-42a1-ab1d-39f7d77f6cb5" containerName="placement-api" containerID="cri-o://2c9bf816e687505a606f4fd060bda3c436925a4077ffaff3b8c6bf59ff37cda1" gracePeriod=30 Feb 27 17:23:21 crc kubenswrapper[4700]: I0227 17:23:21.174089 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 17:23:21 crc kubenswrapper[4700]: I0227 17:23:21.182614 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="3a7b7812-6e5f-4c57-a8d1-507e3dc2217c" containerName="glance-log" containerID="cri-o://71cbecfa6096e107490dd03a4e89a8a7ddf94f1f8f4c66dccfeb084fb8fb85c9" gracePeriod=30 Feb 27 17:23:21 crc kubenswrapper[4700]: I0227 17:23:21.182707 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="3a7b7812-6e5f-4c57-a8d1-507e3dc2217c" containerName="glance-httpd" containerID="cri-o://a6a7945d87c6b8692ed3f3f79afc16e0fa98c0e1a7600285da5d09af361e8f2c" gracePeriod=30 Feb 27 17:23:21 crc kubenswrapper[4700]: I0227 17:23:21.647216 4700 generic.go:334] "Generic (PLEG): container finished" podID="3a7b7812-6e5f-4c57-a8d1-507e3dc2217c" containerID="71cbecfa6096e107490dd03a4e89a8a7ddf94f1f8f4c66dccfeb084fb8fb85c9" exitCode=143 Feb 27 17:23:21 crc kubenswrapper[4700]: I0227 17:23:21.647312 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c","Type":"ContainerDied","Data":"71cbecfa6096e107490dd03a4e89a8a7ddf94f1f8f4c66dccfeb084fb8fb85c9"} Feb 27 17:23:21 crc kubenswrapper[4700]: I0227 17:23:21.661760 4700 generic.go:334] "Generic (PLEG): container finished" podID="df366836-b6d3-42a1-ab1d-39f7d77f6cb5" containerID="f1e4c395d809f499ac4b00dbb1838b8f1f9a5f3addf25b4d24b7c1aa141ea5ce" exitCode=143 Feb 27 17:23:21 crc kubenswrapper[4700]: I0227 17:23:21.662153 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-64fcf8677d-mvkvn" event={"ID":"df366836-b6d3-42a1-ab1d-39f7d77f6cb5","Type":"ContainerDied","Data":"f1e4c395d809f499ac4b00dbb1838b8f1f9a5f3addf25b4d24b7c1aa141ea5ce"} Feb 27 17:23:22 crc kubenswrapper[4700]: I0227 17:23:22.219090 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 17:23:22 crc kubenswrapper[4700]: I0227 17:23:22.219637 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="345c36c6-a3e3-4ece-9ac8-d59a7b31b366" containerName="glance-log" containerID="cri-o://3796d0ad84376c90605ec60063d83fa92dd2c330c228af0dd64e6141ad2e7bd0" gracePeriod=30 Feb 27 17:23:22 crc kubenswrapper[4700]: I0227 17:23:22.219746 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="345c36c6-a3e3-4ece-9ac8-d59a7b31b366" containerName="glance-httpd" containerID="cri-o://f3160c9d9756370992369dfa24520fea2a6602a148ba480ccc954e0d6a434526" gracePeriod=30 Feb 27 17:23:22 crc kubenswrapper[4700]: I0227 17:23:22.672940 4700 generic.go:334] "Generic (PLEG): container finished" podID="df366836-b6d3-42a1-ab1d-39f7d77f6cb5" containerID="2c9bf816e687505a606f4fd060bda3c436925a4077ffaff3b8c6bf59ff37cda1" exitCode=0 Feb 27 17:23:22 crc kubenswrapper[4700]: I0227 17:23:22.673012 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-64fcf8677d-mvkvn" event={"ID":"df366836-b6d3-42a1-ab1d-39f7d77f6cb5","Type":"ContainerDied","Data":"2c9bf816e687505a606f4fd060bda3c436925a4077ffaff3b8c6bf59ff37cda1"} Feb 27 17:23:22 crc kubenswrapper[4700]: I0227 17:23:22.674945 4700 generic.go:334] "Generic (PLEG): container finished" podID="345c36c6-a3e3-4ece-9ac8-d59a7b31b366" containerID="3796d0ad84376c90605ec60063d83fa92dd2c330c228af0dd64e6141ad2e7bd0" exitCode=143 Feb 27 17:23:22 crc kubenswrapper[4700]: I0227 17:23:22.675037 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"345c36c6-a3e3-4ece-9ac8-d59a7b31b366","Type":"ContainerDied","Data":"3796d0ad84376c90605ec60063d83fa92dd2c330c228af0dd64e6141ad2e7bd0"} Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.238918 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-64fcf8677d-mvkvn" Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.246876 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7db57cbc8-t9p8l" Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.259160 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/24cf75b6-81b2-43e8-b75e-c41171313a0f-ovndb-tls-certs\") pod \"24cf75b6-81b2-43e8-b75e-c41171313a0f\" (UID: \"24cf75b6-81b2-43e8-b75e-c41171313a0f\") " Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.259233 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-config-data\") pod \"df366836-b6d3-42a1-ab1d-39f7d77f6cb5\" (UID: \"df366836-b6d3-42a1-ab1d-39f7d77f6cb5\") " Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.259259 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-combined-ca-bundle\") pod \"df366836-b6d3-42a1-ab1d-39f7d77f6cb5\" (UID: \"df366836-b6d3-42a1-ab1d-39f7d77f6cb5\") " Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.259283 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-internal-tls-certs\") pod \"df366836-b6d3-42a1-ab1d-39f7d77f6cb5\" (UID: \"df366836-b6d3-42a1-ab1d-39f7d77f6cb5\") " Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.259309 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz8wr\" (UniqueName: \"kubernetes.io/projected/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-kube-api-access-lz8wr\") pod \"df366836-b6d3-42a1-ab1d-39f7d77f6cb5\" (UID: \"df366836-b6d3-42a1-ab1d-39f7d77f6cb5\") " Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.259488 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/24cf75b6-81b2-43e8-b75e-c41171313a0f-httpd-config\") pod \"24cf75b6-81b2-43e8-b75e-c41171313a0f\" (UID: \"24cf75b6-81b2-43e8-b75e-c41171313a0f\") " Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.259529 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-logs\") pod \"df366836-b6d3-42a1-ab1d-39f7d77f6cb5\" (UID: \"df366836-b6d3-42a1-ab1d-39f7d77f6cb5\") " Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.259553 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/24cf75b6-81b2-43e8-b75e-c41171313a0f-config\") pod \"24cf75b6-81b2-43e8-b75e-c41171313a0f\" (UID: \"24cf75b6-81b2-43e8-b75e-c41171313a0f\") " Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.259572 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24cf75b6-81b2-43e8-b75e-c41171313a0f-combined-ca-bundle\") pod \"24cf75b6-81b2-43e8-b75e-c41171313a0f\" (UID: \"24cf75b6-81b2-43e8-b75e-c41171313a0f\") " Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.259592 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-scripts\") pod \"df366836-b6d3-42a1-ab1d-39f7d77f6cb5\" (UID: \"df366836-b6d3-42a1-ab1d-39f7d77f6cb5\") " Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.259635 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-public-tls-certs\") pod \"df366836-b6d3-42a1-ab1d-39f7d77f6cb5\" (UID: \"df366836-b6d3-42a1-ab1d-39f7d77f6cb5\") " Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.262939 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-logs" (OuterVolumeSpecName: "logs") pod "df366836-b6d3-42a1-ab1d-39f7d77f6cb5" (UID: "df366836-b6d3-42a1-ab1d-39f7d77f6cb5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.303301 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24cf75b6-81b2-43e8-b75e-c41171313a0f-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "24cf75b6-81b2-43e8-b75e-c41171313a0f" (UID: "24cf75b6-81b2-43e8-b75e-c41171313a0f"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.304651 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-kube-api-access-lz8wr" (OuterVolumeSpecName: "kube-api-access-lz8wr") pod "df366836-b6d3-42a1-ab1d-39f7d77f6cb5" (UID: "df366836-b6d3-42a1-ab1d-39f7d77f6cb5"). InnerVolumeSpecName "kube-api-access-lz8wr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.306667 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-scripts" (OuterVolumeSpecName: "scripts") pod "df366836-b6d3-42a1-ab1d-39f7d77f6cb5" (UID: "df366836-b6d3-42a1-ab1d-39f7d77f6cb5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.347434 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-config-data" (OuterVolumeSpecName: "config-data") pod "df366836-b6d3-42a1-ab1d-39f7d77f6cb5" (UID: "df366836-b6d3-42a1-ab1d-39f7d77f6cb5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.370268 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmmqq\" (UniqueName: \"kubernetes.io/projected/24cf75b6-81b2-43e8-b75e-c41171313a0f-kube-api-access-xmmqq\") pod \"24cf75b6-81b2-43e8-b75e-c41171313a0f\" (UID: \"24cf75b6-81b2-43e8-b75e-c41171313a0f\") " Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.371009 4700 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-logs\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.371022 4700 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.371030 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.371039 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz8wr\" (UniqueName: \"kubernetes.io/projected/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-kube-api-access-lz8wr\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.371048 4700 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/24cf75b6-81b2-43e8-b75e-c41171313a0f-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.414628 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24cf75b6-81b2-43e8-b75e-c41171313a0f-kube-api-access-xmmqq" (OuterVolumeSpecName: "kube-api-access-xmmqq") pod "24cf75b6-81b2-43e8-b75e-c41171313a0f" (UID: "24cf75b6-81b2-43e8-b75e-c41171313a0f"). InnerVolumeSpecName "kube-api-access-xmmqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.420556 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24cf75b6-81b2-43e8-b75e-c41171313a0f-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "24cf75b6-81b2-43e8-b75e-c41171313a0f" (UID: "24cf75b6-81b2-43e8-b75e-c41171313a0f"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.421615 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24cf75b6-81b2-43e8-b75e-c41171313a0f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "24cf75b6-81b2-43e8-b75e-c41171313a0f" (UID: "24cf75b6-81b2-43e8-b75e-c41171313a0f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.433042 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "df366836-b6d3-42a1-ab1d-39f7d77f6cb5" (UID: "df366836-b6d3-42a1-ab1d-39f7d77f6cb5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.449505 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24cf75b6-81b2-43e8-b75e-c41171313a0f-config" (OuterVolumeSpecName: "config") pod "24cf75b6-81b2-43e8-b75e-c41171313a0f" (UID: "24cf75b6-81b2-43e8-b75e-c41171313a0f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.472621 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/24cf75b6-81b2-43e8-b75e-c41171313a0f-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.472656 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24cf75b6-81b2-43e8-b75e-c41171313a0f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.472666 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmmqq\" (UniqueName: \"kubernetes.io/projected/24cf75b6-81b2-43e8-b75e-c41171313a0f-kube-api-access-xmmqq\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.472675 4700 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/24cf75b6-81b2-43e8-b75e-c41171313a0f-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.472684 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.474125 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "df366836-b6d3-42a1-ab1d-39f7d77f6cb5" (UID: "df366836-b6d3-42a1-ab1d-39f7d77f6cb5"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.482149 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "df366836-b6d3-42a1-ab1d-39f7d77f6cb5" (UID: "df366836-b6d3-42a1-ab1d-39f7d77f6cb5"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.574931 4700 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.575265 4700 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/df366836-b6d3-42a1-ab1d-39f7d77f6cb5-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.690139 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7db57cbc8-t9p8l" event={"ID":"24cf75b6-81b2-43e8-b75e-c41171313a0f","Type":"ContainerDied","Data":"15635a8ed85c68ba78061124a91a84acf4bc7783426a91c3f6401f894c977b93"} Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.690209 4700 scope.go:117] "RemoveContainer" containerID="994e8654f68b07a71d26635805e774bb44d66badb496942fec0b2817d51b9f27" Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.690371 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7db57cbc8-t9p8l" Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.698769 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-64fcf8677d-mvkvn" event={"ID":"df366836-b6d3-42a1-ab1d-39f7d77f6cb5","Type":"ContainerDied","Data":"2cb5bceb568b9c7c3e8fee04bba319c843a556a14b7e1faa3c05e9e9033c865f"} Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.698879 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-64fcf8677d-mvkvn" Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.706147 4700 generic.go:334] "Generic (PLEG): container finished" podID="345c36c6-a3e3-4ece-9ac8-d59a7b31b366" containerID="f3160c9d9756370992369dfa24520fea2a6602a148ba480ccc954e0d6a434526" exitCode=0 Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.706242 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"345c36c6-a3e3-4ece-9ac8-d59a7b31b366","Type":"ContainerDied","Data":"f3160c9d9756370992369dfa24520fea2a6602a148ba480ccc954e0d6a434526"} Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.708060 4700 generic.go:334] "Generic (PLEG): container finished" podID="3a7b7812-6e5f-4c57-a8d1-507e3dc2217c" containerID="a6a7945d87c6b8692ed3f3f79afc16e0fa98c0e1a7600285da5d09af361e8f2c" exitCode=0 Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.708103 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c","Type":"ContainerDied","Data":"a6a7945d87c6b8692ed3f3f79afc16e0fa98c0e1a7600285da5d09af361e8f2c"} Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.757590 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-64fcf8677d-mvkvn"] Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.769000 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-64fcf8677d-mvkvn"] Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.781584 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7db57cbc8-t9p8l"] Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.784792 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.789956 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7db57cbc8-t9p8l"] Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.796801 4700 scope.go:117] "RemoveContainer" containerID="3675602a5b830e0a8101ab8de56cf97a66e7e05593be9748eed8b289659a40a0" Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.921148 4700 scope.go:117] "RemoveContainer" containerID="2c9bf816e687505a606f4fd060bda3c436925a4077ffaff3b8c6bf59ff37cda1" Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.965783 4700 scope.go:117] "RemoveContainer" containerID="f1e4c395d809f499ac4b00dbb1838b8f1f9a5f3addf25b4d24b7c1aa141ea5ce" Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.981410 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-config-data\") pod \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\" (UID: \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\") " Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.981572 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lt89w\" (UniqueName: \"kubernetes.io/projected/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-kube-api-access-lt89w\") pod \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\" (UID: \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\") " Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.981671 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-scripts\") pod \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\" (UID: \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\") " Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.981694 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-public-tls-certs\") pod \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\" (UID: \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\") " Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.981714 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-combined-ca-bundle\") pod \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\" (UID: \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\") " Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.981769 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-httpd-run\") pod \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\" (UID: \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\") " Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.981968 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-logs\") pod \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\" (UID: \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\") " Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.981987 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\" (UID: \"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c\") " Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.982645 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "3a7b7812-6e5f-4c57-a8d1-507e3dc2217c" (UID: "3a7b7812-6e5f-4c57-a8d1-507e3dc2217c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.982752 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-logs" (OuterVolumeSpecName: "logs") pod "3a7b7812-6e5f-4c57-a8d1-507e3dc2217c" (UID: "3a7b7812-6e5f-4c57-a8d1-507e3dc2217c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.985650 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-kube-api-access-lt89w" (OuterVolumeSpecName: "kube-api-access-lt89w") pod "3a7b7812-6e5f-4c57-a8d1-507e3dc2217c" (UID: "3a7b7812-6e5f-4c57-a8d1-507e3dc2217c"). InnerVolumeSpecName "kube-api-access-lt89w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.988773 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "3a7b7812-6e5f-4c57-a8d1-507e3dc2217c" (UID: "3a7b7812-6e5f-4c57-a8d1-507e3dc2217c"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 27 17:23:23 crc kubenswrapper[4700]: I0227 17:23:23.988784 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-scripts" (OuterVolumeSpecName: "scripts") pod "3a7b7812-6e5f-4c57-a8d1-507e3dc2217c" (UID: "3a7b7812-6e5f-4c57-a8d1-507e3dc2217c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.014959 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3a7b7812-6e5f-4c57-a8d1-507e3dc2217c" (UID: "3a7b7812-6e5f-4c57-a8d1-507e3dc2217c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.060690 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-config-data" (OuterVolumeSpecName: "config-data") pod "3a7b7812-6e5f-4c57-a8d1-507e3dc2217c" (UID: "3a7b7812-6e5f-4c57-a8d1-507e3dc2217c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.083492 4700 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.083596 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.084786 4700 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.084847 4700 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-logs\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.084918 4700 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.084983 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.085036 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lt89w\" (UniqueName: \"kubernetes.io/projected/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-kube-api-access-lt89w\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.097590 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "3a7b7812-6e5f-4c57-a8d1-507e3dc2217c" (UID: "3a7b7812-6e5f-4c57-a8d1-507e3dc2217c"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.107644 4700 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.187378 4700 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.187424 4700 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.214961 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.391079 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\" (UID: \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\") " Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.391130 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zr8ct\" (UniqueName: \"kubernetes.io/projected/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-kube-api-access-zr8ct\") pod \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\" (UID: \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\") " Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.391241 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-logs\") pod \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\" (UID: \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\") " Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.391270 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-scripts\") pod \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\" (UID: \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\") " Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.391291 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-internal-tls-certs\") pod \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\" (UID: \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\") " Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.391344 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-config-data\") pod \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\" (UID: \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\") " Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.391422 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-httpd-run\") pod \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\" (UID: \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\") " Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.391444 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-combined-ca-bundle\") pod \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\" (UID: \"345c36c6-a3e3-4ece-9ac8-d59a7b31b366\") " Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.392735 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "345c36c6-a3e3-4ece-9ac8-d59a7b31b366" (UID: "345c36c6-a3e3-4ece-9ac8-d59a7b31b366"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.392801 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-logs" (OuterVolumeSpecName: "logs") pod "345c36c6-a3e3-4ece-9ac8-d59a7b31b366" (UID: "345c36c6-a3e3-4ece-9ac8-d59a7b31b366"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.396066 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-scripts" (OuterVolumeSpecName: "scripts") pod "345c36c6-a3e3-4ece-9ac8-d59a7b31b366" (UID: "345c36c6-a3e3-4ece-9ac8-d59a7b31b366"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.396240 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-kube-api-access-zr8ct" (OuterVolumeSpecName: "kube-api-access-zr8ct") pod "345c36c6-a3e3-4ece-9ac8-d59a7b31b366" (UID: "345c36c6-a3e3-4ece-9ac8-d59a7b31b366"). InnerVolumeSpecName "kube-api-access-zr8ct". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.404124 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "345c36c6-a3e3-4ece-9ac8-d59a7b31b366" (UID: "345c36c6-a3e3-4ece-9ac8-d59a7b31b366"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.429934 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "345c36c6-a3e3-4ece-9ac8-d59a7b31b366" (UID: "345c36c6-a3e3-4ece-9ac8-d59a7b31b366"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.448850 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-config-data" (OuterVolumeSpecName: "config-data") pod "345c36c6-a3e3-4ece-9ac8-d59a7b31b366" (UID: "345c36c6-a3e3-4ece-9ac8-d59a7b31b366"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.457910 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "345c36c6-a3e3-4ece-9ac8-d59a7b31b366" (UID: "345c36c6-a3e3-4ece-9ac8-d59a7b31b366"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.494029 4700 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-logs\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.494063 4700 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.494073 4700 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.494084 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.494093 4700 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.494101 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.494133 4700 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.494145 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zr8ct\" (UniqueName: \"kubernetes.io/projected/345c36c6-a3e3-4ece-9ac8-d59a7b31b366-kube-api-access-zr8ct\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.523934 4700 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.596183 4700 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.720845 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3a7b7812-6e5f-4c57-a8d1-507e3dc2217c","Type":"ContainerDied","Data":"2db193f558ee0dfae7f8503c267a33ea0acab870e1ad92fbb4a0b47590f23bc3"} Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.720908 4700 scope.go:117] "RemoveContainer" containerID="a6a7945d87c6b8692ed3f3f79afc16e0fa98c0e1a7600285da5d09af361e8f2c" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.721061 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.735971 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-lf45g" event={"ID":"3cb8c359-c338-4350-acb2-b302786f082a","Type":"ContainerStarted","Data":"117d54daaae734326ad4742ce2bd14a1fb2c2af5276ef54492ec642e6f432eca"} Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.743888 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.743925 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"345c36c6-a3e3-4ece-9ac8-d59a7b31b366","Type":"ContainerDied","Data":"a98a8e42a97e9e2975b87c4917ec6b1475249cb171de127d2dc86371a5f57eda"} Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.763383 4700 scope.go:117] "RemoveContainer" containerID="71cbecfa6096e107490dd03a4e89a8a7ddf94f1f8f4c66dccfeb084fb8fb85c9" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.767563 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-lf45g" podStartSLOduration=1.68298831 podStartE2EDuration="12.767443018s" podCreationTimestamp="2026-02-27 17:23:12 +0000 UTC" firstStartedPulling="2026-02-27 17:23:12.882266608 +0000 UTC m=+1352.867579355" lastFinishedPulling="2026-02-27 17:23:23.966721316 +0000 UTC m=+1363.952034063" observedRunningTime="2026-02-27 17:23:24.76033195 +0000 UTC m=+1364.745644727" watchObservedRunningTime="2026-02-27 17:23:24.767443018 +0000 UTC m=+1364.752755765" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.794600 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.801925 4700 scope.go:117] "RemoveContainer" containerID="f3160c9d9756370992369dfa24520fea2a6602a148ba480ccc954e0d6a434526" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.807302 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.851297 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 17:23:24 crc kubenswrapper[4700]: E0227 17:23:24.864367 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24cf75b6-81b2-43e8-b75e-c41171313a0f" containerName="neutron-api" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.864610 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="24cf75b6-81b2-43e8-b75e-c41171313a0f" containerName="neutron-api" Feb 27 17:23:24 crc kubenswrapper[4700]: E0227 17:23:24.864758 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="345c36c6-a3e3-4ece-9ac8-d59a7b31b366" containerName="glance-log" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.864868 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="345c36c6-a3e3-4ece-9ac8-d59a7b31b366" containerName="glance-log" Feb 27 17:23:24 crc kubenswrapper[4700]: E0227 17:23:24.864964 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24cf75b6-81b2-43e8-b75e-c41171313a0f" containerName="neutron-httpd" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.865039 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="24cf75b6-81b2-43e8-b75e-c41171313a0f" containerName="neutron-httpd" Feb 27 17:23:24 crc kubenswrapper[4700]: E0227 17:23:24.865125 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="345c36c6-a3e3-4ece-9ac8-d59a7b31b366" containerName="glance-httpd" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.865242 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="345c36c6-a3e3-4ece-9ac8-d59a7b31b366" containerName="glance-httpd" Feb 27 17:23:24 crc kubenswrapper[4700]: E0227 17:23:24.865344 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df366836-b6d3-42a1-ab1d-39f7d77f6cb5" containerName="placement-api" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.865419 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="df366836-b6d3-42a1-ab1d-39f7d77f6cb5" containerName="placement-api" Feb 27 17:23:24 crc kubenswrapper[4700]: E0227 17:23:24.865523 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df366836-b6d3-42a1-ab1d-39f7d77f6cb5" containerName="placement-log" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.865600 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="df366836-b6d3-42a1-ab1d-39f7d77f6cb5" containerName="placement-log" Feb 27 17:23:24 crc kubenswrapper[4700]: E0227 17:23:24.865694 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a7b7812-6e5f-4c57-a8d1-507e3dc2217c" containerName="glance-httpd" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.865809 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a7b7812-6e5f-4c57-a8d1-507e3dc2217c" containerName="glance-httpd" Feb 27 17:23:24 crc kubenswrapper[4700]: E0227 17:23:24.865963 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a7b7812-6e5f-4c57-a8d1-507e3dc2217c" containerName="glance-log" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.866089 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a7b7812-6e5f-4c57-a8d1-507e3dc2217c" containerName="glance-log" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.866556 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="24cf75b6-81b2-43e8-b75e-c41171313a0f" containerName="neutron-httpd" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.866680 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="345c36c6-a3e3-4ece-9ac8-d59a7b31b366" containerName="glance-log" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.866772 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="24cf75b6-81b2-43e8-b75e-c41171313a0f" containerName="neutron-api" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.866863 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="df366836-b6d3-42a1-ab1d-39f7d77f6cb5" containerName="placement-api" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.866955 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="345c36c6-a3e3-4ece-9ac8-d59a7b31b366" containerName="glance-httpd" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.867043 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a7b7812-6e5f-4c57-a8d1-507e3dc2217c" containerName="glance-log" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.867138 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a7b7812-6e5f-4c57-a8d1-507e3dc2217c" containerName="glance-httpd" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.867223 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="df366836-b6d3-42a1-ab1d-39f7d77f6cb5" containerName="placement-log" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.864701 4700 scope.go:117] "RemoveContainer" containerID="3796d0ad84376c90605ec60063d83fa92dd2c330c228af0dd64e6141ad2e7bd0" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.869165 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.869285 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.871170 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-758nx" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.871372 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.871660 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.874754 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.884504 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.912951 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.927448 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.929182 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.933819 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.934785 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.944256 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.991332 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24cf75b6-81b2-43e8-b75e-c41171313a0f" path="/var/lib/kubelet/pods/24cf75b6-81b2-43e8-b75e-c41171313a0f/volumes" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.992263 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="345c36c6-a3e3-4ece-9ac8-d59a7b31b366" path="/var/lib/kubelet/pods/345c36c6-a3e3-4ece-9ac8-d59a7b31b366/volumes" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.993009 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a7b7812-6e5f-4c57-a8d1-507e3dc2217c" path="/var/lib/kubelet/pods/3a7b7812-6e5f-4c57-a8d1-507e3dc2217c/volumes" Feb 27 17:23:24 crc kubenswrapper[4700]: I0227 17:23:24.994171 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df366836-b6d3-42a1-ab1d-39f7d77f6cb5" path="/var/lib/kubelet/pods/df366836-b6d3-42a1-ab1d-39f7d77f6cb5/volumes" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.011637 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4bba45c6-8787-499a-a2a4-9f90e3ef1d7d-logs\") pod \"glance-default-external-api-0\" (UID: \"4bba45c6-8787-499a-a2a4-9f90e3ef1d7d\") " pod="openstack/glance-default-external-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.011688 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4bba45c6-8787-499a-a2a4-9f90e3ef1d7d-scripts\") pod \"glance-default-external-api-0\" (UID: \"4bba45c6-8787-499a-a2a4-9f90e3ef1d7d\") " pod="openstack/glance-default-external-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.011978 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7wq8\" (UniqueName: \"kubernetes.io/projected/4bba45c6-8787-499a-a2a4-9f90e3ef1d7d-kube-api-access-m7wq8\") pod \"glance-default-external-api-0\" (UID: \"4bba45c6-8787-499a-a2a4-9f90e3ef1d7d\") " pod="openstack/glance-default-external-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.012152 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bba45c6-8787-499a-a2a4-9f90e3ef1d7d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4bba45c6-8787-499a-a2a4-9f90e3ef1d7d\") " pod="openstack/glance-default-external-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.012324 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bba45c6-8787-499a-a2a4-9f90e3ef1d7d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"4bba45c6-8787-499a-a2a4-9f90e3ef1d7d\") " pod="openstack/glance-default-external-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.012378 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"4bba45c6-8787-499a-a2a4-9f90e3ef1d7d\") " pod="openstack/glance-default-external-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.012684 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4bba45c6-8787-499a-a2a4-9f90e3ef1d7d-config-data\") pod \"glance-default-external-api-0\" (UID: \"4bba45c6-8787-499a-a2a4-9f90e3ef1d7d\") " pod="openstack/glance-default-external-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.012719 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4bba45c6-8787-499a-a2a4-9f90e3ef1d7d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4bba45c6-8787-499a-a2a4-9f90e3ef1d7d\") " pod="openstack/glance-default-external-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.114818 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99391efa-6166-404b-9113-6e28f50d7136-logs\") pod \"glance-default-internal-api-0\" (UID: \"99391efa-6166-404b-9113-6e28f50d7136\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.114878 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4bba45c6-8787-499a-a2a4-9f90e3ef1d7d-config-data\") pod \"glance-default-external-api-0\" (UID: \"4bba45c6-8787-499a-a2a4-9f90e3ef1d7d\") " pod="openstack/glance-default-external-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.114905 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4bba45c6-8787-499a-a2a4-9f90e3ef1d7d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4bba45c6-8787-499a-a2a4-9f90e3ef1d7d\") " pod="openstack/glance-default-external-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.114945 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"99391efa-6166-404b-9113-6e28f50d7136\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.114968 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4bba45c6-8787-499a-a2a4-9f90e3ef1d7d-logs\") pod \"glance-default-external-api-0\" (UID: \"4bba45c6-8787-499a-a2a4-9f90e3ef1d7d\") " pod="openstack/glance-default-external-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.114991 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7mbz\" (UniqueName: \"kubernetes.io/projected/99391efa-6166-404b-9113-6e28f50d7136-kube-api-access-v7mbz\") pod \"glance-default-internal-api-0\" (UID: \"99391efa-6166-404b-9113-6e28f50d7136\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.115015 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4bba45c6-8787-499a-a2a4-9f90e3ef1d7d-scripts\") pod \"glance-default-external-api-0\" (UID: \"4bba45c6-8787-499a-a2a4-9f90e3ef1d7d\") " pod="openstack/glance-default-external-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.115032 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99391efa-6166-404b-9113-6e28f50d7136-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"99391efa-6166-404b-9113-6e28f50d7136\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.115071 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7wq8\" (UniqueName: \"kubernetes.io/projected/4bba45c6-8787-499a-a2a4-9f90e3ef1d7d-kube-api-access-m7wq8\") pod \"glance-default-external-api-0\" (UID: \"4bba45c6-8787-499a-a2a4-9f90e3ef1d7d\") " pod="openstack/glance-default-external-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.115096 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99391efa-6166-404b-9113-6e28f50d7136-config-data\") pod \"glance-default-internal-api-0\" (UID: \"99391efa-6166-404b-9113-6e28f50d7136\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.115128 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bba45c6-8787-499a-a2a4-9f90e3ef1d7d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4bba45c6-8787-499a-a2a4-9f90e3ef1d7d\") " pod="openstack/glance-default-external-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.115165 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99391efa-6166-404b-9113-6e28f50d7136-scripts\") pod \"glance-default-internal-api-0\" (UID: \"99391efa-6166-404b-9113-6e28f50d7136\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.115189 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/99391efa-6166-404b-9113-6e28f50d7136-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"99391efa-6166-404b-9113-6e28f50d7136\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.115212 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bba45c6-8787-499a-a2a4-9f90e3ef1d7d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"4bba45c6-8787-499a-a2a4-9f90e3ef1d7d\") " pod="openstack/glance-default-external-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.115233 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/99391efa-6166-404b-9113-6e28f50d7136-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"99391efa-6166-404b-9113-6e28f50d7136\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.115253 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"4bba45c6-8787-499a-a2a4-9f90e3ef1d7d\") " pod="openstack/glance-default-external-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.115654 4700 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"4bba45c6-8787-499a-a2a4-9f90e3ef1d7d\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.116308 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4bba45c6-8787-499a-a2a4-9f90e3ef1d7d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4bba45c6-8787-499a-a2a4-9f90e3ef1d7d\") " pod="openstack/glance-default-external-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.117019 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4bba45c6-8787-499a-a2a4-9f90e3ef1d7d-logs\") pod \"glance-default-external-api-0\" (UID: \"4bba45c6-8787-499a-a2a4-9f90e3ef1d7d\") " pod="openstack/glance-default-external-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.120886 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4bba45c6-8787-499a-a2a4-9f90e3ef1d7d-config-data\") pod \"glance-default-external-api-0\" (UID: \"4bba45c6-8787-499a-a2a4-9f90e3ef1d7d\") " pod="openstack/glance-default-external-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.124292 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4bba45c6-8787-499a-a2a4-9f90e3ef1d7d-scripts\") pod \"glance-default-external-api-0\" (UID: \"4bba45c6-8787-499a-a2a4-9f90e3ef1d7d\") " pod="openstack/glance-default-external-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.125750 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bba45c6-8787-499a-a2a4-9f90e3ef1d7d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4bba45c6-8787-499a-a2a4-9f90e3ef1d7d\") " pod="openstack/glance-default-external-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.129122 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bba45c6-8787-499a-a2a4-9f90e3ef1d7d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"4bba45c6-8787-499a-a2a4-9f90e3ef1d7d\") " pod="openstack/glance-default-external-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.137291 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7wq8\" (UniqueName: \"kubernetes.io/projected/4bba45c6-8787-499a-a2a4-9f90e3ef1d7d-kube-api-access-m7wq8\") pod \"glance-default-external-api-0\" (UID: \"4bba45c6-8787-499a-a2a4-9f90e3ef1d7d\") " pod="openstack/glance-default-external-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.150739 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"4bba45c6-8787-499a-a2a4-9f90e3ef1d7d\") " pod="openstack/glance-default-external-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.195183 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.216374 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99391efa-6166-404b-9113-6e28f50d7136-logs\") pod \"glance-default-internal-api-0\" (UID: \"99391efa-6166-404b-9113-6e28f50d7136\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.216446 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"99391efa-6166-404b-9113-6e28f50d7136\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.216487 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7mbz\" (UniqueName: \"kubernetes.io/projected/99391efa-6166-404b-9113-6e28f50d7136-kube-api-access-v7mbz\") pod \"glance-default-internal-api-0\" (UID: \"99391efa-6166-404b-9113-6e28f50d7136\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.216516 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99391efa-6166-404b-9113-6e28f50d7136-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"99391efa-6166-404b-9113-6e28f50d7136\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.216570 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99391efa-6166-404b-9113-6e28f50d7136-config-data\") pod \"glance-default-internal-api-0\" (UID: \"99391efa-6166-404b-9113-6e28f50d7136\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.216611 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99391efa-6166-404b-9113-6e28f50d7136-scripts\") pod \"glance-default-internal-api-0\" (UID: \"99391efa-6166-404b-9113-6e28f50d7136\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.216638 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/99391efa-6166-404b-9113-6e28f50d7136-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"99391efa-6166-404b-9113-6e28f50d7136\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.216666 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/99391efa-6166-404b-9113-6e28f50d7136-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"99391efa-6166-404b-9113-6e28f50d7136\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.216800 4700 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"99391efa-6166-404b-9113-6e28f50d7136\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.216910 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99391efa-6166-404b-9113-6e28f50d7136-logs\") pod \"glance-default-internal-api-0\" (UID: \"99391efa-6166-404b-9113-6e28f50d7136\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.220590 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/99391efa-6166-404b-9113-6e28f50d7136-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"99391efa-6166-404b-9113-6e28f50d7136\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.225479 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99391efa-6166-404b-9113-6e28f50d7136-config-data\") pod \"glance-default-internal-api-0\" (UID: \"99391efa-6166-404b-9113-6e28f50d7136\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.226244 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99391efa-6166-404b-9113-6e28f50d7136-scripts\") pod \"glance-default-internal-api-0\" (UID: \"99391efa-6166-404b-9113-6e28f50d7136\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.232140 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/99391efa-6166-404b-9113-6e28f50d7136-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"99391efa-6166-404b-9113-6e28f50d7136\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.242446 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99391efa-6166-404b-9113-6e28f50d7136-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"99391efa-6166-404b-9113-6e28f50d7136\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.249157 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7mbz\" (UniqueName: \"kubernetes.io/projected/99391efa-6166-404b-9113-6e28f50d7136-kube-api-access-v7mbz\") pod \"glance-default-internal-api-0\" (UID: \"99391efa-6166-404b-9113-6e28f50d7136\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.337866 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"99391efa-6166-404b-9113-6e28f50d7136\") " pod="openstack/glance-default-internal-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.552337 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 27 17:23:25 crc kubenswrapper[4700]: I0227 17:23:25.848335 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 17:23:26 crc kubenswrapper[4700]: I0227 17:23:26.088774 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 17:23:26 crc kubenswrapper[4700]: W0227 17:23:26.096290 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99391efa_6166_404b_9113_6e28f50d7136.slice/crio-8dcd52cec43053d6261c162e73ee0ee2fd2f3c5a4a49e61e13405bfcce04882b WatchSource:0}: Error finding container 8dcd52cec43053d6261c162e73ee0ee2fd2f3c5a4a49e61e13405bfcce04882b: Status 404 returned error can't find the container with id 8dcd52cec43053d6261c162e73ee0ee2fd2f3c5a4a49e61e13405bfcce04882b Feb 27 17:23:26 crc kubenswrapper[4700]: I0227 17:23:26.824807 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4bba45c6-8787-499a-a2a4-9f90e3ef1d7d","Type":"ContainerStarted","Data":"3df76f95fc75e2fc51a4b887e13e08ebc3375888d8ac90005028b6c38de0663f"} Feb 27 17:23:26 crc kubenswrapper[4700]: I0227 17:23:26.825083 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4bba45c6-8787-499a-a2a4-9f90e3ef1d7d","Type":"ContainerStarted","Data":"4535c41dc18abf8a9228bb993f0d8d24f9409d127f310e980e770e092369c146"} Feb 27 17:23:26 crc kubenswrapper[4700]: I0227 17:23:26.830571 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"99391efa-6166-404b-9113-6e28f50d7136","Type":"ContainerStarted","Data":"41e6dbf34f94e03a7a0228a30d68c6bc05347d4a9a577da7b7daad36a766e925"} Feb 27 17:23:26 crc kubenswrapper[4700]: I0227 17:23:26.830600 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"99391efa-6166-404b-9113-6e28f50d7136","Type":"ContainerStarted","Data":"8dcd52cec43053d6261c162e73ee0ee2fd2f3c5a4a49e61e13405bfcce04882b"} Feb 27 17:23:30 crc kubenswrapper[4700]: I0227 17:23:30.885867 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"99391efa-6166-404b-9113-6e28f50d7136","Type":"ContainerStarted","Data":"4a23fbea898d16f6eea15ca18460bfc571f69e1cb587fd16224347266c2ed646"} Feb 27 17:23:31 crc kubenswrapper[4700]: I0227 17:23:31.904497 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4bba45c6-8787-499a-a2a4-9f90e3ef1d7d","Type":"ContainerStarted","Data":"8da001ee954d8c4c8071c63f66d7d5d1f24550b9df6f6b848c4e825eab6ec9e6"} Feb 27 17:23:31 crc kubenswrapper[4700]: I0227 17:23:31.940323 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.940300483 podStartE2EDuration="7.940300483s" podCreationTimestamp="2026-02-27 17:23:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:23:31.929849297 +0000 UTC m=+1371.915162054" watchObservedRunningTime="2026-02-27 17:23:31.940300483 +0000 UTC m=+1371.925613240" Feb 27 17:23:31 crc kubenswrapper[4700]: I0227 17:23:31.962109 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=7.962086998 podStartE2EDuration="7.962086998s" podCreationTimestamp="2026-02-27 17:23:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:23:31.953799149 +0000 UTC m=+1371.939111906" watchObservedRunningTime="2026-02-27 17:23:31.962086998 +0000 UTC m=+1371.947399745" Feb 27 17:23:31 crc kubenswrapper[4700]: I0227 17:23:31.981131 4700 scope.go:117] "RemoveContainer" containerID="2d0da9f450ceed9ed3cba22dade279b2f7018efcb70068e35e804ea9219036d5" Feb 27 17:23:31 crc kubenswrapper[4700]: E0227 17:23:31.981555 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 40s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(0f1ae433-28e4-45cf-9635-dc6edbf16f6e)\"" pod="openstack/watcher-decision-engine-0" podUID="0f1ae433-28e4-45cf-9635-dc6edbf16f6e" Feb 27 17:23:33 crc kubenswrapper[4700]: I0227 17:23:33.946761 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="82104760-b734-4a3b-89c7-c6774f2ace27" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Feb 27 17:23:35 crc kubenswrapper[4700]: I0227 17:23:35.196208 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 27 17:23:35 crc kubenswrapper[4700]: I0227 17:23:35.196276 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 27 17:23:35 crc kubenswrapper[4700]: I0227 17:23:35.244650 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 27 17:23:35 crc kubenswrapper[4700]: I0227 17:23:35.250268 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 27 17:23:35 crc kubenswrapper[4700]: I0227 17:23:35.553436 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 27 17:23:35 crc kubenswrapper[4700]: I0227 17:23:35.553520 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 27 17:23:35 crc kubenswrapper[4700]: I0227 17:23:35.599455 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 27 17:23:35 crc kubenswrapper[4700]: I0227 17:23:35.607834 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 27 17:23:35 crc kubenswrapper[4700]: I0227 17:23:35.977355 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 27 17:23:35 crc kubenswrapper[4700]: I0227 17:23:35.977638 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 27 17:23:35 crc kubenswrapper[4700]: I0227 17:23:35.977649 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 27 17:23:35 crc kubenswrapper[4700]: I0227 17:23:35.977660 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 27 17:23:37 crc kubenswrapper[4700]: I0227 17:23:37.905606 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 27 17:23:37 crc kubenswrapper[4700]: I0227 17:23:37.924844 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 27 17:23:37 crc kubenswrapper[4700]: I0227 17:23:37.950958 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 27 17:23:37 crc kubenswrapper[4700]: I0227 17:23:37.955181 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 27 17:23:38 crc kubenswrapper[4700]: I0227 17:23:38.867982 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 17:23:38 crc kubenswrapper[4700]: I0227 17:23:38.950292 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82104760-b734-4a3b-89c7-c6774f2ace27-combined-ca-bundle\") pod \"82104760-b734-4a3b-89c7-c6774f2ace27\" (UID: \"82104760-b734-4a3b-89c7-c6774f2ace27\") " Feb 27 17:23:38 crc kubenswrapper[4700]: I0227 17:23:38.950357 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82104760-b734-4a3b-89c7-c6774f2ace27-log-httpd\") pod \"82104760-b734-4a3b-89c7-c6774f2ace27\" (UID: \"82104760-b734-4a3b-89c7-c6774f2ace27\") " Feb 27 17:23:38 crc kubenswrapper[4700]: I0227 17:23:38.950401 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bn9d8\" (UniqueName: \"kubernetes.io/projected/82104760-b734-4a3b-89c7-c6774f2ace27-kube-api-access-bn9d8\") pod \"82104760-b734-4a3b-89c7-c6774f2ace27\" (UID: \"82104760-b734-4a3b-89c7-c6774f2ace27\") " Feb 27 17:23:38 crc kubenswrapper[4700]: I0227 17:23:38.950431 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82104760-b734-4a3b-89c7-c6774f2ace27-scripts\") pod \"82104760-b734-4a3b-89c7-c6774f2ace27\" (UID: \"82104760-b734-4a3b-89c7-c6774f2ace27\") " Feb 27 17:23:38 crc kubenswrapper[4700]: I0227 17:23:38.950549 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82104760-b734-4a3b-89c7-c6774f2ace27-run-httpd\") pod \"82104760-b734-4a3b-89c7-c6774f2ace27\" (UID: \"82104760-b734-4a3b-89c7-c6774f2ace27\") " Feb 27 17:23:38 crc kubenswrapper[4700]: I0227 17:23:38.950626 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/82104760-b734-4a3b-89c7-c6774f2ace27-sg-core-conf-yaml\") pod \"82104760-b734-4a3b-89c7-c6774f2ace27\" (UID: \"82104760-b734-4a3b-89c7-c6774f2ace27\") " Feb 27 17:23:38 crc kubenswrapper[4700]: I0227 17:23:38.950649 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82104760-b734-4a3b-89c7-c6774f2ace27-config-data\") pod \"82104760-b734-4a3b-89c7-c6774f2ace27\" (UID: \"82104760-b734-4a3b-89c7-c6774f2ace27\") " Feb 27 17:23:38 crc kubenswrapper[4700]: I0227 17:23:38.952354 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82104760-b734-4a3b-89c7-c6774f2ace27-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "82104760-b734-4a3b-89c7-c6774f2ace27" (UID: "82104760-b734-4a3b-89c7-c6774f2ace27"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:23:38 crc kubenswrapper[4700]: I0227 17:23:38.952448 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82104760-b734-4a3b-89c7-c6774f2ace27-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "82104760-b734-4a3b-89c7-c6774f2ace27" (UID: "82104760-b734-4a3b-89c7-c6774f2ace27"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:23:38 crc kubenswrapper[4700]: I0227 17:23:38.957850 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82104760-b734-4a3b-89c7-c6774f2ace27-kube-api-access-bn9d8" (OuterVolumeSpecName: "kube-api-access-bn9d8") pod "82104760-b734-4a3b-89c7-c6774f2ace27" (UID: "82104760-b734-4a3b-89c7-c6774f2ace27"). InnerVolumeSpecName "kube-api-access-bn9d8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:23:38 crc kubenswrapper[4700]: I0227 17:23:38.958509 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82104760-b734-4a3b-89c7-c6774f2ace27-scripts" (OuterVolumeSpecName: "scripts") pod "82104760-b734-4a3b-89c7-c6774f2ace27" (UID: "82104760-b734-4a3b-89c7-c6774f2ace27"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:38 crc kubenswrapper[4700]: I0227 17:23:38.989674 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82104760-b734-4a3b-89c7-c6774f2ace27-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "82104760-b734-4a3b-89c7-c6774f2ace27" (UID: "82104760-b734-4a3b-89c7-c6774f2ace27"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.009725 4700 generic.go:334] "Generic (PLEG): container finished" podID="82104760-b734-4a3b-89c7-c6774f2ace27" containerID="9955ee7b957831bf50a5646e6ef0a6dd7561229e7e807d415c2c1aa5735544de" exitCode=137 Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.010740 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.048302 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82104760-b734-4a3b-89c7-c6774f2ace27-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "82104760-b734-4a3b-89c7-c6774f2ace27" (UID: "82104760-b734-4a3b-89c7-c6774f2ace27"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.053827 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82104760-b734-4a3b-89c7-c6774f2ace27-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.053859 4700 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82104760-b734-4a3b-89c7-c6774f2ace27-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.053871 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bn9d8\" (UniqueName: \"kubernetes.io/projected/82104760-b734-4a3b-89c7-c6774f2ace27-kube-api-access-bn9d8\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.053881 4700 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82104760-b734-4a3b-89c7-c6774f2ace27-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.053891 4700 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82104760-b734-4a3b-89c7-c6774f2ace27-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.053902 4700 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/82104760-b734-4a3b-89c7-c6774f2ace27-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.062949 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82104760-b734-4a3b-89c7-c6774f2ace27","Type":"ContainerDied","Data":"9955ee7b957831bf50a5646e6ef0a6dd7561229e7e807d415c2c1aa5735544de"} Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.063219 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82104760-b734-4a3b-89c7-c6774f2ace27","Type":"ContainerDied","Data":"c3f3c402d90fb56137e21ddac03eb3d3df93ebb104f63e6c1b858a3153a046ac"} Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.063336 4700 scope.go:117] "RemoveContainer" containerID="9955ee7b957831bf50a5646e6ef0a6dd7561229e7e807d415c2c1aa5735544de" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.082355 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82104760-b734-4a3b-89c7-c6774f2ace27-config-data" (OuterVolumeSpecName: "config-data") pod "82104760-b734-4a3b-89c7-c6774f2ace27" (UID: "82104760-b734-4a3b-89c7-c6774f2ace27"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.091933 4700 scope.go:117] "RemoveContainer" containerID="3c5f4b118989a58fc0d031ddfb3e2d6e548520a4718c3f9320a3df03d5564b40" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.109065 4700 scope.go:117] "RemoveContainer" containerID="f98afe6cc5de5a8cc7c6c4b99e4d6445bf8fc3572b8e74beeffc813c2c415e76" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.136366 4700 scope.go:117] "RemoveContainer" containerID="2fa25eca5355fa202d4ac28fe3ddfa48510e0c2e220d76a114db47de5a2e77c5" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.156195 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82104760-b734-4a3b-89c7-c6774f2ace27-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.158618 4700 scope.go:117] "RemoveContainer" containerID="9955ee7b957831bf50a5646e6ef0a6dd7561229e7e807d415c2c1aa5735544de" Feb 27 17:23:39 crc kubenswrapper[4700]: E0227 17:23:39.159015 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9955ee7b957831bf50a5646e6ef0a6dd7561229e7e807d415c2c1aa5735544de\": container with ID starting with 9955ee7b957831bf50a5646e6ef0a6dd7561229e7e807d415c2c1aa5735544de not found: ID does not exist" containerID="9955ee7b957831bf50a5646e6ef0a6dd7561229e7e807d415c2c1aa5735544de" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.159106 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9955ee7b957831bf50a5646e6ef0a6dd7561229e7e807d415c2c1aa5735544de"} err="failed to get container status \"9955ee7b957831bf50a5646e6ef0a6dd7561229e7e807d415c2c1aa5735544de\": rpc error: code = NotFound desc = could not find container \"9955ee7b957831bf50a5646e6ef0a6dd7561229e7e807d415c2c1aa5735544de\": container with ID starting with 9955ee7b957831bf50a5646e6ef0a6dd7561229e7e807d415c2c1aa5735544de not found: ID does not exist" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.159211 4700 scope.go:117] "RemoveContainer" containerID="3c5f4b118989a58fc0d031ddfb3e2d6e548520a4718c3f9320a3df03d5564b40" Feb 27 17:23:39 crc kubenswrapper[4700]: E0227 17:23:39.159482 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c5f4b118989a58fc0d031ddfb3e2d6e548520a4718c3f9320a3df03d5564b40\": container with ID starting with 3c5f4b118989a58fc0d031ddfb3e2d6e548520a4718c3f9320a3df03d5564b40 not found: ID does not exist" containerID="3c5f4b118989a58fc0d031ddfb3e2d6e548520a4718c3f9320a3df03d5564b40" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.159598 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c5f4b118989a58fc0d031ddfb3e2d6e548520a4718c3f9320a3df03d5564b40"} err="failed to get container status \"3c5f4b118989a58fc0d031ddfb3e2d6e548520a4718c3f9320a3df03d5564b40\": rpc error: code = NotFound desc = could not find container \"3c5f4b118989a58fc0d031ddfb3e2d6e548520a4718c3f9320a3df03d5564b40\": container with ID starting with 3c5f4b118989a58fc0d031ddfb3e2d6e548520a4718c3f9320a3df03d5564b40 not found: ID does not exist" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.159713 4700 scope.go:117] "RemoveContainer" containerID="f98afe6cc5de5a8cc7c6c4b99e4d6445bf8fc3572b8e74beeffc813c2c415e76" Feb 27 17:23:39 crc kubenswrapper[4700]: E0227 17:23:39.160085 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f98afe6cc5de5a8cc7c6c4b99e4d6445bf8fc3572b8e74beeffc813c2c415e76\": container with ID starting with f98afe6cc5de5a8cc7c6c4b99e4d6445bf8fc3572b8e74beeffc813c2c415e76 not found: ID does not exist" containerID="f98afe6cc5de5a8cc7c6c4b99e4d6445bf8fc3572b8e74beeffc813c2c415e76" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.160166 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f98afe6cc5de5a8cc7c6c4b99e4d6445bf8fc3572b8e74beeffc813c2c415e76"} err="failed to get container status \"f98afe6cc5de5a8cc7c6c4b99e4d6445bf8fc3572b8e74beeffc813c2c415e76\": rpc error: code = NotFound desc = could not find container \"f98afe6cc5de5a8cc7c6c4b99e4d6445bf8fc3572b8e74beeffc813c2c415e76\": container with ID starting with f98afe6cc5de5a8cc7c6c4b99e4d6445bf8fc3572b8e74beeffc813c2c415e76 not found: ID does not exist" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.160230 4700 scope.go:117] "RemoveContainer" containerID="2fa25eca5355fa202d4ac28fe3ddfa48510e0c2e220d76a114db47de5a2e77c5" Feb 27 17:23:39 crc kubenswrapper[4700]: E0227 17:23:39.160553 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fa25eca5355fa202d4ac28fe3ddfa48510e0c2e220d76a114db47de5a2e77c5\": container with ID starting with 2fa25eca5355fa202d4ac28fe3ddfa48510e0c2e220d76a114db47de5a2e77c5 not found: ID does not exist" containerID="2fa25eca5355fa202d4ac28fe3ddfa48510e0c2e220d76a114db47de5a2e77c5" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.160643 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fa25eca5355fa202d4ac28fe3ddfa48510e0c2e220d76a114db47de5a2e77c5"} err="failed to get container status \"2fa25eca5355fa202d4ac28fe3ddfa48510e0c2e220d76a114db47de5a2e77c5\": rpc error: code = NotFound desc = could not find container \"2fa25eca5355fa202d4ac28fe3ddfa48510e0c2e220d76a114db47de5a2e77c5\": container with ID starting with 2fa25eca5355fa202d4ac28fe3ddfa48510e0c2e220d76a114db47de5a2e77c5 not found: ID does not exist" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.355306 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.377819 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.405286 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:23:39 crc kubenswrapper[4700]: E0227 17:23:39.406014 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82104760-b734-4a3b-89c7-c6774f2ace27" containerName="sg-core" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.406108 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="82104760-b734-4a3b-89c7-c6774f2ace27" containerName="sg-core" Feb 27 17:23:39 crc kubenswrapper[4700]: E0227 17:23:39.406194 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82104760-b734-4a3b-89c7-c6774f2ace27" containerName="ceilometer-central-agent" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.406296 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="82104760-b734-4a3b-89c7-c6774f2ace27" containerName="ceilometer-central-agent" Feb 27 17:23:39 crc kubenswrapper[4700]: E0227 17:23:39.406392 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82104760-b734-4a3b-89c7-c6774f2ace27" containerName="proxy-httpd" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.406544 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="82104760-b734-4a3b-89c7-c6774f2ace27" containerName="proxy-httpd" Feb 27 17:23:39 crc kubenswrapper[4700]: E0227 17:23:39.406645 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82104760-b734-4a3b-89c7-c6774f2ace27" containerName="ceilometer-notification-agent" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.406722 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="82104760-b734-4a3b-89c7-c6774f2ace27" containerName="ceilometer-notification-agent" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.407047 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="82104760-b734-4a3b-89c7-c6774f2ace27" containerName="ceilometer-central-agent" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.407135 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="82104760-b734-4a3b-89c7-c6774f2ace27" containerName="ceilometer-notification-agent" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.407248 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="82104760-b734-4a3b-89c7-c6774f2ace27" containerName="sg-core" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.407326 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="82104760-b734-4a3b-89c7-c6774f2ace27" containerName="proxy-httpd" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.409430 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.411851 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.415967 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.427864 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.461817 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5237ed58-b5d6-49e9-b111-869ead2c5c9f-log-httpd\") pod \"ceilometer-0\" (UID: \"5237ed58-b5d6-49e9-b111-869ead2c5c9f\") " pod="openstack/ceilometer-0" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.462081 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5237ed58-b5d6-49e9-b111-869ead2c5c9f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5237ed58-b5d6-49e9-b111-869ead2c5c9f\") " pod="openstack/ceilometer-0" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.462190 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5237ed58-b5d6-49e9-b111-869ead2c5c9f-config-data\") pod \"ceilometer-0\" (UID: \"5237ed58-b5d6-49e9-b111-869ead2c5c9f\") " pod="openstack/ceilometer-0" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.462409 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5237ed58-b5d6-49e9-b111-869ead2c5c9f-run-httpd\") pod \"ceilometer-0\" (UID: \"5237ed58-b5d6-49e9-b111-869ead2c5c9f\") " pod="openstack/ceilometer-0" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.462539 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5237ed58-b5d6-49e9-b111-869ead2c5c9f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5237ed58-b5d6-49e9-b111-869ead2c5c9f\") " pod="openstack/ceilometer-0" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.462846 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5237ed58-b5d6-49e9-b111-869ead2c5c9f-scripts\") pod \"ceilometer-0\" (UID: \"5237ed58-b5d6-49e9-b111-869ead2c5c9f\") " pod="openstack/ceilometer-0" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.462901 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8wzz\" (UniqueName: \"kubernetes.io/projected/5237ed58-b5d6-49e9-b111-869ead2c5c9f-kube-api-access-l8wzz\") pod \"ceilometer-0\" (UID: \"5237ed58-b5d6-49e9-b111-869ead2c5c9f\") " pod="openstack/ceilometer-0" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.564267 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5237ed58-b5d6-49e9-b111-869ead2c5c9f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5237ed58-b5d6-49e9-b111-869ead2c5c9f\") " pod="openstack/ceilometer-0" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.564628 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5237ed58-b5d6-49e9-b111-869ead2c5c9f-scripts\") pod \"ceilometer-0\" (UID: \"5237ed58-b5d6-49e9-b111-869ead2c5c9f\") " pod="openstack/ceilometer-0" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.564734 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8wzz\" (UniqueName: \"kubernetes.io/projected/5237ed58-b5d6-49e9-b111-869ead2c5c9f-kube-api-access-l8wzz\") pod \"ceilometer-0\" (UID: \"5237ed58-b5d6-49e9-b111-869ead2c5c9f\") " pod="openstack/ceilometer-0" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.564852 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5237ed58-b5d6-49e9-b111-869ead2c5c9f-log-httpd\") pod \"ceilometer-0\" (UID: \"5237ed58-b5d6-49e9-b111-869ead2c5c9f\") " pod="openstack/ceilometer-0" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.564939 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5237ed58-b5d6-49e9-b111-869ead2c5c9f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5237ed58-b5d6-49e9-b111-869ead2c5c9f\") " pod="openstack/ceilometer-0" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.565019 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5237ed58-b5d6-49e9-b111-869ead2c5c9f-config-data\") pod \"ceilometer-0\" (UID: \"5237ed58-b5d6-49e9-b111-869ead2c5c9f\") " pod="openstack/ceilometer-0" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.565115 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5237ed58-b5d6-49e9-b111-869ead2c5c9f-run-httpd\") pod \"ceilometer-0\" (UID: \"5237ed58-b5d6-49e9-b111-869ead2c5c9f\") " pod="openstack/ceilometer-0" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.565523 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5237ed58-b5d6-49e9-b111-869ead2c5c9f-log-httpd\") pod \"ceilometer-0\" (UID: \"5237ed58-b5d6-49e9-b111-869ead2c5c9f\") " pod="openstack/ceilometer-0" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.565526 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5237ed58-b5d6-49e9-b111-869ead2c5c9f-run-httpd\") pod \"ceilometer-0\" (UID: \"5237ed58-b5d6-49e9-b111-869ead2c5c9f\") " pod="openstack/ceilometer-0" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.569059 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5237ed58-b5d6-49e9-b111-869ead2c5c9f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5237ed58-b5d6-49e9-b111-869ead2c5c9f\") " pod="openstack/ceilometer-0" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.572040 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5237ed58-b5d6-49e9-b111-869ead2c5c9f-scripts\") pod \"ceilometer-0\" (UID: \"5237ed58-b5d6-49e9-b111-869ead2c5c9f\") " pod="openstack/ceilometer-0" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.573142 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5237ed58-b5d6-49e9-b111-869ead2c5c9f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5237ed58-b5d6-49e9-b111-869ead2c5c9f\") " pod="openstack/ceilometer-0" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.573481 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5237ed58-b5d6-49e9-b111-869ead2c5c9f-config-data\") pod \"ceilometer-0\" (UID: \"5237ed58-b5d6-49e9-b111-869ead2c5c9f\") " pod="openstack/ceilometer-0" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.587608 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8wzz\" (UniqueName: \"kubernetes.io/projected/5237ed58-b5d6-49e9-b111-869ead2c5c9f-kube-api-access-l8wzz\") pod \"ceilometer-0\" (UID: \"5237ed58-b5d6-49e9-b111-869ead2c5c9f\") " pod="openstack/ceilometer-0" Feb 27 17:23:39 crc kubenswrapper[4700]: I0227 17:23:39.735754 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 17:23:40 crc kubenswrapper[4700]: I0227 17:23:40.223966 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:23:40 crc kubenswrapper[4700]: I0227 17:23:40.996707 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82104760-b734-4a3b-89c7-c6774f2ace27" path="/var/lib/kubelet/pods/82104760-b734-4a3b-89c7-c6774f2ace27/volumes" Feb 27 17:23:41 crc kubenswrapper[4700]: I0227 17:23:41.032362 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5237ed58-b5d6-49e9-b111-869ead2c5c9f","Type":"ContainerStarted","Data":"638a481b0dbf3e74fd9ec14224b127b722e31365890f6c7c4fbcdc6bc6383a89"} Feb 27 17:23:42 crc kubenswrapper[4700]: I0227 17:23:42.042120 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5237ed58-b5d6-49e9-b111-869ead2c5c9f","Type":"ContainerStarted","Data":"334c312b612849dc2eaad796bf75c4d3602478eb0ecdfb1b2ef242ed92b2ef48"} Feb 27 17:23:42 crc kubenswrapper[4700]: I0227 17:23:42.042627 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5237ed58-b5d6-49e9-b111-869ead2c5c9f","Type":"ContainerStarted","Data":"a7cf6456251254e5e1f7842969bc696e066f23a4b3f515b567d5c7c009f24ca5"} Feb 27 17:23:42 crc kubenswrapper[4700]: I0227 17:23:42.042657 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5237ed58-b5d6-49e9-b111-869ead2c5c9f","Type":"ContainerStarted","Data":"6d0a7dc194ff2f1c820c5b7eca3f67bae50f2437c78183ca1216491944741066"} Feb 27 17:23:45 crc kubenswrapper[4700]: I0227 17:23:45.069209 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5237ed58-b5d6-49e9-b111-869ead2c5c9f","Type":"ContainerStarted","Data":"3dd648808dd281515db809b33588cda5361358829f3653e1b1381314de447ce3"} Feb 27 17:23:45 crc kubenswrapper[4700]: I0227 17:23:45.069471 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 27 17:23:45 crc kubenswrapper[4700]: I0227 17:23:45.096090 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.84075009 podStartE2EDuration="6.096075595s" podCreationTimestamp="2026-02-27 17:23:39 +0000 UTC" firstStartedPulling="2026-02-27 17:23:40.213349462 +0000 UTC m=+1380.198662209" lastFinishedPulling="2026-02-27 17:23:44.468674967 +0000 UTC m=+1384.453987714" observedRunningTime="2026-02-27 17:23:45.088958678 +0000 UTC m=+1385.074271425" watchObservedRunningTime="2026-02-27 17:23:45.096075595 +0000 UTC m=+1385.081388342" Feb 27 17:23:45 crc kubenswrapper[4700]: I0227 17:23:45.981809 4700 scope.go:117] "RemoveContainer" containerID="2d0da9f450ceed9ed3cba22dade279b2f7018efcb70068e35e804ea9219036d5" Feb 27 17:23:45 crc kubenswrapper[4700]: E0227 17:23:45.982433 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 40s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(0f1ae433-28e4-45cf-9635-dc6edbf16f6e)\"" pod="openstack/watcher-decision-engine-0" podUID="0f1ae433-28e4-45cf-9635-dc6edbf16f6e" Feb 27 17:23:47 crc kubenswrapper[4700]: I0227 17:23:47.610361 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:23:47 crc kubenswrapper[4700]: I0227 17:23:47.611136 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5237ed58-b5d6-49e9-b111-869ead2c5c9f" containerName="ceilometer-central-agent" containerID="cri-o://6d0a7dc194ff2f1c820c5b7eca3f67bae50f2437c78183ca1216491944741066" gracePeriod=30 Feb 27 17:23:47 crc kubenswrapper[4700]: I0227 17:23:47.611255 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5237ed58-b5d6-49e9-b111-869ead2c5c9f" containerName="sg-core" containerID="cri-o://334c312b612849dc2eaad796bf75c4d3602478eb0ecdfb1b2ef242ed92b2ef48" gracePeriod=30 Feb 27 17:23:47 crc kubenswrapper[4700]: I0227 17:23:47.611428 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5237ed58-b5d6-49e9-b111-869ead2c5c9f" containerName="proxy-httpd" containerID="cri-o://3dd648808dd281515db809b33588cda5361358829f3653e1b1381314de447ce3" gracePeriod=30 Feb 27 17:23:47 crc kubenswrapper[4700]: I0227 17:23:47.611486 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5237ed58-b5d6-49e9-b111-869ead2c5c9f" containerName="ceilometer-notification-agent" containerID="cri-o://a7cf6456251254e5e1f7842969bc696e066f23a4b3f515b567d5c7c009f24ca5" gracePeriod=30 Feb 27 17:23:48 crc kubenswrapper[4700]: I0227 17:23:48.098986 4700 generic.go:334] "Generic (PLEG): container finished" podID="5237ed58-b5d6-49e9-b111-869ead2c5c9f" containerID="3dd648808dd281515db809b33588cda5361358829f3653e1b1381314de447ce3" exitCode=0 Feb 27 17:23:48 crc kubenswrapper[4700]: I0227 17:23:48.099317 4700 generic.go:334] "Generic (PLEG): container finished" podID="5237ed58-b5d6-49e9-b111-869ead2c5c9f" containerID="334c312b612849dc2eaad796bf75c4d3602478eb0ecdfb1b2ef242ed92b2ef48" exitCode=2 Feb 27 17:23:48 crc kubenswrapper[4700]: I0227 17:23:48.099080 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5237ed58-b5d6-49e9-b111-869ead2c5c9f","Type":"ContainerDied","Data":"3dd648808dd281515db809b33588cda5361358829f3653e1b1381314de447ce3"} Feb 27 17:23:48 crc kubenswrapper[4700]: I0227 17:23:48.099362 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5237ed58-b5d6-49e9-b111-869ead2c5c9f","Type":"ContainerDied","Data":"334c312b612849dc2eaad796bf75c4d3602478eb0ecdfb1b2ef242ed92b2ef48"} Feb 27 17:23:49 crc kubenswrapper[4700]: I0227 17:23:49.109874 4700 generic.go:334] "Generic (PLEG): container finished" podID="5237ed58-b5d6-49e9-b111-869ead2c5c9f" containerID="a7cf6456251254e5e1f7842969bc696e066f23a4b3f515b567d5c7c009f24ca5" exitCode=0 Feb 27 17:23:49 crc kubenswrapper[4700]: I0227 17:23:49.109921 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5237ed58-b5d6-49e9-b111-869ead2c5c9f","Type":"ContainerDied","Data":"a7cf6456251254e5e1f7842969bc696e066f23a4b3f515b567d5c7c009f24ca5"} Feb 27 17:23:50 crc kubenswrapper[4700]: I0227 17:23:50.123766 4700 generic.go:334] "Generic (PLEG): container finished" podID="3cb8c359-c338-4350-acb2-b302786f082a" containerID="117d54daaae734326ad4742ce2bd14a1fb2c2af5276ef54492ec642e6f432eca" exitCode=0 Feb 27 17:23:50 crc kubenswrapper[4700]: I0227 17:23:50.124586 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-lf45g" event={"ID":"3cb8c359-c338-4350-acb2-b302786f082a","Type":"ContainerDied","Data":"117d54daaae734326ad4742ce2bd14a1fb2c2af5276ef54492ec642e6f432eca"} Feb 27 17:23:51 crc kubenswrapper[4700]: I0227 17:23:51.595621 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-lf45g" Feb 27 17:23:51 crc kubenswrapper[4700]: I0227 17:23:51.660571 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cb8c359-c338-4350-acb2-b302786f082a-combined-ca-bundle\") pod \"3cb8c359-c338-4350-acb2-b302786f082a\" (UID: \"3cb8c359-c338-4350-acb2-b302786f082a\") " Feb 27 17:23:51 crc kubenswrapper[4700]: I0227 17:23:51.660674 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zp8v4\" (UniqueName: \"kubernetes.io/projected/3cb8c359-c338-4350-acb2-b302786f082a-kube-api-access-zp8v4\") pod \"3cb8c359-c338-4350-acb2-b302786f082a\" (UID: \"3cb8c359-c338-4350-acb2-b302786f082a\") " Feb 27 17:23:51 crc kubenswrapper[4700]: I0227 17:23:51.660756 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cb8c359-c338-4350-acb2-b302786f082a-scripts\") pod \"3cb8c359-c338-4350-acb2-b302786f082a\" (UID: \"3cb8c359-c338-4350-acb2-b302786f082a\") " Feb 27 17:23:51 crc kubenswrapper[4700]: I0227 17:23:51.660802 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cb8c359-c338-4350-acb2-b302786f082a-config-data\") pod \"3cb8c359-c338-4350-acb2-b302786f082a\" (UID: \"3cb8c359-c338-4350-acb2-b302786f082a\") " Feb 27 17:23:51 crc kubenswrapper[4700]: I0227 17:23:51.668593 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb8c359-c338-4350-acb2-b302786f082a-kube-api-access-zp8v4" (OuterVolumeSpecName: "kube-api-access-zp8v4") pod "3cb8c359-c338-4350-acb2-b302786f082a" (UID: "3cb8c359-c338-4350-acb2-b302786f082a"). InnerVolumeSpecName "kube-api-access-zp8v4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:23:51 crc kubenswrapper[4700]: I0227 17:23:51.673288 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cb8c359-c338-4350-acb2-b302786f082a-scripts" (OuterVolumeSpecName: "scripts") pod "3cb8c359-c338-4350-acb2-b302786f082a" (UID: "3cb8c359-c338-4350-acb2-b302786f082a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:51 crc kubenswrapper[4700]: I0227 17:23:51.703112 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cb8c359-c338-4350-acb2-b302786f082a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3cb8c359-c338-4350-acb2-b302786f082a" (UID: "3cb8c359-c338-4350-acb2-b302786f082a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:51 crc kubenswrapper[4700]: I0227 17:23:51.724639 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cb8c359-c338-4350-acb2-b302786f082a-config-data" (OuterVolumeSpecName: "config-data") pod "3cb8c359-c338-4350-acb2-b302786f082a" (UID: "3cb8c359-c338-4350-acb2-b302786f082a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:51 crc kubenswrapper[4700]: I0227 17:23:51.765839 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cb8c359-c338-4350-acb2-b302786f082a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:51 crc kubenswrapper[4700]: I0227 17:23:51.766187 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zp8v4\" (UniqueName: \"kubernetes.io/projected/3cb8c359-c338-4350-acb2-b302786f082a-kube-api-access-zp8v4\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:51 crc kubenswrapper[4700]: I0227 17:23:51.766239 4700 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cb8c359-c338-4350-acb2-b302786f082a-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:51 crc kubenswrapper[4700]: I0227 17:23:51.766254 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cb8c359-c338-4350-acb2-b302786f082a-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:51 crc kubenswrapper[4700]: I0227 17:23:51.836792 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 17:23:51 crc kubenswrapper[4700]: I0227 17:23:51.969217 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5237ed58-b5d6-49e9-b111-869ead2c5c9f-scripts\") pod \"5237ed58-b5d6-49e9-b111-869ead2c5c9f\" (UID: \"5237ed58-b5d6-49e9-b111-869ead2c5c9f\") " Feb 27 17:23:51 crc kubenswrapper[4700]: I0227 17:23:51.969384 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5237ed58-b5d6-49e9-b111-869ead2c5c9f-run-httpd\") pod \"5237ed58-b5d6-49e9-b111-869ead2c5c9f\" (UID: \"5237ed58-b5d6-49e9-b111-869ead2c5c9f\") " Feb 27 17:23:51 crc kubenswrapper[4700]: I0227 17:23:51.969443 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5237ed58-b5d6-49e9-b111-869ead2c5c9f-combined-ca-bundle\") pod \"5237ed58-b5d6-49e9-b111-869ead2c5c9f\" (UID: \"5237ed58-b5d6-49e9-b111-869ead2c5c9f\") " Feb 27 17:23:51 crc kubenswrapper[4700]: I0227 17:23:51.969502 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5237ed58-b5d6-49e9-b111-869ead2c5c9f-log-httpd\") pod \"5237ed58-b5d6-49e9-b111-869ead2c5c9f\" (UID: \"5237ed58-b5d6-49e9-b111-869ead2c5c9f\") " Feb 27 17:23:51 crc kubenswrapper[4700]: I0227 17:23:51.969562 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5237ed58-b5d6-49e9-b111-869ead2c5c9f-config-data\") pod \"5237ed58-b5d6-49e9-b111-869ead2c5c9f\" (UID: \"5237ed58-b5d6-49e9-b111-869ead2c5c9f\") " Feb 27 17:23:51 crc kubenswrapper[4700]: I0227 17:23:51.969584 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8wzz\" (UniqueName: \"kubernetes.io/projected/5237ed58-b5d6-49e9-b111-869ead2c5c9f-kube-api-access-l8wzz\") pod \"5237ed58-b5d6-49e9-b111-869ead2c5c9f\" (UID: \"5237ed58-b5d6-49e9-b111-869ead2c5c9f\") " Feb 27 17:23:51 crc kubenswrapper[4700]: I0227 17:23:51.969638 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5237ed58-b5d6-49e9-b111-869ead2c5c9f-sg-core-conf-yaml\") pod \"5237ed58-b5d6-49e9-b111-869ead2c5c9f\" (UID: \"5237ed58-b5d6-49e9-b111-869ead2c5c9f\") " Feb 27 17:23:51 crc kubenswrapper[4700]: I0227 17:23:51.969922 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5237ed58-b5d6-49e9-b111-869ead2c5c9f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5237ed58-b5d6-49e9-b111-869ead2c5c9f" (UID: "5237ed58-b5d6-49e9-b111-869ead2c5c9f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:23:51 crc kubenswrapper[4700]: I0227 17:23:51.970151 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5237ed58-b5d6-49e9-b111-869ead2c5c9f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5237ed58-b5d6-49e9-b111-869ead2c5c9f" (UID: "5237ed58-b5d6-49e9-b111-869ead2c5c9f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:23:51 crc kubenswrapper[4700]: I0227 17:23:51.970220 4700 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5237ed58-b5d6-49e9-b111-869ead2c5c9f-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:51 crc kubenswrapper[4700]: I0227 17:23:51.973210 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5237ed58-b5d6-49e9-b111-869ead2c5c9f-scripts" (OuterVolumeSpecName: "scripts") pod "5237ed58-b5d6-49e9-b111-869ead2c5c9f" (UID: "5237ed58-b5d6-49e9-b111-869ead2c5c9f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:51 crc kubenswrapper[4700]: I0227 17:23:51.974256 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5237ed58-b5d6-49e9-b111-869ead2c5c9f-kube-api-access-l8wzz" (OuterVolumeSpecName: "kube-api-access-l8wzz") pod "5237ed58-b5d6-49e9-b111-869ead2c5c9f" (UID: "5237ed58-b5d6-49e9-b111-869ead2c5c9f"). InnerVolumeSpecName "kube-api-access-l8wzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:23:51 crc kubenswrapper[4700]: I0227 17:23:51.993631 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5237ed58-b5d6-49e9-b111-869ead2c5c9f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5237ed58-b5d6-49e9-b111-869ead2c5c9f" (UID: "5237ed58-b5d6-49e9-b111-869ead2c5c9f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.057319 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5237ed58-b5d6-49e9-b111-869ead2c5c9f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5237ed58-b5d6-49e9-b111-869ead2c5c9f" (UID: "5237ed58-b5d6-49e9-b111-869ead2c5c9f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.072323 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8wzz\" (UniqueName: \"kubernetes.io/projected/5237ed58-b5d6-49e9-b111-869ead2c5c9f-kube-api-access-l8wzz\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.072361 4700 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5237ed58-b5d6-49e9-b111-869ead2c5c9f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.072376 4700 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5237ed58-b5d6-49e9-b111-869ead2c5c9f-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.072388 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5237ed58-b5d6-49e9-b111-869ead2c5c9f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.072400 4700 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5237ed58-b5d6-49e9-b111-869ead2c5c9f-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.091998 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5237ed58-b5d6-49e9-b111-869ead2c5c9f-config-data" (OuterVolumeSpecName: "config-data") pod "5237ed58-b5d6-49e9-b111-869ead2c5c9f" (UID: "5237ed58-b5d6-49e9-b111-869ead2c5c9f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.152136 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-lf45g" event={"ID":"3cb8c359-c338-4350-acb2-b302786f082a","Type":"ContainerDied","Data":"b892a4e5b88cec8f104a9fdf9c9a43f0327efb4ebbe73c0d70e9743328a44dd5"} Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.152182 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b892a4e5b88cec8f104a9fdf9c9a43f0327efb4ebbe73c0d70e9743328a44dd5" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.152254 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-lf45g" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.155991 4700 generic.go:334] "Generic (PLEG): container finished" podID="5237ed58-b5d6-49e9-b111-869ead2c5c9f" containerID="6d0a7dc194ff2f1c820c5b7eca3f67bae50f2437c78183ca1216491944741066" exitCode=0 Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.156040 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.156087 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5237ed58-b5d6-49e9-b111-869ead2c5c9f","Type":"ContainerDied","Data":"6d0a7dc194ff2f1c820c5b7eca3f67bae50f2437c78183ca1216491944741066"} Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.156711 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5237ed58-b5d6-49e9-b111-869ead2c5c9f","Type":"ContainerDied","Data":"638a481b0dbf3e74fd9ec14224b127b722e31365890f6c7c4fbcdc6bc6383a89"} Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.156768 4700 scope.go:117] "RemoveContainer" containerID="3dd648808dd281515db809b33588cda5361358829f3653e1b1381314de447ce3" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.181718 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5237ed58-b5d6-49e9-b111-869ead2c5c9f-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.200260 4700 scope.go:117] "RemoveContainer" containerID="334c312b612849dc2eaad796bf75c4d3602478eb0ecdfb1b2ef242ed92b2ef48" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.218340 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.241374 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.252607 4700 scope.go:117] "RemoveContainer" containerID="a7cf6456251254e5e1f7842969bc696e066f23a4b3f515b567d5c7c009f24ca5" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.263284 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:23:52 crc kubenswrapper[4700]: E0227 17:23:52.263711 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5237ed58-b5d6-49e9-b111-869ead2c5c9f" containerName="ceilometer-notification-agent" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.263724 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="5237ed58-b5d6-49e9-b111-869ead2c5c9f" containerName="ceilometer-notification-agent" Feb 27 17:23:52 crc kubenswrapper[4700]: E0227 17:23:52.263738 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cb8c359-c338-4350-acb2-b302786f082a" containerName="nova-cell0-conductor-db-sync" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.263744 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cb8c359-c338-4350-acb2-b302786f082a" containerName="nova-cell0-conductor-db-sync" Feb 27 17:23:52 crc kubenswrapper[4700]: E0227 17:23:52.263757 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5237ed58-b5d6-49e9-b111-869ead2c5c9f" containerName="proxy-httpd" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.263762 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="5237ed58-b5d6-49e9-b111-869ead2c5c9f" containerName="proxy-httpd" Feb 27 17:23:52 crc kubenswrapper[4700]: E0227 17:23:52.263773 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5237ed58-b5d6-49e9-b111-869ead2c5c9f" containerName="sg-core" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.263779 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="5237ed58-b5d6-49e9-b111-869ead2c5c9f" containerName="sg-core" Feb 27 17:23:52 crc kubenswrapper[4700]: E0227 17:23:52.263788 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5237ed58-b5d6-49e9-b111-869ead2c5c9f" containerName="ceilometer-central-agent" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.263793 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="5237ed58-b5d6-49e9-b111-869ead2c5c9f" containerName="ceilometer-central-agent" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.263967 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="5237ed58-b5d6-49e9-b111-869ead2c5c9f" containerName="sg-core" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.263983 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="5237ed58-b5d6-49e9-b111-869ead2c5c9f" containerName="proxy-httpd" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.263992 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cb8c359-c338-4350-acb2-b302786f082a" containerName="nova-cell0-conductor-db-sync" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.264001 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="5237ed58-b5d6-49e9-b111-869ead2c5c9f" containerName="ceilometer-central-agent" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.264016 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="5237ed58-b5d6-49e9-b111-869ead2c5c9f" containerName="ceilometer-notification-agent" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.265829 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.270799 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.271040 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.281393 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.311537 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.312811 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.314584 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-ctcbq" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.316046 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.316630 4700 scope.go:117] "RemoveContainer" containerID="6d0a7dc194ff2f1c820c5b7eca3f67bae50f2437c78183ca1216491944741066" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.341406 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.343373 4700 scope.go:117] "RemoveContainer" containerID="3dd648808dd281515db809b33588cda5361358829f3653e1b1381314de447ce3" Feb 27 17:23:52 crc kubenswrapper[4700]: E0227 17:23:52.343803 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3dd648808dd281515db809b33588cda5361358829f3653e1b1381314de447ce3\": container with ID starting with 3dd648808dd281515db809b33588cda5361358829f3653e1b1381314de447ce3 not found: ID does not exist" containerID="3dd648808dd281515db809b33588cda5361358829f3653e1b1381314de447ce3" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.343848 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3dd648808dd281515db809b33588cda5361358829f3653e1b1381314de447ce3"} err="failed to get container status \"3dd648808dd281515db809b33588cda5361358829f3653e1b1381314de447ce3\": rpc error: code = NotFound desc = could not find container \"3dd648808dd281515db809b33588cda5361358829f3653e1b1381314de447ce3\": container with ID starting with 3dd648808dd281515db809b33588cda5361358829f3653e1b1381314de447ce3 not found: ID does not exist" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.343866 4700 scope.go:117] "RemoveContainer" containerID="334c312b612849dc2eaad796bf75c4d3602478eb0ecdfb1b2ef242ed92b2ef48" Feb 27 17:23:52 crc kubenswrapper[4700]: E0227 17:23:52.344088 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"334c312b612849dc2eaad796bf75c4d3602478eb0ecdfb1b2ef242ed92b2ef48\": container with ID starting with 334c312b612849dc2eaad796bf75c4d3602478eb0ecdfb1b2ef242ed92b2ef48 not found: ID does not exist" containerID="334c312b612849dc2eaad796bf75c4d3602478eb0ecdfb1b2ef242ed92b2ef48" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.344102 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"334c312b612849dc2eaad796bf75c4d3602478eb0ecdfb1b2ef242ed92b2ef48"} err="failed to get container status \"334c312b612849dc2eaad796bf75c4d3602478eb0ecdfb1b2ef242ed92b2ef48\": rpc error: code = NotFound desc = could not find container \"334c312b612849dc2eaad796bf75c4d3602478eb0ecdfb1b2ef242ed92b2ef48\": container with ID starting with 334c312b612849dc2eaad796bf75c4d3602478eb0ecdfb1b2ef242ed92b2ef48 not found: ID does not exist" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.344113 4700 scope.go:117] "RemoveContainer" containerID="a7cf6456251254e5e1f7842969bc696e066f23a4b3f515b567d5c7c009f24ca5" Feb 27 17:23:52 crc kubenswrapper[4700]: E0227 17:23:52.344320 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7cf6456251254e5e1f7842969bc696e066f23a4b3f515b567d5c7c009f24ca5\": container with ID starting with a7cf6456251254e5e1f7842969bc696e066f23a4b3f515b567d5c7c009f24ca5 not found: ID does not exist" containerID="a7cf6456251254e5e1f7842969bc696e066f23a4b3f515b567d5c7c009f24ca5" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.344333 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7cf6456251254e5e1f7842969bc696e066f23a4b3f515b567d5c7c009f24ca5"} err="failed to get container status \"a7cf6456251254e5e1f7842969bc696e066f23a4b3f515b567d5c7c009f24ca5\": rpc error: code = NotFound desc = could not find container \"a7cf6456251254e5e1f7842969bc696e066f23a4b3f515b567d5c7c009f24ca5\": container with ID starting with a7cf6456251254e5e1f7842969bc696e066f23a4b3f515b567d5c7c009f24ca5 not found: ID does not exist" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.344344 4700 scope.go:117] "RemoveContainer" containerID="6d0a7dc194ff2f1c820c5b7eca3f67bae50f2437c78183ca1216491944741066" Feb 27 17:23:52 crc kubenswrapper[4700]: E0227 17:23:52.345412 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d0a7dc194ff2f1c820c5b7eca3f67bae50f2437c78183ca1216491944741066\": container with ID starting with 6d0a7dc194ff2f1c820c5b7eca3f67bae50f2437c78183ca1216491944741066 not found: ID does not exist" containerID="6d0a7dc194ff2f1c820c5b7eca3f67bae50f2437c78183ca1216491944741066" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.345431 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d0a7dc194ff2f1c820c5b7eca3f67bae50f2437c78183ca1216491944741066"} err="failed to get container status \"6d0a7dc194ff2f1c820c5b7eca3f67bae50f2437c78183ca1216491944741066\": rpc error: code = NotFound desc = could not find container \"6d0a7dc194ff2f1c820c5b7eca3f67bae50f2437c78183ca1216491944741066\": container with ID starting with 6d0a7dc194ff2f1c820c5b7eca3f67bae50f2437c78183ca1216491944741066 not found: ID does not exist" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.384633 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/338e2de1-bc7a-4224-a7c1-e90c99eba72e-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"338e2de1-bc7a-4224-a7c1-e90c99eba72e\") " pod="openstack/nova-cell0-conductor-0" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.384714 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ce63ee0-bc69-4649-b644-34854c003845-scripts\") pod \"ceilometer-0\" (UID: \"7ce63ee0-bc69-4649-b644-34854c003845\") " pod="openstack/ceilometer-0" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.384935 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68z8n\" (UniqueName: \"kubernetes.io/projected/7ce63ee0-bc69-4649-b644-34854c003845-kube-api-access-68z8n\") pod \"ceilometer-0\" (UID: \"7ce63ee0-bc69-4649-b644-34854c003845\") " pod="openstack/ceilometer-0" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.385032 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ce63ee0-bc69-4649-b644-34854c003845-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7ce63ee0-bc69-4649-b644-34854c003845\") " pod="openstack/ceilometer-0" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.385057 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ce63ee0-bc69-4649-b644-34854c003845-config-data\") pod \"ceilometer-0\" (UID: \"7ce63ee0-bc69-4649-b644-34854c003845\") " pod="openstack/ceilometer-0" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.385197 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/338e2de1-bc7a-4224-a7c1-e90c99eba72e-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"338e2de1-bc7a-4224-a7c1-e90c99eba72e\") " pod="openstack/nova-cell0-conductor-0" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.385348 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7ce63ee0-bc69-4649-b644-34854c003845-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7ce63ee0-bc69-4649-b644-34854c003845\") " pod="openstack/ceilometer-0" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.385384 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ce63ee0-bc69-4649-b644-34854c003845-log-httpd\") pod \"ceilometer-0\" (UID: \"7ce63ee0-bc69-4649-b644-34854c003845\") " pod="openstack/ceilometer-0" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.385450 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ce63ee0-bc69-4649-b644-34854c003845-run-httpd\") pod \"ceilometer-0\" (UID: \"7ce63ee0-bc69-4649-b644-34854c003845\") " pod="openstack/ceilometer-0" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.385547 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qzjp\" (UniqueName: \"kubernetes.io/projected/338e2de1-bc7a-4224-a7c1-e90c99eba72e-kube-api-access-4qzjp\") pod \"nova-cell0-conductor-0\" (UID: \"338e2de1-bc7a-4224-a7c1-e90c99eba72e\") " pod="openstack/nova-cell0-conductor-0" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.487900 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ce63ee0-bc69-4649-b644-34854c003845-scripts\") pod \"ceilometer-0\" (UID: \"7ce63ee0-bc69-4649-b644-34854c003845\") " pod="openstack/ceilometer-0" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.488014 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68z8n\" (UniqueName: \"kubernetes.io/projected/7ce63ee0-bc69-4649-b644-34854c003845-kube-api-access-68z8n\") pod \"ceilometer-0\" (UID: \"7ce63ee0-bc69-4649-b644-34854c003845\") " pod="openstack/ceilometer-0" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.488073 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ce63ee0-bc69-4649-b644-34854c003845-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7ce63ee0-bc69-4649-b644-34854c003845\") " pod="openstack/ceilometer-0" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.488110 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ce63ee0-bc69-4649-b644-34854c003845-config-data\") pod \"ceilometer-0\" (UID: \"7ce63ee0-bc69-4649-b644-34854c003845\") " pod="openstack/ceilometer-0" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.488390 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/338e2de1-bc7a-4224-a7c1-e90c99eba72e-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"338e2de1-bc7a-4224-a7c1-e90c99eba72e\") " pod="openstack/nova-cell0-conductor-0" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.489163 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7ce63ee0-bc69-4649-b644-34854c003845-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7ce63ee0-bc69-4649-b644-34854c003845\") " pod="openstack/ceilometer-0" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.489260 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ce63ee0-bc69-4649-b644-34854c003845-log-httpd\") pod \"ceilometer-0\" (UID: \"7ce63ee0-bc69-4649-b644-34854c003845\") " pod="openstack/ceilometer-0" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.490219 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ce63ee0-bc69-4649-b644-34854c003845-log-httpd\") pod \"ceilometer-0\" (UID: \"7ce63ee0-bc69-4649-b644-34854c003845\") " pod="openstack/ceilometer-0" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.491018 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ce63ee0-bc69-4649-b644-34854c003845-run-httpd\") pod \"ceilometer-0\" (UID: \"7ce63ee0-bc69-4649-b644-34854c003845\") " pod="openstack/ceilometer-0" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.491194 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qzjp\" (UniqueName: \"kubernetes.io/projected/338e2de1-bc7a-4224-a7c1-e90c99eba72e-kube-api-access-4qzjp\") pod \"nova-cell0-conductor-0\" (UID: \"338e2de1-bc7a-4224-a7c1-e90c99eba72e\") " pod="openstack/nova-cell0-conductor-0" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.491297 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/338e2de1-bc7a-4224-a7c1-e90c99eba72e-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"338e2de1-bc7a-4224-a7c1-e90c99eba72e\") " pod="openstack/nova-cell0-conductor-0" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.492845 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ce63ee0-bc69-4649-b644-34854c003845-run-httpd\") pod \"ceilometer-0\" (UID: \"7ce63ee0-bc69-4649-b644-34854c003845\") " pod="openstack/ceilometer-0" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.494093 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ce63ee0-bc69-4649-b644-34854c003845-config-data\") pod \"ceilometer-0\" (UID: \"7ce63ee0-bc69-4649-b644-34854c003845\") " pod="openstack/ceilometer-0" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.494774 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ce63ee0-bc69-4649-b644-34854c003845-scripts\") pod \"ceilometer-0\" (UID: \"7ce63ee0-bc69-4649-b644-34854c003845\") " pod="openstack/ceilometer-0" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.495667 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/338e2de1-bc7a-4224-a7c1-e90c99eba72e-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"338e2de1-bc7a-4224-a7c1-e90c99eba72e\") " pod="openstack/nova-cell0-conductor-0" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.496235 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ce63ee0-bc69-4649-b644-34854c003845-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7ce63ee0-bc69-4649-b644-34854c003845\") " pod="openstack/ceilometer-0" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.496727 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7ce63ee0-bc69-4649-b644-34854c003845-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7ce63ee0-bc69-4649-b644-34854c003845\") " pod="openstack/ceilometer-0" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.506242 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/338e2de1-bc7a-4224-a7c1-e90c99eba72e-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"338e2de1-bc7a-4224-a7c1-e90c99eba72e\") " pod="openstack/nova-cell0-conductor-0" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.520349 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qzjp\" (UniqueName: \"kubernetes.io/projected/338e2de1-bc7a-4224-a7c1-e90c99eba72e-kube-api-access-4qzjp\") pod \"nova-cell0-conductor-0\" (UID: \"338e2de1-bc7a-4224-a7c1-e90c99eba72e\") " pod="openstack/nova-cell0-conductor-0" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.520568 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68z8n\" (UniqueName: \"kubernetes.io/projected/7ce63ee0-bc69-4649-b644-34854c003845-kube-api-access-68z8n\") pod \"ceilometer-0\" (UID: \"7ce63ee0-bc69-4649-b644-34854c003845\") " pod="openstack/ceilometer-0" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.601612 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.635638 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 27 17:23:52 crc kubenswrapper[4700]: I0227 17:23:52.994038 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5237ed58-b5d6-49e9-b111-869ead2c5c9f" path="/var/lib/kubelet/pods/5237ed58-b5d6-49e9-b111-869ead2c5c9f/volumes" Feb 27 17:23:53 crc kubenswrapper[4700]: W0227 17:23:53.117219 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ce63ee0_bc69_4649_b644_34854c003845.slice/crio-44b24c6e8002529a4461c4d4cb652bfe388826c1770a837cc27fb733c2604652 WatchSource:0}: Error finding container 44b24c6e8002529a4461c4d4cb652bfe388826c1770a837cc27fb733c2604652: Status 404 returned error can't find the container with id 44b24c6e8002529a4461c4d4cb652bfe388826c1770a837cc27fb733c2604652 Feb 27 17:23:53 crc kubenswrapper[4700]: I0227 17:23:53.118968 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:23:53 crc kubenswrapper[4700]: I0227 17:23:53.168881 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ce63ee0-bc69-4649-b644-34854c003845","Type":"ContainerStarted","Data":"44b24c6e8002529a4461c4d4cb652bfe388826c1770a837cc27fb733c2604652"} Feb 27 17:23:53 crc kubenswrapper[4700]: I0227 17:23:53.229826 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 27 17:23:54 crc kubenswrapper[4700]: I0227 17:23:54.180498 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"338e2de1-bc7a-4224-a7c1-e90c99eba72e","Type":"ContainerStarted","Data":"c6ea170a1ba6e64f7584f1b2dc9e65ab5e8bc3c08876889f9224acd46481c52c"} Feb 27 17:23:54 crc kubenswrapper[4700]: I0227 17:23:54.180922 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Feb 27 17:23:54 crc kubenswrapper[4700]: I0227 17:23:54.180937 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"338e2de1-bc7a-4224-a7c1-e90c99eba72e","Type":"ContainerStarted","Data":"60ae40a553439985def3ed68f686915ef7625be9a1057aeae99beec19cd51917"} Feb 27 17:23:54 crc kubenswrapper[4700]: I0227 17:23:54.182488 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ce63ee0-bc69-4649-b644-34854c003845","Type":"ContainerStarted","Data":"42ec230b3f62902d00c3fd43df8526b7308de78bc5499217ccae8575e86dc256"} Feb 27 17:23:54 crc kubenswrapper[4700]: I0227 17:23:54.182523 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ce63ee0-bc69-4649-b644-34854c003845","Type":"ContainerStarted","Data":"986968e0c42c3b559f64654abac2a6fc19896236cf264f9c9cf307667e85b157"} Feb 27 17:23:54 crc kubenswrapper[4700]: I0227 17:23:54.203521 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.203502625 podStartE2EDuration="2.203502625s" podCreationTimestamp="2026-02-27 17:23:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:23:54.201533793 +0000 UTC m=+1394.186846540" watchObservedRunningTime="2026-02-27 17:23:54.203502625 +0000 UTC m=+1394.188815372" Feb 27 17:23:55 crc kubenswrapper[4700]: I0227 17:23:55.197847 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ce63ee0-bc69-4649-b644-34854c003845","Type":"ContainerStarted","Data":"2ca94d3bd4e956f51cb9ac69bbc45b31e426962d22261e81abab6037c2e03fa8"} Feb 27 17:23:57 crc kubenswrapper[4700]: I0227 17:23:57.217542 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ce63ee0-bc69-4649-b644-34854c003845","Type":"ContainerStarted","Data":"35d6afc4239d670644f97c066dc5c882b65e5365dfc28b389ce805563136bf19"} Feb 27 17:23:57 crc kubenswrapper[4700]: I0227 17:23:57.218122 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 27 17:23:57 crc kubenswrapper[4700]: I0227 17:23:57.238527 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.8556846949999999 podStartE2EDuration="5.238512813s" podCreationTimestamp="2026-02-27 17:23:52 +0000 UTC" firstStartedPulling="2026-02-27 17:23:53.120254357 +0000 UTC m=+1393.105567134" lastFinishedPulling="2026-02-27 17:23:56.503082485 +0000 UTC m=+1396.488395252" observedRunningTime="2026-02-27 17:23:57.234691652 +0000 UTC m=+1397.220004409" watchObservedRunningTime="2026-02-27 17:23:57.238512813 +0000 UTC m=+1397.223825560" Feb 27 17:23:57 crc kubenswrapper[4700]: I0227 17:23:57.980651 4700 scope.go:117] "RemoveContainer" containerID="2d0da9f450ceed9ed3cba22dade279b2f7018efcb70068e35e804ea9219036d5" Feb 27 17:23:58 crc kubenswrapper[4700]: I0227 17:23:58.247147 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"0f1ae433-28e4-45cf-9635-dc6edbf16f6e","Type":"ContainerStarted","Data":"fd78392ba78e151375f967806c2c7011c5e3b2bd6832596a30adb5d8da70a937"} Feb 27 17:24:00 crc kubenswrapper[4700]: I0227 17:24:00.149332 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536884-25wnt"] Feb 27 17:24:00 crc kubenswrapper[4700]: I0227 17:24:00.152016 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536884-25wnt" Feb 27 17:24:00 crc kubenswrapper[4700]: I0227 17:24:00.155038 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:24:00 crc kubenswrapper[4700]: I0227 17:24:00.155324 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:24:00 crc kubenswrapper[4700]: I0227 17:24:00.155332 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d62kc" Feb 27 17:24:00 crc kubenswrapper[4700]: I0227 17:24:00.175346 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536884-25wnt"] Feb 27 17:24:00 crc kubenswrapper[4700]: I0227 17:24:00.284889 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmrk8\" (UniqueName: \"kubernetes.io/projected/950c16a0-025a-44bb-8ebd-1b69dea4d6b8-kube-api-access-nmrk8\") pod \"auto-csr-approver-29536884-25wnt\" (UID: \"950c16a0-025a-44bb-8ebd-1b69dea4d6b8\") " pod="openshift-infra/auto-csr-approver-29536884-25wnt" Feb 27 17:24:00 crc kubenswrapper[4700]: I0227 17:24:00.387599 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmrk8\" (UniqueName: \"kubernetes.io/projected/950c16a0-025a-44bb-8ebd-1b69dea4d6b8-kube-api-access-nmrk8\") pod \"auto-csr-approver-29536884-25wnt\" (UID: \"950c16a0-025a-44bb-8ebd-1b69dea4d6b8\") " pod="openshift-infra/auto-csr-approver-29536884-25wnt" Feb 27 17:24:00 crc kubenswrapper[4700]: I0227 17:24:00.414031 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmrk8\" (UniqueName: \"kubernetes.io/projected/950c16a0-025a-44bb-8ebd-1b69dea4d6b8-kube-api-access-nmrk8\") pod \"auto-csr-approver-29536884-25wnt\" (UID: \"950c16a0-025a-44bb-8ebd-1b69dea4d6b8\") " pod="openshift-infra/auto-csr-approver-29536884-25wnt" Feb 27 17:24:00 crc kubenswrapper[4700]: I0227 17:24:00.479391 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536884-25wnt" Feb 27 17:24:00 crc kubenswrapper[4700]: I0227 17:24:00.953721 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536884-25wnt"] Feb 27 17:24:00 crc kubenswrapper[4700]: W0227 17:24:00.955921 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod950c16a0_025a_44bb_8ebd_1b69dea4d6b8.slice/crio-45dedff16948f463947ef07cd4bed2f02c05c30117946e170a5eea996e5d8079 WatchSource:0}: Error finding container 45dedff16948f463947ef07cd4bed2f02c05c30117946e170a5eea996e5d8079: Status 404 returned error can't find the container with id 45dedff16948f463947ef07cd4bed2f02c05c30117946e170a5eea996e5d8079 Feb 27 17:24:01 crc kubenswrapper[4700]: I0227 17:24:01.275321 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536884-25wnt" event={"ID":"950c16a0-025a-44bb-8ebd-1b69dea4d6b8","Type":"ContainerStarted","Data":"45dedff16948f463947ef07cd4bed2f02c05c30117946e170a5eea996e5d8079"} Feb 27 17:24:02 crc kubenswrapper[4700]: I0227 17:24:02.308716 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536884-25wnt" event={"ID":"950c16a0-025a-44bb-8ebd-1b69dea4d6b8","Type":"ContainerStarted","Data":"dc2763324c06b0b064ffa165970464ae60e3ecfda6de3837813bddacda25008a"} Feb 27 17:24:02 crc kubenswrapper[4700]: I0227 17:24:02.333101 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536884-25wnt" podStartSLOduration=1.402812185 podStartE2EDuration="2.333080988s" podCreationTimestamp="2026-02-27 17:24:00 +0000 UTC" firstStartedPulling="2026-02-27 17:24:00.971068711 +0000 UTC m=+1400.956381498" lastFinishedPulling="2026-02-27 17:24:01.901337534 +0000 UTC m=+1401.886650301" observedRunningTime="2026-02-27 17:24:02.323992718 +0000 UTC m=+1402.309305485" watchObservedRunningTime="2026-02-27 17:24:02.333080988 +0000 UTC m=+1402.318393745" Feb 27 17:24:02 crc kubenswrapper[4700]: I0227 17:24:02.679923 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.305995 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-xc2m6"] Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.307902 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-xc2m6" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.311547 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.311612 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.324439 4700 generic.go:334] "Generic (PLEG): container finished" podID="950c16a0-025a-44bb-8ebd-1b69dea4d6b8" containerID="dc2763324c06b0b064ffa165970464ae60e3ecfda6de3837813bddacda25008a" exitCode=0 Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.324515 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536884-25wnt" event={"ID":"950c16a0-025a-44bb-8ebd-1b69dea4d6b8","Type":"ContainerDied","Data":"dc2763324c06b0b064ffa165970464ae60e3ecfda6de3837813bddacda25008a"} Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.333636 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-xc2m6"] Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.372537 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e412b441-8c49-41fe-b795-7c50d4ef09d0-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-xc2m6\" (UID: \"e412b441-8c49-41fe-b795-7c50d4ef09d0\") " pod="openstack/nova-cell0-cell-mapping-xc2m6" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.372608 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ldld\" (UniqueName: \"kubernetes.io/projected/e412b441-8c49-41fe-b795-7c50d4ef09d0-kube-api-access-7ldld\") pod \"nova-cell0-cell-mapping-xc2m6\" (UID: \"e412b441-8c49-41fe-b795-7c50d4ef09d0\") " pod="openstack/nova-cell0-cell-mapping-xc2m6" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.372689 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e412b441-8c49-41fe-b795-7c50d4ef09d0-scripts\") pod \"nova-cell0-cell-mapping-xc2m6\" (UID: \"e412b441-8c49-41fe-b795-7c50d4ef09d0\") " pod="openstack/nova-cell0-cell-mapping-xc2m6" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.372800 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e412b441-8c49-41fe-b795-7c50d4ef09d0-config-data\") pod \"nova-cell0-cell-mapping-xc2m6\" (UID: \"e412b441-8c49-41fe-b795-7c50d4ef09d0\") " pod="openstack/nova-cell0-cell-mapping-xc2m6" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.477721 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e412b441-8c49-41fe-b795-7c50d4ef09d0-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-xc2m6\" (UID: \"e412b441-8c49-41fe-b795-7c50d4ef09d0\") " pod="openstack/nova-cell0-cell-mapping-xc2m6" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.477837 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ldld\" (UniqueName: \"kubernetes.io/projected/e412b441-8c49-41fe-b795-7c50d4ef09d0-kube-api-access-7ldld\") pod \"nova-cell0-cell-mapping-xc2m6\" (UID: \"e412b441-8c49-41fe-b795-7c50d4ef09d0\") " pod="openstack/nova-cell0-cell-mapping-xc2m6" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.477925 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e412b441-8c49-41fe-b795-7c50d4ef09d0-scripts\") pod \"nova-cell0-cell-mapping-xc2m6\" (UID: \"e412b441-8c49-41fe-b795-7c50d4ef09d0\") " pod="openstack/nova-cell0-cell-mapping-xc2m6" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.477957 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e412b441-8c49-41fe-b795-7c50d4ef09d0-config-data\") pod \"nova-cell0-cell-mapping-xc2m6\" (UID: \"e412b441-8c49-41fe-b795-7c50d4ef09d0\") " pod="openstack/nova-cell0-cell-mapping-xc2m6" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.487517 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e412b441-8c49-41fe-b795-7c50d4ef09d0-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-xc2m6\" (UID: \"e412b441-8c49-41fe-b795-7c50d4ef09d0\") " pod="openstack/nova-cell0-cell-mapping-xc2m6" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.494087 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e412b441-8c49-41fe-b795-7c50d4ef09d0-config-data\") pod \"nova-cell0-cell-mapping-xc2m6\" (UID: \"e412b441-8c49-41fe-b795-7c50d4ef09d0\") " pod="openstack/nova-cell0-cell-mapping-xc2m6" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.496033 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e412b441-8c49-41fe-b795-7c50d4ef09d0-scripts\") pod \"nova-cell0-cell-mapping-xc2m6\" (UID: \"e412b441-8c49-41fe-b795-7c50d4ef09d0\") " pod="openstack/nova-cell0-cell-mapping-xc2m6" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.498056 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ldld\" (UniqueName: \"kubernetes.io/projected/e412b441-8c49-41fe-b795-7c50d4ef09d0-kube-api-access-7ldld\") pod \"nova-cell0-cell-mapping-xc2m6\" (UID: \"e412b441-8c49-41fe-b795-7c50d4ef09d0\") " pod="openstack/nova-cell0-cell-mapping-xc2m6" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.518357 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.525335 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.527283 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.531509 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.538902 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.540402 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.542351 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.548633 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.590511 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.592029 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.597336 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.642667 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.643320 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-xc2m6" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.644204 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.652572 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.657102 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.681762 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/121fbb5a-aad2-4a23-bee5-79a13beebcc7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"121fbb5a-aad2-4a23-bee5-79a13beebcc7\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.681812 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5tvj\" (UniqueName: \"kubernetes.io/projected/5eecbe2a-5069-402b-a74f-f38cf0c8b48a-kube-api-access-h5tvj\") pod \"nova-api-0\" (UID: \"5eecbe2a-5069-402b-a74f-f38cf0c8b48a\") " pod="openstack/nova-api-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.681886 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5eecbe2a-5069-402b-a74f-f38cf0c8b48a-logs\") pod \"nova-api-0\" (UID: \"5eecbe2a-5069-402b-a74f-f38cf0c8b48a\") " pod="openstack/nova-api-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.681906 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5eecbe2a-5069-402b-a74f-f38cf0c8b48a-config-data\") pod \"nova-api-0\" (UID: \"5eecbe2a-5069-402b-a74f-f38cf0c8b48a\") " pod="openstack/nova-api-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.681930 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04cedaeb-c4e6-4287-9826-29690ab72dce-config-data\") pod \"nova-scheduler-0\" (UID: \"04cedaeb-c4e6-4287-9826-29690ab72dce\") " pod="openstack/nova-scheduler-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.681965 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04cedaeb-c4e6-4287-9826-29690ab72dce-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"04cedaeb-c4e6-4287-9826-29690ab72dce\") " pod="openstack/nova-scheduler-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.681985 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/121fbb5a-aad2-4a23-bee5-79a13beebcc7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"121fbb5a-aad2-4a23-bee5-79a13beebcc7\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.682005 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfdt2\" (UniqueName: \"kubernetes.io/projected/04cedaeb-c4e6-4287-9826-29690ab72dce-kube-api-access-gfdt2\") pod \"nova-scheduler-0\" (UID: \"04cedaeb-c4e6-4287-9826-29690ab72dce\") " pod="openstack/nova-scheduler-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.682052 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqcw9\" (UniqueName: \"kubernetes.io/projected/121fbb5a-aad2-4a23-bee5-79a13beebcc7-kube-api-access-lqcw9\") pod \"nova-cell1-novncproxy-0\" (UID: \"121fbb5a-aad2-4a23-bee5-79a13beebcc7\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.682091 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5eecbe2a-5069-402b-a74f-f38cf0c8b48a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5eecbe2a-5069-402b-a74f-f38cf0c8b48a\") " pod="openstack/nova-api-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.705727 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.777519 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bb48d97cc-t2pb6"] Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.779212 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bb48d97cc-t2pb6" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.784561 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b296eaf-d6d2-4028-bfac-2573f32841c5-config-data\") pod \"nova-metadata-0\" (UID: \"4b296eaf-d6d2-4028-bfac-2573f32841c5\") " pod="openstack/nova-metadata-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.784685 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fcww\" (UniqueName: \"kubernetes.io/projected/4b296eaf-d6d2-4028-bfac-2573f32841c5-kube-api-access-4fcww\") pod \"nova-metadata-0\" (UID: \"4b296eaf-d6d2-4028-bfac-2573f32841c5\") " pod="openstack/nova-metadata-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.784804 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqcw9\" (UniqueName: \"kubernetes.io/projected/121fbb5a-aad2-4a23-bee5-79a13beebcc7-kube-api-access-lqcw9\") pod \"nova-cell1-novncproxy-0\" (UID: \"121fbb5a-aad2-4a23-bee5-79a13beebcc7\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.784905 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5eecbe2a-5069-402b-a74f-f38cf0c8b48a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5eecbe2a-5069-402b-a74f-f38cf0c8b48a\") " pod="openstack/nova-api-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.785011 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/121fbb5a-aad2-4a23-bee5-79a13beebcc7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"121fbb5a-aad2-4a23-bee5-79a13beebcc7\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.785100 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5tvj\" (UniqueName: \"kubernetes.io/projected/5eecbe2a-5069-402b-a74f-f38cf0c8b48a-kube-api-access-h5tvj\") pod \"nova-api-0\" (UID: \"5eecbe2a-5069-402b-a74f-f38cf0c8b48a\") " pod="openstack/nova-api-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.785184 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5eecbe2a-5069-402b-a74f-f38cf0c8b48a-logs\") pod \"nova-api-0\" (UID: \"5eecbe2a-5069-402b-a74f-f38cf0c8b48a\") " pod="openstack/nova-api-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.785259 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5eecbe2a-5069-402b-a74f-f38cf0c8b48a-config-data\") pod \"nova-api-0\" (UID: \"5eecbe2a-5069-402b-a74f-f38cf0c8b48a\") " pod="openstack/nova-api-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.785325 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b296eaf-d6d2-4028-bfac-2573f32841c5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4b296eaf-d6d2-4028-bfac-2573f32841c5\") " pod="openstack/nova-metadata-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.785398 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04cedaeb-c4e6-4287-9826-29690ab72dce-config-data\") pod \"nova-scheduler-0\" (UID: \"04cedaeb-c4e6-4287-9826-29690ab72dce\") " pod="openstack/nova-scheduler-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.785506 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04cedaeb-c4e6-4287-9826-29690ab72dce-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"04cedaeb-c4e6-4287-9826-29690ab72dce\") " pod="openstack/nova-scheduler-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.785594 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/121fbb5a-aad2-4a23-bee5-79a13beebcc7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"121fbb5a-aad2-4a23-bee5-79a13beebcc7\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.785667 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b296eaf-d6d2-4028-bfac-2573f32841c5-logs\") pod \"nova-metadata-0\" (UID: \"4b296eaf-d6d2-4028-bfac-2573f32841c5\") " pod="openstack/nova-metadata-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.785745 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfdt2\" (UniqueName: \"kubernetes.io/projected/04cedaeb-c4e6-4287-9826-29690ab72dce-kube-api-access-gfdt2\") pod \"nova-scheduler-0\" (UID: \"04cedaeb-c4e6-4287-9826-29690ab72dce\") " pod="openstack/nova-scheduler-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.786330 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5eecbe2a-5069-402b-a74f-f38cf0c8b48a-logs\") pod \"nova-api-0\" (UID: \"5eecbe2a-5069-402b-a74f-f38cf0c8b48a\") " pod="openstack/nova-api-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.788888 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5eecbe2a-5069-402b-a74f-f38cf0c8b48a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5eecbe2a-5069-402b-a74f-f38cf0c8b48a\") " pod="openstack/nova-api-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.789614 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/121fbb5a-aad2-4a23-bee5-79a13beebcc7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"121fbb5a-aad2-4a23-bee5-79a13beebcc7\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.790992 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5eecbe2a-5069-402b-a74f-f38cf0c8b48a-config-data\") pod \"nova-api-0\" (UID: \"5eecbe2a-5069-402b-a74f-f38cf0c8b48a\") " pod="openstack/nova-api-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.792552 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/121fbb5a-aad2-4a23-bee5-79a13beebcc7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"121fbb5a-aad2-4a23-bee5-79a13beebcc7\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.793422 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04cedaeb-c4e6-4287-9826-29690ab72dce-config-data\") pod \"nova-scheduler-0\" (UID: \"04cedaeb-c4e6-4287-9826-29690ab72dce\") " pod="openstack/nova-scheduler-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.808004 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bb48d97cc-t2pb6"] Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.815198 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04cedaeb-c4e6-4287-9826-29690ab72dce-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"04cedaeb-c4e6-4287-9826-29690ab72dce\") " pod="openstack/nova-scheduler-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.815926 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5tvj\" (UniqueName: \"kubernetes.io/projected/5eecbe2a-5069-402b-a74f-f38cf0c8b48a-kube-api-access-h5tvj\") pod \"nova-api-0\" (UID: \"5eecbe2a-5069-402b-a74f-f38cf0c8b48a\") " pod="openstack/nova-api-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.819319 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqcw9\" (UniqueName: \"kubernetes.io/projected/121fbb5a-aad2-4a23-bee5-79a13beebcc7-kube-api-access-lqcw9\") pod \"nova-cell1-novncproxy-0\" (UID: \"121fbb5a-aad2-4a23-bee5-79a13beebcc7\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.832042 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfdt2\" (UniqueName: \"kubernetes.io/projected/04cedaeb-c4e6-4287-9826-29690ab72dce-kube-api-access-gfdt2\") pod \"nova-scheduler-0\" (UID: \"04cedaeb-c4e6-4287-9826-29690ab72dce\") " pod="openstack/nova-scheduler-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.902289 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.902510 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b296eaf-d6d2-4028-bfac-2573f32841c5-logs\") pod \"nova-metadata-0\" (UID: \"4b296eaf-d6d2-4028-bfac-2573f32841c5\") " pod="openstack/nova-metadata-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.902571 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/884d838f-ded4-4ef7-92d5-7a6a1fcdecf2-config\") pod \"dnsmasq-dns-bb48d97cc-t2pb6\" (UID: \"884d838f-ded4-4ef7-92d5-7a6a1fcdecf2\") " pod="openstack/dnsmasq-dns-bb48d97cc-t2pb6" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.902597 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b296eaf-d6d2-4028-bfac-2573f32841c5-config-data\") pod \"nova-metadata-0\" (UID: \"4b296eaf-d6d2-4028-bfac-2573f32841c5\") " pod="openstack/nova-metadata-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.902613 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fcww\" (UniqueName: \"kubernetes.io/projected/4b296eaf-d6d2-4028-bfac-2573f32841c5-kube-api-access-4fcww\") pod \"nova-metadata-0\" (UID: \"4b296eaf-d6d2-4028-bfac-2573f32841c5\") " pod="openstack/nova-metadata-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.902688 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/884d838f-ded4-4ef7-92d5-7a6a1fcdecf2-ovsdbserver-sb\") pod \"dnsmasq-dns-bb48d97cc-t2pb6\" (UID: \"884d838f-ded4-4ef7-92d5-7a6a1fcdecf2\") " pod="openstack/dnsmasq-dns-bb48d97cc-t2pb6" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.902710 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/884d838f-ded4-4ef7-92d5-7a6a1fcdecf2-ovsdbserver-nb\") pod \"dnsmasq-dns-bb48d97cc-t2pb6\" (UID: \"884d838f-ded4-4ef7-92d5-7a6a1fcdecf2\") " pod="openstack/dnsmasq-dns-bb48d97cc-t2pb6" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.902730 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/884d838f-ded4-4ef7-92d5-7a6a1fcdecf2-dns-svc\") pod \"dnsmasq-dns-bb48d97cc-t2pb6\" (UID: \"884d838f-ded4-4ef7-92d5-7a6a1fcdecf2\") " pod="openstack/dnsmasq-dns-bb48d97cc-t2pb6" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.902756 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7c7j\" (UniqueName: \"kubernetes.io/projected/884d838f-ded4-4ef7-92d5-7a6a1fcdecf2-kube-api-access-c7c7j\") pod \"dnsmasq-dns-bb48d97cc-t2pb6\" (UID: \"884d838f-ded4-4ef7-92d5-7a6a1fcdecf2\") " pod="openstack/dnsmasq-dns-bb48d97cc-t2pb6" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.902797 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b296eaf-d6d2-4028-bfac-2573f32841c5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4b296eaf-d6d2-4028-bfac-2573f32841c5\") " pod="openstack/nova-metadata-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.902818 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/884d838f-ded4-4ef7-92d5-7a6a1fcdecf2-dns-swift-storage-0\") pod \"dnsmasq-dns-bb48d97cc-t2pb6\" (UID: \"884d838f-ded4-4ef7-92d5-7a6a1fcdecf2\") " pod="openstack/dnsmasq-dns-bb48d97cc-t2pb6" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.902891 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b296eaf-d6d2-4028-bfac-2573f32841c5-logs\") pod \"nova-metadata-0\" (UID: \"4b296eaf-d6d2-4028-bfac-2573f32841c5\") " pod="openstack/nova-metadata-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.912431 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b296eaf-d6d2-4028-bfac-2573f32841c5-config-data\") pod \"nova-metadata-0\" (UID: \"4b296eaf-d6d2-4028-bfac-2573f32841c5\") " pod="openstack/nova-metadata-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.916974 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.918045 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b296eaf-d6d2-4028-bfac-2573f32841c5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4b296eaf-d6d2-4028-bfac-2573f32841c5\") " pod="openstack/nova-metadata-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.928188 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fcww\" (UniqueName: \"kubernetes.io/projected/4b296eaf-d6d2-4028-bfac-2573f32841c5-kube-api-access-4fcww\") pod \"nova-metadata-0\" (UID: \"4b296eaf-d6d2-4028-bfac-2573f32841c5\") " pod="openstack/nova-metadata-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.929214 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 27 17:24:03 crc kubenswrapper[4700]: I0227 17:24:03.996922 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 17:24:04 crc kubenswrapper[4700]: I0227 17:24:04.010702 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/884d838f-ded4-4ef7-92d5-7a6a1fcdecf2-config\") pod \"dnsmasq-dns-bb48d97cc-t2pb6\" (UID: \"884d838f-ded4-4ef7-92d5-7a6a1fcdecf2\") " pod="openstack/dnsmasq-dns-bb48d97cc-t2pb6" Feb 27 17:24:04 crc kubenswrapper[4700]: I0227 17:24:04.010798 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/884d838f-ded4-4ef7-92d5-7a6a1fcdecf2-ovsdbserver-sb\") pod \"dnsmasq-dns-bb48d97cc-t2pb6\" (UID: \"884d838f-ded4-4ef7-92d5-7a6a1fcdecf2\") " pod="openstack/dnsmasq-dns-bb48d97cc-t2pb6" Feb 27 17:24:04 crc kubenswrapper[4700]: I0227 17:24:04.010820 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/884d838f-ded4-4ef7-92d5-7a6a1fcdecf2-ovsdbserver-nb\") pod \"dnsmasq-dns-bb48d97cc-t2pb6\" (UID: \"884d838f-ded4-4ef7-92d5-7a6a1fcdecf2\") " pod="openstack/dnsmasq-dns-bb48d97cc-t2pb6" Feb 27 17:24:04 crc kubenswrapper[4700]: I0227 17:24:04.010839 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/884d838f-ded4-4ef7-92d5-7a6a1fcdecf2-dns-svc\") pod \"dnsmasq-dns-bb48d97cc-t2pb6\" (UID: \"884d838f-ded4-4ef7-92d5-7a6a1fcdecf2\") " pod="openstack/dnsmasq-dns-bb48d97cc-t2pb6" Feb 27 17:24:04 crc kubenswrapper[4700]: I0227 17:24:04.010863 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7c7j\" (UniqueName: \"kubernetes.io/projected/884d838f-ded4-4ef7-92d5-7a6a1fcdecf2-kube-api-access-c7c7j\") pod \"dnsmasq-dns-bb48d97cc-t2pb6\" (UID: \"884d838f-ded4-4ef7-92d5-7a6a1fcdecf2\") " pod="openstack/dnsmasq-dns-bb48d97cc-t2pb6" Feb 27 17:24:04 crc kubenswrapper[4700]: I0227 17:24:04.010907 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/884d838f-ded4-4ef7-92d5-7a6a1fcdecf2-dns-swift-storage-0\") pod \"dnsmasq-dns-bb48d97cc-t2pb6\" (UID: \"884d838f-ded4-4ef7-92d5-7a6a1fcdecf2\") " pod="openstack/dnsmasq-dns-bb48d97cc-t2pb6" Feb 27 17:24:04 crc kubenswrapper[4700]: I0227 17:24:04.013296 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/884d838f-ded4-4ef7-92d5-7a6a1fcdecf2-config\") pod \"dnsmasq-dns-bb48d97cc-t2pb6\" (UID: \"884d838f-ded4-4ef7-92d5-7a6a1fcdecf2\") " pod="openstack/dnsmasq-dns-bb48d97cc-t2pb6" Feb 27 17:24:04 crc kubenswrapper[4700]: I0227 17:24:04.013990 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/884d838f-ded4-4ef7-92d5-7a6a1fcdecf2-dns-swift-storage-0\") pod \"dnsmasq-dns-bb48d97cc-t2pb6\" (UID: \"884d838f-ded4-4ef7-92d5-7a6a1fcdecf2\") " pod="openstack/dnsmasq-dns-bb48d97cc-t2pb6" Feb 27 17:24:04 crc kubenswrapper[4700]: I0227 17:24:04.014050 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/884d838f-ded4-4ef7-92d5-7a6a1fcdecf2-ovsdbserver-nb\") pod \"dnsmasq-dns-bb48d97cc-t2pb6\" (UID: \"884d838f-ded4-4ef7-92d5-7a6a1fcdecf2\") " pod="openstack/dnsmasq-dns-bb48d97cc-t2pb6" Feb 27 17:24:04 crc kubenswrapper[4700]: I0227 17:24:04.014528 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/884d838f-ded4-4ef7-92d5-7a6a1fcdecf2-dns-svc\") pod \"dnsmasq-dns-bb48d97cc-t2pb6\" (UID: \"884d838f-ded4-4ef7-92d5-7a6a1fcdecf2\") " pod="openstack/dnsmasq-dns-bb48d97cc-t2pb6" Feb 27 17:24:04 crc kubenswrapper[4700]: I0227 17:24:04.014665 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/884d838f-ded4-4ef7-92d5-7a6a1fcdecf2-ovsdbserver-sb\") pod \"dnsmasq-dns-bb48d97cc-t2pb6\" (UID: \"884d838f-ded4-4ef7-92d5-7a6a1fcdecf2\") " pod="openstack/dnsmasq-dns-bb48d97cc-t2pb6" Feb 27 17:24:04 crc kubenswrapper[4700]: I0227 17:24:04.112207 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7c7j\" (UniqueName: \"kubernetes.io/projected/884d838f-ded4-4ef7-92d5-7a6a1fcdecf2-kube-api-access-c7c7j\") pod \"dnsmasq-dns-bb48d97cc-t2pb6\" (UID: \"884d838f-ded4-4ef7-92d5-7a6a1fcdecf2\") " pod="openstack/dnsmasq-dns-bb48d97cc-t2pb6" Feb 27 17:24:04 crc kubenswrapper[4700]: I0227 17:24:04.170970 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bb48d97cc-t2pb6" Feb 27 17:24:04 crc kubenswrapper[4700]: I0227 17:24:04.273550 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-xc2m6"] Feb 27 17:24:04 crc kubenswrapper[4700]: I0227 17:24:04.385288 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-xc2m6" event={"ID":"e412b441-8c49-41fe-b795-7c50d4ef09d0","Type":"ContainerStarted","Data":"54a2226ca79913fded1290e53c9afbd65e0c1f311bff3241c910439c79fbb0f1"} Feb 27 17:24:04 crc kubenswrapper[4700]: I0227 17:24:04.532500 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 27 17:24:04 crc kubenswrapper[4700]: W0227 17:24:04.545735 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod121fbb5a_aad2_4a23_bee5_79a13beebcc7.slice/crio-bd475a1378478065ed09b931231a301ad986f3f7ecde3b45984e96a8e81c3c44 WatchSource:0}: Error finding container bd475a1378478065ed09b931231a301ad986f3f7ecde3b45984e96a8e81c3c44: Status 404 returned error can't find the container with id bd475a1378478065ed09b931231a301ad986f3f7ecde3b45984e96a8e81c3c44 Feb 27 17:24:04 crc kubenswrapper[4700]: I0227 17:24:04.720820 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 27 17:24:04 crc kubenswrapper[4700]: I0227 17:24:04.973177 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8djkd"] Feb 27 17:24:04 crc kubenswrapper[4700]: I0227 17:24:04.974642 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8djkd" Feb 27 17:24:04 crc kubenswrapper[4700]: I0227 17:24:04.979995 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Feb 27 17:24:04 crc kubenswrapper[4700]: I0227 17:24:04.980391 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 27 17:24:05 crc kubenswrapper[4700]: I0227 17:24:05.022531 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8djkd"] Feb 27 17:24:05 crc kubenswrapper[4700]: I0227 17:24:05.023526 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 17:24:05 crc kubenswrapper[4700]: I0227 17:24:05.040144 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 17:24:05 crc kubenswrapper[4700]: I0227 17:24:05.063715 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536884-25wnt" Feb 27 17:24:05 crc kubenswrapper[4700]: I0227 17:24:05.142266 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmrk8\" (UniqueName: \"kubernetes.io/projected/950c16a0-025a-44bb-8ebd-1b69dea4d6b8-kube-api-access-nmrk8\") pod \"950c16a0-025a-44bb-8ebd-1b69dea4d6b8\" (UID: \"950c16a0-025a-44bb-8ebd-1b69dea4d6b8\") " Feb 27 17:24:05 crc kubenswrapper[4700]: I0227 17:24:05.142607 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c828f754-e6cd-46ff-9c5b-b3db50639317-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8djkd\" (UID: \"c828f754-e6cd-46ff-9c5b-b3db50639317\") " pod="openstack/nova-cell1-conductor-db-sync-8djkd" Feb 27 17:24:05 crc kubenswrapper[4700]: I0227 17:24:05.142691 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c828f754-e6cd-46ff-9c5b-b3db50639317-config-data\") pod \"nova-cell1-conductor-db-sync-8djkd\" (UID: \"c828f754-e6cd-46ff-9c5b-b3db50639317\") " pod="openstack/nova-cell1-conductor-db-sync-8djkd" Feb 27 17:24:05 crc kubenswrapper[4700]: I0227 17:24:05.142825 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tk2c\" (UniqueName: \"kubernetes.io/projected/c828f754-e6cd-46ff-9c5b-b3db50639317-kube-api-access-2tk2c\") pod \"nova-cell1-conductor-db-sync-8djkd\" (UID: \"c828f754-e6cd-46ff-9c5b-b3db50639317\") " pod="openstack/nova-cell1-conductor-db-sync-8djkd" Feb 27 17:24:05 crc kubenswrapper[4700]: I0227 17:24:05.142874 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c828f754-e6cd-46ff-9c5b-b3db50639317-scripts\") pod \"nova-cell1-conductor-db-sync-8djkd\" (UID: \"c828f754-e6cd-46ff-9c5b-b3db50639317\") " pod="openstack/nova-cell1-conductor-db-sync-8djkd" Feb 27 17:24:05 crc kubenswrapper[4700]: I0227 17:24:05.148691 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/950c16a0-025a-44bb-8ebd-1b69dea4d6b8-kube-api-access-nmrk8" (OuterVolumeSpecName: "kube-api-access-nmrk8") pod "950c16a0-025a-44bb-8ebd-1b69dea4d6b8" (UID: "950c16a0-025a-44bb-8ebd-1b69dea4d6b8"). InnerVolumeSpecName "kube-api-access-nmrk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:24:05 crc kubenswrapper[4700]: I0227 17:24:05.151189 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bb48d97cc-t2pb6"] Feb 27 17:24:05 crc kubenswrapper[4700]: W0227 17:24:05.162317 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod884d838f_ded4_4ef7_92d5_7a6a1fcdecf2.slice/crio-95a2721b3473ea4df6ef723bbaf8aede130791e2c0e1a34de3b67bf5106ad20d WatchSource:0}: Error finding container 95a2721b3473ea4df6ef723bbaf8aede130791e2c0e1a34de3b67bf5106ad20d: Status 404 returned error can't find the container with id 95a2721b3473ea4df6ef723bbaf8aede130791e2c0e1a34de3b67bf5106ad20d Feb 27 17:24:05 crc kubenswrapper[4700]: I0227 17:24:05.245049 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tk2c\" (UniqueName: \"kubernetes.io/projected/c828f754-e6cd-46ff-9c5b-b3db50639317-kube-api-access-2tk2c\") pod \"nova-cell1-conductor-db-sync-8djkd\" (UID: \"c828f754-e6cd-46ff-9c5b-b3db50639317\") " pod="openstack/nova-cell1-conductor-db-sync-8djkd" Feb 27 17:24:05 crc kubenswrapper[4700]: I0227 17:24:05.245090 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c828f754-e6cd-46ff-9c5b-b3db50639317-scripts\") pod \"nova-cell1-conductor-db-sync-8djkd\" (UID: \"c828f754-e6cd-46ff-9c5b-b3db50639317\") " pod="openstack/nova-cell1-conductor-db-sync-8djkd" Feb 27 17:24:05 crc kubenswrapper[4700]: I0227 17:24:05.245187 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c828f754-e6cd-46ff-9c5b-b3db50639317-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8djkd\" (UID: \"c828f754-e6cd-46ff-9c5b-b3db50639317\") " pod="openstack/nova-cell1-conductor-db-sync-8djkd" Feb 27 17:24:05 crc kubenswrapper[4700]: I0227 17:24:05.245259 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c828f754-e6cd-46ff-9c5b-b3db50639317-config-data\") pod \"nova-cell1-conductor-db-sync-8djkd\" (UID: \"c828f754-e6cd-46ff-9c5b-b3db50639317\") " pod="openstack/nova-cell1-conductor-db-sync-8djkd" Feb 27 17:24:05 crc kubenswrapper[4700]: I0227 17:24:05.245332 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmrk8\" (UniqueName: \"kubernetes.io/projected/950c16a0-025a-44bb-8ebd-1b69dea4d6b8-kube-api-access-nmrk8\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:05 crc kubenswrapper[4700]: I0227 17:24:05.248638 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c828f754-e6cd-46ff-9c5b-b3db50639317-config-data\") pod \"nova-cell1-conductor-db-sync-8djkd\" (UID: \"c828f754-e6cd-46ff-9c5b-b3db50639317\") " pod="openstack/nova-cell1-conductor-db-sync-8djkd" Feb 27 17:24:05 crc kubenswrapper[4700]: I0227 17:24:05.249614 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c828f754-e6cd-46ff-9c5b-b3db50639317-scripts\") pod \"nova-cell1-conductor-db-sync-8djkd\" (UID: \"c828f754-e6cd-46ff-9c5b-b3db50639317\") " pod="openstack/nova-cell1-conductor-db-sync-8djkd" Feb 27 17:24:05 crc kubenswrapper[4700]: I0227 17:24:05.251740 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c828f754-e6cd-46ff-9c5b-b3db50639317-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8djkd\" (UID: \"c828f754-e6cd-46ff-9c5b-b3db50639317\") " pod="openstack/nova-cell1-conductor-db-sync-8djkd" Feb 27 17:24:05 crc kubenswrapper[4700]: I0227 17:24:05.262832 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tk2c\" (UniqueName: \"kubernetes.io/projected/c828f754-e6cd-46ff-9c5b-b3db50639317-kube-api-access-2tk2c\") pod \"nova-cell1-conductor-db-sync-8djkd\" (UID: \"c828f754-e6cd-46ff-9c5b-b3db50639317\") " pod="openstack/nova-cell1-conductor-db-sync-8djkd" Feb 27 17:24:05 crc kubenswrapper[4700]: I0227 17:24:05.372028 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8djkd" Feb 27 17:24:05 crc kubenswrapper[4700]: I0227 17:24:05.418178 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536878-zp9lh"] Feb 27 17:24:05 crc kubenswrapper[4700]: I0227 17:24:05.426743 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"121fbb5a-aad2-4a23-bee5-79a13beebcc7","Type":"ContainerStarted","Data":"bd475a1378478065ed09b931231a301ad986f3f7ecde3b45984e96a8e81c3c44"} Feb 27 17:24:05 crc kubenswrapper[4700]: I0227 17:24:05.430160 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536878-zp9lh"] Feb 27 17:24:05 crc kubenswrapper[4700]: I0227 17:24:05.456349 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-xc2m6" event={"ID":"e412b441-8c49-41fe-b795-7c50d4ef09d0","Type":"ContainerStarted","Data":"90785a0c53b789b80497313e888385c92d2327f8d5a03457ee7ccc9658666729"} Feb 27 17:24:05 crc kubenswrapper[4700]: I0227 17:24:05.463076 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"04cedaeb-c4e6-4287-9826-29690ab72dce","Type":"ContainerStarted","Data":"b44439856568de28db2c973f63928b3cf97368a4fa707560c8574ae3f88757c6"} Feb 27 17:24:05 crc kubenswrapper[4700]: I0227 17:24:05.465310 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4b296eaf-d6d2-4028-bfac-2573f32841c5","Type":"ContainerStarted","Data":"74d08da09a7eb33950cb58a98bd8e41e77dc92b8c82041b7024e8328f7db2d53"} Feb 27 17:24:05 crc kubenswrapper[4700]: I0227 17:24:05.469528 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5eecbe2a-5069-402b-a74f-f38cf0c8b48a","Type":"ContainerStarted","Data":"516de32a2a22bda02aa72a4101902c855b52f3d369f039e29d193e020d91969f"} Feb 27 17:24:05 crc kubenswrapper[4700]: I0227 17:24:05.478450 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bb48d97cc-t2pb6" event={"ID":"884d838f-ded4-4ef7-92d5-7a6a1fcdecf2","Type":"ContainerStarted","Data":"95a2721b3473ea4df6ef723bbaf8aede130791e2c0e1a34de3b67bf5106ad20d"} Feb 27 17:24:05 crc kubenswrapper[4700]: I0227 17:24:05.482811 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536884-25wnt" event={"ID":"950c16a0-025a-44bb-8ebd-1b69dea4d6b8","Type":"ContainerDied","Data":"45dedff16948f463947ef07cd4bed2f02c05c30117946e170a5eea996e5d8079"} Feb 27 17:24:05 crc kubenswrapper[4700]: I0227 17:24:05.482851 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536884-25wnt" Feb 27 17:24:05 crc kubenswrapper[4700]: I0227 17:24:05.482841 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45dedff16948f463947ef07cd4bed2f02c05c30117946e170a5eea996e5d8079" Feb 27 17:24:05 crc kubenswrapper[4700]: I0227 17:24:05.484792 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-xc2m6" podStartSLOduration=2.484777817 podStartE2EDuration="2.484777817s" podCreationTimestamp="2026-02-27 17:24:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:24:05.476433536 +0000 UTC m=+1405.461746283" watchObservedRunningTime="2026-02-27 17:24:05.484777817 +0000 UTC m=+1405.470090564" Feb 27 17:24:05 crc kubenswrapper[4700]: I0227 17:24:05.949220 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8djkd"] Feb 27 17:24:06 crc kubenswrapper[4700]: I0227 17:24:06.410324 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:24:06 crc kubenswrapper[4700]: I0227 17:24:06.410701 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:24:06 crc kubenswrapper[4700]: I0227 17:24:06.521111 4700 generic.go:334] "Generic (PLEG): container finished" podID="884d838f-ded4-4ef7-92d5-7a6a1fcdecf2" containerID="311929f6d9e48ad9e0020e036a90ed5783e7cbce50894a14a3efda6d9a10cc5b" exitCode=0 Feb 27 17:24:06 crc kubenswrapper[4700]: I0227 17:24:06.521201 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bb48d97cc-t2pb6" event={"ID":"884d838f-ded4-4ef7-92d5-7a6a1fcdecf2","Type":"ContainerDied","Data":"311929f6d9e48ad9e0020e036a90ed5783e7cbce50894a14a3efda6d9a10cc5b"} Feb 27 17:24:06 crc kubenswrapper[4700]: I0227 17:24:06.993096 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7765df9b-4c2e-4ac9-b862-4c300d69acb5" path="/var/lib/kubelet/pods/7765df9b-4c2e-4ac9-b862-4c300d69acb5/volumes" Feb 27 17:24:07 crc kubenswrapper[4700]: I0227 17:24:07.383049 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 17:24:07 crc kubenswrapper[4700]: I0227 17:24:07.391248 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 27 17:24:07 crc kubenswrapper[4700]: I0227 17:24:07.536104 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8djkd" event={"ID":"c828f754-e6cd-46ff-9c5b-b3db50639317","Type":"ContainerStarted","Data":"0b863820a179f4779b2fa0a6ad29853dbcb0ecc56cf0ee13dddd0531f98a960d"} Feb 27 17:24:07 crc kubenswrapper[4700]: I0227 17:24:07.857379 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Feb 27 17:24:07 crc kubenswrapper[4700]: I0227 17:24:07.941563 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-decision-engine-0" Feb 27 17:24:08 crc kubenswrapper[4700]: I0227 17:24:08.546745 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4b296eaf-d6d2-4028-bfac-2573f32841c5","Type":"ContainerStarted","Data":"3a64e23bc32ed82280a4c504ee2a7f4ceffd3f339803744211f6d83907e043c2"} Feb 27 17:24:08 crc kubenswrapper[4700]: I0227 17:24:08.547044 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4b296eaf-d6d2-4028-bfac-2573f32841c5","Type":"ContainerStarted","Data":"0d953a3f45063d782dbc5280789fecc41036ec232071b3c63c22625c59b99b55"} Feb 27 17:24:08 crc kubenswrapper[4700]: I0227 17:24:08.547127 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4b296eaf-d6d2-4028-bfac-2573f32841c5" containerName="nova-metadata-log" containerID="cri-o://0d953a3f45063d782dbc5280789fecc41036ec232071b3c63c22625c59b99b55" gracePeriod=30 Feb 27 17:24:08 crc kubenswrapper[4700]: I0227 17:24:08.547286 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4b296eaf-d6d2-4028-bfac-2573f32841c5" containerName="nova-metadata-metadata" containerID="cri-o://3a64e23bc32ed82280a4c504ee2a7f4ceffd3f339803744211f6d83907e043c2" gracePeriod=30 Feb 27 17:24:08 crc kubenswrapper[4700]: I0227 17:24:08.549916 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8djkd" event={"ID":"c828f754-e6cd-46ff-9c5b-b3db50639317","Type":"ContainerStarted","Data":"8009591760c8c226bb147d170c36329029be33c21271f32af0d70481c9d9cc6e"} Feb 27 17:24:08 crc kubenswrapper[4700]: I0227 17:24:08.552104 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"04cedaeb-c4e6-4287-9826-29690ab72dce","Type":"ContainerStarted","Data":"a760d630bf8db09e3b132243d30e1d2e128ba2de8394fde4e801ef125df76d73"} Feb 27 17:24:08 crc kubenswrapper[4700]: I0227 17:24:08.556562 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5eecbe2a-5069-402b-a74f-f38cf0c8b48a","Type":"ContainerStarted","Data":"7f5fba4bd94dc76296228ff84afffd2e2939936211b46ddee429ae242376daed"} Feb 27 17:24:08 crc kubenswrapper[4700]: I0227 17:24:08.556589 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5eecbe2a-5069-402b-a74f-f38cf0c8b48a","Type":"ContainerStarted","Data":"d6625309a4014a2de803d7a5de878f2a92784bd7a91a306dcbde3232203d1f24"} Feb 27 17:24:08 crc kubenswrapper[4700]: I0227 17:24:08.558943 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bb48d97cc-t2pb6" event={"ID":"884d838f-ded4-4ef7-92d5-7a6a1fcdecf2","Type":"ContainerStarted","Data":"a18083759d87601fd4c2f06b18c7f74979cf39abb7274d034a6781bb0f2eebc4"} Feb 27 17:24:08 crc kubenswrapper[4700]: I0227 17:24:08.559341 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bb48d97cc-t2pb6" Feb 27 17:24:08 crc kubenswrapper[4700]: I0227 17:24:08.561412 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"121fbb5a-aad2-4a23-bee5-79a13beebcc7","Type":"ContainerStarted","Data":"5332ac22fe4901ba743f6617ea0e18ccfad50da9adf328dd75043d3078428062"} Feb 27 17:24:08 crc kubenswrapper[4700]: I0227 17:24:08.561554 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Feb 27 17:24:08 crc kubenswrapper[4700]: I0227 17:24:08.561592 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="121fbb5a-aad2-4a23-bee5-79a13beebcc7" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://5332ac22fe4901ba743f6617ea0e18ccfad50da9adf328dd75043d3078428062" gracePeriod=30 Feb 27 17:24:08 crc kubenswrapper[4700]: I0227 17:24:08.592215 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.89185315 podStartE2EDuration="5.592195366s" podCreationTimestamp="2026-02-27 17:24:03 +0000 UTC" firstStartedPulling="2026-02-27 17:24:05.115367237 +0000 UTC m=+1405.100679984" lastFinishedPulling="2026-02-27 17:24:07.815709453 +0000 UTC m=+1407.801022200" observedRunningTime="2026-02-27 17:24:08.569268691 +0000 UTC m=+1408.554581428" watchObservedRunningTime="2026-02-27 17:24:08.592195366 +0000 UTC m=+1408.577508113" Feb 27 17:24:08 crc kubenswrapper[4700]: I0227 17:24:08.596549 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.5740510629999998 podStartE2EDuration="5.596531621s" podCreationTimestamp="2026-02-27 17:24:03 +0000 UTC" firstStartedPulling="2026-02-27 17:24:04.775571699 +0000 UTC m=+1404.760884446" lastFinishedPulling="2026-02-27 17:24:07.798052257 +0000 UTC m=+1407.783365004" observedRunningTime="2026-02-27 17:24:08.593706776 +0000 UTC m=+1408.579019523" watchObservedRunningTime="2026-02-27 17:24:08.596531621 +0000 UTC m=+1408.581844368" Feb 27 17:24:08 crc kubenswrapper[4700]: I0227 17:24:08.605065 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-decision-engine-0" Feb 27 17:24:08 crc kubenswrapper[4700]: I0227 17:24:08.607995 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.345379008 podStartE2EDuration="5.607977303s" podCreationTimestamp="2026-02-27 17:24:03 +0000 UTC" firstStartedPulling="2026-02-27 17:24:04.550048177 +0000 UTC m=+1404.535360924" lastFinishedPulling="2026-02-27 17:24:07.812646472 +0000 UTC m=+1407.797959219" observedRunningTime="2026-02-27 17:24:08.606196726 +0000 UTC m=+1408.591509473" watchObservedRunningTime="2026-02-27 17:24:08.607977303 +0000 UTC m=+1408.593290050" Feb 27 17:24:08 crc kubenswrapper[4700]: I0227 17:24:08.645931 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.945099755 podStartE2EDuration="5.645912574s" podCreationTimestamp="2026-02-27 17:24:03 +0000 UTC" firstStartedPulling="2026-02-27 17:24:05.114997197 +0000 UTC m=+1405.100309944" lastFinishedPulling="2026-02-27 17:24:07.815810016 +0000 UTC m=+1407.801122763" observedRunningTime="2026-02-27 17:24:08.626116162 +0000 UTC m=+1408.611428909" watchObservedRunningTime="2026-02-27 17:24:08.645912574 +0000 UTC m=+1408.631225321" Feb 27 17:24:08 crc kubenswrapper[4700]: I0227 17:24:08.651090 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bb48d97cc-t2pb6" podStartSLOduration=5.65107681 podStartE2EDuration="5.65107681s" podCreationTimestamp="2026-02-27 17:24:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:24:08.646650663 +0000 UTC m=+1408.631963400" watchObservedRunningTime="2026-02-27 17:24:08.65107681 +0000 UTC m=+1408.636389547" Feb 27 17:24:08 crc kubenswrapper[4700]: I0227 17:24:08.697502 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-8djkd" podStartSLOduration=4.697483405 podStartE2EDuration="4.697483405s" podCreationTimestamp="2026-02-27 17:24:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:24:08.695602855 +0000 UTC m=+1408.680915602" watchObservedRunningTime="2026-02-27 17:24:08.697483405 +0000 UTC m=+1408.682796152" Feb 27 17:24:08 crc kubenswrapper[4700]: I0227 17:24:08.760937 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-decision-engine-0"] Feb 27 17:24:08 crc kubenswrapper[4700]: I0227 17:24:08.903071 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 27 17:24:08 crc kubenswrapper[4700]: I0227 17:24:08.929649 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 27 17:24:08 crc kubenswrapper[4700]: I0227 17:24:08.997472 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 27 17:24:08 crc kubenswrapper[4700]: I0227 17:24:08.997507 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 27 17:24:09 crc kubenswrapper[4700]: I0227 17:24:09.570248 4700 generic.go:334] "Generic (PLEG): container finished" podID="4b296eaf-d6d2-4028-bfac-2573f32841c5" containerID="0d953a3f45063d782dbc5280789fecc41036ec232071b3c63c22625c59b99b55" exitCode=143 Feb 27 17:24:09 crc kubenswrapper[4700]: I0227 17:24:09.570395 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4b296eaf-d6d2-4028-bfac-2573f32841c5","Type":"ContainerDied","Data":"0d953a3f45063d782dbc5280789fecc41036ec232071b3c63c22625c59b99b55"} Feb 27 17:24:10 crc kubenswrapper[4700]: I0227 17:24:10.579298 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-decision-engine-0" podUID="0f1ae433-28e4-45cf-9635-dc6edbf16f6e" containerName="watcher-decision-engine" containerID="cri-o://fd78392ba78e151375f967806c2c7011c5e3b2bd6832596a30adb5d8da70a937" gracePeriod=30 Feb 27 17:24:10 crc kubenswrapper[4700]: I0227 17:24:10.788058 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-applier-0"] Feb 27 17:24:10 crc kubenswrapper[4700]: I0227 17:24:10.788323 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-applier-0" podUID="95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4" containerName="watcher-applier" containerID="cri-o://e1d9ecb97d7cf929a9aa95da731edfb7dc7ac87d203cef3af34b708aba7c33ca" gracePeriod=30 Feb 27 17:24:10 crc kubenswrapper[4700]: I0227 17:24:10.818756 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Feb 27 17:24:10 crc kubenswrapper[4700]: I0227 17:24:10.819207 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="bcc9554e-7836-4db9-aab8-d5b0bef440e1" containerName="watcher-api-log" containerID="cri-o://c224457d145e0f2fbebb305640c1cb7af141262e5054691b554ed0b3d3c26da1" gracePeriod=30 Feb 27 17:24:10 crc kubenswrapper[4700]: I0227 17:24:10.819291 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="bcc9554e-7836-4db9-aab8-d5b0bef440e1" containerName="watcher-api" containerID="cri-o://4ccfe5c870e5cdb285edb2596fa13cabf7b4950cdebee135126075a3f1383f21" gracePeriod=30 Feb 27 17:24:11 crc kubenswrapper[4700]: I0227 17:24:11.600090 4700 generic.go:334] "Generic (PLEG): container finished" podID="bcc9554e-7836-4db9-aab8-d5b0bef440e1" containerID="4ccfe5c870e5cdb285edb2596fa13cabf7b4950cdebee135126075a3f1383f21" exitCode=0 Feb 27 17:24:11 crc kubenswrapper[4700]: I0227 17:24:11.600340 4700 generic.go:334] "Generic (PLEG): container finished" podID="bcc9554e-7836-4db9-aab8-d5b0bef440e1" containerID="c224457d145e0f2fbebb305640c1cb7af141262e5054691b554ed0b3d3c26da1" exitCode=143 Feb 27 17:24:11 crc kubenswrapper[4700]: I0227 17:24:11.600183 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"bcc9554e-7836-4db9-aab8-d5b0bef440e1","Type":"ContainerDied","Data":"4ccfe5c870e5cdb285edb2596fa13cabf7b4950cdebee135126075a3f1383f21"} Feb 27 17:24:11 crc kubenswrapper[4700]: I0227 17:24:11.600376 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"bcc9554e-7836-4db9-aab8-d5b0bef440e1","Type":"ContainerDied","Data":"c224457d145e0f2fbebb305640c1cb7af141262e5054691b554ed0b3d3c26da1"} Feb 27 17:24:11 crc kubenswrapper[4700]: I0227 17:24:11.809360 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Feb 27 17:24:11 crc kubenswrapper[4700]: I0227 17:24:11.907045 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcc9554e-7836-4db9-aab8-d5b0bef440e1-public-tls-certs\") pod \"bcc9554e-7836-4db9-aab8-d5b0bef440e1\" (UID: \"bcc9554e-7836-4db9-aab8-d5b0bef440e1\") " Feb 27 17:24:11 crc kubenswrapper[4700]: I0227 17:24:11.907101 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tprg6\" (UniqueName: \"kubernetes.io/projected/bcc9554e-7836-4db9-aab8-d5b0bef440e1-kube-api-access-tprg6\") pod \"bcc9554e-7836-4db9-aab8-d5b0bef440e1\" (UID: \"bcc9554e-7836-4db9-aab8-d5b0bef440e1\") " Feb 27 17:24:11 crc kubenswrapper[4700]: I0227 17:24:11.907242 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/bcc9554e-7836-4db9-aab8-d5b0bef440e1-custom-prometheus-ca\") pod \"bcc9554e-7836-4db9-aab8-d5b0bef440e1\" (UID: \"bcc9554e-7836-4db9-aab8-d5b0bef440e1\") " Feb 27 17:24:11 crc kubenswrapper[4700]: I0227 17:24:11.907266 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bcc9554e-7836-4db9-aab8-d5b0bef440e1-logs\") pod \"bcc9554e-7836-4db9-aab8-d5b0bef440e1\" (UID: \"bcc9554e-7836-4db9-aab8-d5b0bef440e1\") " Feb 27 17:24:11 crc kubenswrapper[4700]: I0227 17:24:11.907307 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcc9554e-7836-4db9-aab8-d5b0bef440e1-combined-ca-bundle\") pod \"bcc9554e-7836-4db9-aab8-d5b0bef440e1\" (UID: \"bcc9554e-7836-4db9-aab8-d5b0bef440e1\") " Feb 27 17:24:11 crc kubenswrapper[4700]: I0227 17:24:11.907394 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcc9554e-7836-4db9-aab8-d5b0bef440e1-internal-tls-certs\") pod \"bcc9554e-7836-4db9-aab8-d5b0bef440e1\" (UID: \"bcc9554e-7836-4db9-aab8-d5b0bef440e1\") " Feb 27 17:24:11 crc kubenswrapper[4700]: I0227 17:24:11.907419 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcc9554e-7836-4db9-aab8-d5b0bef440e1-config-data\") pod \"bcc9554e-7836-4db9-aab8-d5b0bef440e1\" (UID: \"bcc9554e-7836-4db9-aab8-d5b0bef440e1\") " Feb 27 17:24:11 crc kubenswrapper[4700]: I0227 17:24:11.910081 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bcc9554e-7836-4db9-aab8-d5b0bef440e1-logs" (OuterVolumeSpecName: "logs") pod "bcc9554e-7836-4db9-aab8-d5b0bef440e1" (UID: "bcc9554e-7836-4db9-aab8-d5b0bef440e1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:24:11 crc kubenswrapper[4700]: I0227 17:24:11.931022 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcc9554e-7836-4db9-aab8-d5b0bef440e1-kube-api-access-tprg6" (OuterVolumeSpecName: "kube-api-access-tprg6") pod "bcc9554e-7836-4db9-aab8-d5b0bef440e1" (UID: "bcc9554e-7836-4db9-aab8-d5b0bef440e1"). InnerVolumeSpecName "kube-api-access-tprg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:24:11 crc kubenswrapper[4700]: I0227 17:24:11.964997 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcc9554e-7836-4db9-aab8-d5b0bef440e1-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "bcc9554e-7836-4db9-aab8-d5b0bef440e1" (UID: "bcc9554e-7836-4db9-aab8-d5b0bef440e1"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:11 crc kubenswrapper[4700]: I0227 17:24:11.993251 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcc9554e-7836-4db9-aab8-d5b0bef440e1-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "bcc9554e-7836-4db9-aab8-d5b0bef440e1" (UID: "bcc9554e-7836-4db9-aab8-d5b0bef440e1"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.009269 4700 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcc9554e-7836-4db9-aab8-d5b0bef440e1-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.009293 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tprg6\" (UniqueName: \"kubernetes.io/projected/bcc9554e-7836-4db9-aab8-d5b0bef440e1-kube-api-access-tprg6\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.009305 4700 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/bcc9554e-7836-4db9-aab8-d5b0bef440e1-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.009313 4700 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bcc9554e-7836-4db9-aab8-d5b0bef440e1-logs\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.009945 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcc9554e-7836-4db9-aab8-d5b0bef440e1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bcc9554e-7836-4db9-aab8-d5b0bef440e1" (UID: "bcc9554e-7836-4db9-aab8-d5b0bef440e1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.020292 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcc9554e-7836-4db9-aab8-d5b0bef440e1-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "bcc9554e-7836-4db9-aab8-d5b0bef440e1" (UID: "bcc9554e-7836-4db9-aab8-d5b0bef440e1"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.060760 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcc9554e-7836-4db9-aab8-d5b0bef440e1-config-data" (OuterVolumeSpecName: "config-data") pod "bcc9554e-7836-4db9-aab8-d5b0bef440e1" (UID: "bcc9554e-7836-4db9-aab8-d5b0bef440e1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.111293 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcc9554e-7836-4db9-aab8-d5b0bef440e1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.111323 4700 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcc9554e-7836-4db9-aab8-d5b0bef440e1-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.111333 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcc9554e-7836-4db9-aab8-d5b0bef440e1-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.616752 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"bcc9554e-7836-4db9-aab8-d5b0bef440e1","Type":"ContainerDied","Data":"47ee76e3ef58d41b1dea9e1705ee3548bd938013999ee9d4c980575d96c1fe21"} Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.616864 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.616905 4700 scope.go:117] "RemoveContainer" containerID="4ccfe5c870e5cdb285edb2596fa13cabf7b4950cdebee135126075a3f1383f21" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.652952 4700 scope.go:117] "RemoveContainer" containerID="c224457d145e0f2fbebb305640c1cb7af141262e5054691b554ed0b3d3c26da1" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.666931 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.693799 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-api-0"] Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.702178 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-api-0"] Feb 27 17:24:12 crc kubenswrapper[4700]: E0227 17:24:12.702660 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcc9554e-7836-4db9-aab8-d5b0bef440e1" containerName="watcher-api" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.702677 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcc9554e-7836-4db9-aab8-d5b0bef440e1" containerName="watcher-api" Feb 27 17:24:12 crc kubenswrapper[4700]: E0227 17:24:12.702707 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="950c16a0-025a-44bb-8ebd-1b69dea4d6b8" containerName="oc" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.702714 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="950c16a0-025a-44bb-8ebd-1b69dea4d6b8" containerName="oc" Feb 27 17:24:12 crc kubenswrapper[4700]: E0227 17:24:12.702732 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcc9554e-7836-4db9-aab8-d5b0bef440e1" containerName="watcher-api-log" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.702749 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcc9554e-7836-4db9-aab8-d5b0bef440e1" containerName="watcher-api-log" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.702915 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcc9554e-7836-4db9-aab8-d5b0bef440e1" containerName="watcher-api-log" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.702935 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcc9554e-7836-4db9-aab8-d5b0bef440e1" containerName="watcher-api" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.702954 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="950c16a0-025a-44bb-8ebd-1b69dea4d6b8" containerName="oc" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.704017 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.706905 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-watcher-public-svc" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.707166 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-api-config-data" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.707336 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-watcher-internal-svc" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.711931 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.746266 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36944f75-7dc8-4a1a-b20f-ab423b8282fa-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"36944f75-7dc8-4a1a-b20f-ab423b8282fa\") " pod="openstack/watcher-api-0" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.746312 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36944f75-7dc8-4a1a-b20f-ab423b8282fa-config-data\") pod \"watcher-api-0\" (UID: \"36944f75-7dc8-4a1a-b20f-ab423b8282fa\") " pod="openstack/watcher-api-0" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.746343 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msjms\" (UniqueName: \"kubernetes.io/projected/36944f75-7dc8-4a1a-b20f-ab423b8282fa-kube-api-access-msjms\") pod \"watcher-api-0\" (UID: \"36944f75-7dc8-4a1a-b20f-ab423b8282fa\") " pod="openstack/watcher-api-0" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.746503 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/36944f75-7dc8-4a1a-b20f-ab423b8282fa-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"36944f75-7dc8-4a1a-b20f-ab423b8282fa\") " pod="openstack/watcher-api-0" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.746667 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/36944f75-7dc8-4a1a-b20f-ab423b8282fa-public-tls-certs\") pod \"watcher-api-0\" (UID: \"36944f75-7dc8-4a1a-b20f-ab423b8282fa\") " pod="openstack/watcher-api-0" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.746696 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/36944f75-7dc8-4a1a-b20f-ab423b8282fa-logs\") pod \"watcher-api-0\" (UID: \"36944f75-7dc8-4a1a-b20f-ab423b8282fa\") " pod="openstack/watcher-api-0" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.746964 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/36944f75-7dc8-4a1a-b20f-ab423b8282fa-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"36944f75-7dc8-4a1a-b20f-ab423b8282fa\") " pod="openstack/watcher-api-0" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.848804 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/36944f75-7dc8-4a1a-b20f-ab423b8282fa-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"36944f75-7dc8-4a1a-b20f-ab423b8282fa\") " pod="openstack/watcher-api-0" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.848915 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36944f75-7dc8-4a1a-b20f-ab423b8282fa-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"36944f75-7dc8-4a1a-b20f-ab423b8282fa\") " pod="openstack/watcher-api-0" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.848972 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36944f75-7dc8-4a1a-b20f-ab423b8282fa-config-data\") pod \"watcher-api-0\" (UID: \"36944f75-7dc8-4a1a-b20f-ab423b8282fa\") " pod="openstack/watcher-api-0" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.849034 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msjms\" (UniqueName: \"kubernetes.io/projected/36944f75-7dc8-4a1a-b20f-ab423b8282fa-kube-api-access-msjms\") pod \"watcher-api-0\" (UID: \"36944f75-7dc8-4a1a-b20f-ab423b8282fa\") " pod="openstack/watcher-api-0" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.849115 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/36944f75-7dc8-4a1a-b20f-ab423b8282fa-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"36944f75-7dc8-4a1a-b20f-ab423b8282fa\") " pod="openstack/watcher-api-0" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.849197 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/36944f75-7dc8-4a1a-b20f-ab423b8282fa-public-tls-certs\") pod \"watcher-api-0\" (UID: \"36944f75-7dc8-4a1a-b20f-ab423b8282fa\") " pod="openstack/watcher-api-0" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.849220 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/36944f75-7dc8-4a1a-b20f-ab423b8282fa-logs\") pod \"watcher-api-0\" (UID: \"36944f75-7dc8-4a1a-b20f-ab423b8282fa\") " pod="openstack/watcher-api-0" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.850002 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/36944f75-7dc8-4a1a-b20f-ab423b8282fa-logs\") pod \"watcher-api-0\" (UID: \"36944f75-7dc8-4a1a-b20f-ab423b8282fa\") " pod="openstack/watcher-api-0" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.855293 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36944f75-7dc8-4a1a-b20f-ab423b8282fa-config-data\") pod \"watcher-api-0\" (UID: \"36944f75-7dc8-4a1a-b20f-ab423b8282fa\") " pod="openstack/watcher-api-0" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.859403 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/36944f75-7dc8-4a1a-b20f-ab423b8282fa-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"36944f75-7dc8-4a1a-b20f-ab423b8282fa\") " pod="openstack/watcher-api-0" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.862106 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36944f75-7dc8-4a1a-b20f-ab423b8282fa-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"36944f75-7dc8-4a1a-b20f-ab423b8282fa\") " pod="openstack/watcher-api-0" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.862132 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/36944f75-7dc8-4a1a-b20f-ab423b8282fa-public-tls-certs\") pod \"watcher-api-0\" (UID: \"36944f75-7dc8-4a1a-b20f-ab423b8282fa\") " pod="openstack/watcher-api-0" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.863040 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/36944f75-7dc8-4a1a-b20f-ab423b8282fa-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"36944f75-7dc8-4a1a-b20f-ab423b8282fa\") " pod="openstack/watcher-api-0" Feb 27 17:24:12 crc kubenswrapper[4700]: E0227 17:24:12.872055 4700 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e1d9ecb97d7cf929a9aa95da731edfb7dc7ac87d203cef3af34b708aba7c33ca is running failed: container process not found" containerID="e1d9ecb97d7cf929a9aa95da731edfb7dc7ac87d203cef3af34b708aba7c33ca" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Feb 27 17:24:12 crc kubenswrapper[4700]: E0227 17:24:12.872439 4700 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e1d9ecb97d7cf929a9aa95da731edfb7dc7ac87d203cef3af34b708aba7c33ca is running failed: container process not found" containerID="e1d9ecb97d7cf929a9aa95da731edfb7dc7ac87d203cef3af34b708aba7c33ca" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Feb 27 17:24:12 crc kubenswrapper[4700]: E0227 17:24:12.872820 4700 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e1d9ecb97d7cf929a9aa95da731edfb7dc7ac87d203cef3af34b708aba7c33ca is running failed: container process not found" containerID="e1d9ecb97d7cf929a9aa95da731edfb7dc7ac87d203cef3af34b708aba7c33ca" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Feb 27 17:24:12 crc kubenswrapper[4700]: E0227 17:24:12.872887 4700 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e1d9ecb97d7cf929a9aa95da731edfb7dc7ac87d203cef3af34b708aba7c33ca is running failed: container process not found" probeType="Readiness" pod="openstack/watcher-applier-0" podUID="95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4" containerName="watcher-applier" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.887957 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msjms\" (UniqueName: \"kubernetes.io/projected/36944f75-7dc8-4a1a-b20f-ab423b8282fa-kube-api-access-msjms\") pod \"watcher-api-0\" (UID: \"36944f75-7dc8-4a1a-b20f-ab423b8282fa\") " pod="openstack/watcher-api-0" Feb 27 17:24:12 crc kubenswrapper[4700]: I0227 17:24:12.918057 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.013453 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcc9554e-7836-4db9-aab8-d5b0bef440e1" path="/var/lib/kubelet/pods/bcc9554e-7836-4db9-aab8-d5b0bef440e1/volumes" Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.375189 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.460069 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4-logs\") pod \"95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4\" (UID: \"95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4\") " Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.460174 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7zsv\" (UniqueName: \"kubernetes.io/projected/95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4-kube-api-access-n7zsv\") pod \"95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4\" (UID: \"95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4\") " Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.460205 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4-config-data\") pod \"95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4\" (UID: \"95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4\") " Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.460233 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4-combined-ca-bundle\") pod \"95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4\" (UID: \"95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4\") " Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.460562 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4-logs" (OuterVolumeSpecName: "logs") pod "95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4" (UID: "95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.460876 4700 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4-logs\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.470663 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4-kube-api-access-n7zsv" (OuterVolumeSpecName: "kube-api-access-n7zsv") pod "95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4" (UID: "95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4"). InnerVolumeSpecName "kube-api-access-n7zsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.517282 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4" (UID: "95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:13 crc kubenswrapper[4700]: W0227 17:24:13.544195 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod36944f75_7dc8_4a1a_b20f_ab423b8282fa.slice/crio-ea341a614621de987158a30dae6b2f8369c224a4b7606d4b8ea539d66864f295 WatchSource:0}: Error finding container ea341a614621de987158a30dae6b2f8369c224a4b7606d4b8ea539d66864f295: Status 404 returned error can't find the container with id ea341a614621de987158a30dae6b2f8369c224a4b7606d4b8ea539d66864f295 Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.550555 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4-config-data" (OuterVolumeSpecName: "config-data") pod "95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4" (UID: "95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.559537 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.562135 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.562158 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7zsv\" (UniqueName: \"kubernetes.io/projected/95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4-kube-api-access-n7zsv\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.562168 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.632664 4700 generic.go:334] "Generic (PLEG): container finished" podID="95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4" containerID="e1d9ecb97d7cf929a9aa95da731edfb7dc7ac87d203cef3af34b708aba7c33ca" exitCode=0 Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.632950 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.632861 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4","Type":"ContainerDied","Data":"e1d9ecb97d7cf929a9aa95da731edfb7dc7ac87d203cef3af34b708aba7c33ca"} Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.633086 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4","Type":"ContainerDied","Data":"7689a23a6c5601dffd3ce204470c772926997a63d771ee14f650d21e1f2d58ea"} Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.633112 4700 scope.go:117] "RemoveContainer" containerID="e1d9ecb97d7cf929a9aa95da731edfb7dc7ac87d203cef3af34b708aba7c33ca" Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.637218 4700 generic.go:334] "Generic (PLEG): container finished" podID="e412b441-8c49-41fe-b795-7c50d4ef09d0" containerID="90785a0c53b789b80497313e888385c92d2327f8d5a03457ee7ccc9658666729" exitCode=0 Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.637308 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-xc2m6" event={"ID":"e412b441-8c49-41fe-b795-7c50d4ef09d0","Type":"ContainerDied","Data":"90785a0c53b789b80497313e888385c92d2327f8d5a03457ee7ccc9658666729"} Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.640187 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"36944f75-7dc8-4a1a-b20f-ab423b8282fa","Type":"ContainerStarted","Data":"ea341a614621de987158a30dae6b2f8369c224a4b7606d4b8ea539d66864f295"} Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.684808 4700 scope.go:117] "RemoveContainer" containerID="e1d9ecb97d7cf929a9aa95da731edfb7dc7ac87d203cef3af34b708aba7c33ca" Feb 27 17:24:13 crc kubenswrapper[4700]: E0227 17:24:13.687216 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1d9ecb97d7cf929a9aa95da731edfb7dc7ac87d203cef3af34b708aba7c33ca\": container with ID starting with e1d9ecb97d7cf929a9aa95da731edfb7dc7ac87d203cef3af34b708aba7c33ca not found: ID does not exist" containerID="e1d9ecb97d7cf929a9aa95da731edfb7dc7ac87d203cef3af34b708aba7c33ca" Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.687268 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1d9ecb97d7cf929a9aa95da731edfb7dc7ac87d203cef3af34b708aba7c33ca"} err="failed to get container status \"e1d9ecb97d7cf929a9aa95da731edfb7dc7ac87d203cef3af34b708aba7c33ca\": rpc error: code = NotFound desc = could not find container \"e1d9ecb97d7cf929a9aa95da731edfb7dc7ac87d203cef3af34b708aba7c33ca\": container with ID starting with e1d9ecb97d7cf929a9aa95da731edfb7dc7ac87d203cef3af34b708aba7c33ca not found: ID does not exist" Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.694924 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-applier-0"] Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.709255 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-applier-0"] Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.739084 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-applier-0"] Feb 27 17:24:13 crc kubenswrapper[4700]: E0227 17:24:13.740072 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4" containerName="watcher-applier" Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.740105 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4" containerName="watcher-applier" Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.740411 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4" containerName="watcher-applier" Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.743334 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.745697 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-applier-config-data" Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.751392 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.867055 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3bb7b7a-9fbd-47ed-99f5-ac0dd93ecceb-config-data\") pod \"watcher-applier-0\" (UID: \"e3bb7b7a-9fbd-47ed-99f5-ac0dd93ecceb\") " pod="openstack/watcher-applier-0" Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.867654 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3bb7b7a-9fbd-47ed-99f5-ac0dd93ecceb-logs\") pod \"watcher-applier-0\" (UID: \"e3bb7b7a-9fbd-47ed-99f5-ac0dd93ecceb\") " pod="openstack/watcher-applier-0" Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.867709 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjxnd\" (UniqueName: \"kubernetes.io/projected/e3bb7b7a-9fbd-47ed-99f5-ac0dd93ecceb-kube-api-access-hjxnd\") pod \"watcher-applier-0\" (UID: \"e3bb7b7a-9fbd-47ed-99f5-ac0dd93ecceb\") " pod="openstack/watcher-applier-0" Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.867742 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3bb7b7a-9fbd-47ed-99f5-ac0dd93ecceb-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"e3bb7b7a-9fbd-47ed-99f5-ac0dd93ecceb\") " pod="openstack/watcher-applier-0" Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.917705 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.918123 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.929662 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.960126 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.971060 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3bb7b7a-9fbd-47ed-99f5-ac0dd93ecceb-config-data\") pod \"watcher-applier-0\" (UID: \"e3bb7b7a-9fbd-47ed-99f5-ac0dd93ecceb\") " pod="openstack/watcher-applier-0" Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.971205 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3bb7b7a-9fbd-47ed-99f5-ac0dd93ecceb-logs\") pod \"watcher-applier-0\" (UID: \"e3bb7b7a-9fbd-47ed-99f5-ac0dd93ecceb\") " pod="openstack/watcher-applier-0" Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.971879 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3bb7b7a-9fbd-47ed-99f5-ac0dd93ecceb-logs\") pod \"watcher-applier-0\" (UID: \"e3bb7b7a-9fbd-47ed-99f5-ac0dd93ecceb\") " pod="openstack/watcher-applier-0" Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.972144 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjxnd\" (UniqueName: \"kubernetes.io/projected/e3bb7b7a-9fbd-47ed-99f5-ac0dd93ecceb-kube-api-access-hjxnd\") pod \"watcher-applier-0\" (UID: \"e3bb7b7a-9fbd-47ed-99f5-ac0dd93ecceb\") " pod="openstack/watcher-applier-0" Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.972208 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3bb7b7a-9fbd-47ed-99f5-ac0dd93ecceb-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"e3bb7b7a-9fbd-47ed-99f5-ac0dd93ecceb\") " pod="openstack/watcher-applier-0" Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.983166 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3bb7b7a-9fbd-47ed-99f5-ac0dd93ecceb-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"e3bb7b7a-9fbd-47ed-99f5-ac0dd93ecceb\") " pod="openstack/watcher-applier-0" Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.983166 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3bb7b7a-9fbd-47ed-99f5-ac0dd93ecceb-config-data\") pod \"watcher-applier-0\" (UID: \"e3bb7b7a-9fbd-47ed-99f5-ac0dd93ecceb\") " pod="openstack/watcher-applier-0" Feb 27 17:24:13 crc kubenswrapper[4700]: I0227 17:24:13.991078 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjxnd\" (UniqueName: \"kubernetes.io/projected/e3bb7b7a-9fbd-47ed-99f5-ac0dd93ecceb-kube-api-access-hjxnd\") pod \"watcher-applier-0\" (UID: \"e3bb7b7a-9fbd-47ed-99f5-ac0dd93ecceb\") " pod="openstack/watcher-applier-0" Feb 27 17:24:14 crc kubenswrapper[4700]: I0227 17:24:14.066087 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Feb 27 17:24:14 crc kubenswrapper[4700]: I0227 17:24:14.175923 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bb48d97cc-t2pb6" Feb 27 17:24:14 crc kubenswrapper[4700]: I0227 17:24:14.252248 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b6dff65bc-t5czk"] Feb 27 17:24:14 crc kubenswrapper[4700]: I0227 17:24:14.252481 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7b6dff65bc-t5czk" podUID="18c2be27-ec7d-4d27-a61b-8c0c5d332ebe" containerName="dnsmasq-dns" containerID="cri-o://e2a8677fd655199d727e66834fa84ee4f4ce05a81fe53e4fa222115722e3e594" gracePeriod=10 Feb 27 17:24:14 crc kubenswrapper[4700]: I0227 17:24:14.559427 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Feb 27 17:24:14 crc kubenswrapper[4700]: I0227 17:24:14.678580 4700 generic.go:334] "Generic (PLEG): container finished" podID="18c2be27-ec7d-4d27-a61b-8c0c5d332ebe" containerID="e2a8677fd655199d727e66834fa84ee4f4ce05a81fe53e4fa222115722e3e594" exitCode=0 Feb 27 17:24:14 crc kubenswrapper[4700]: I0227 17:24:14.678663 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b6dff65bc-t5czk" event={"ID":"18c2be27-ec7d-4d27-a61b-8c0c5d332ebe","Type":"ContainerDied","Data":"e2a8677fd655199d727e66834fa84ee4f4ce05a81fe53e4fa222115722e3e594"} Feb 27 17:24:14 crc kubenswrapper[4700]: I0227 17:24:14.684845 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"e3bb7b7a-9fbd-47ed-99f5-ac0dd93ecceb","Type":"ContainerStarted","Data":"03137dccb45ccc74fffb3b9245f78ac7f1e3caf74ea1e3c5a44d4b6b1c682900"} Feb 27 17:24:14 crc kubenswrapper[4700]: I0227 17:24:14.714023 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"36944f75-7dc8-4a1a-b20f-ab423b8282fa","Type":"ContainerStarted","Data":"4b7f4978951853a91c116e5bb0960550c0eb9fc634d2767de691474bed5159e7"} Feb 27 17:24:14 crc kubenswrapper[4700]: I0227 17:24:14.714082 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"36944f75-7dc8-4a1a-b20f-ab423b8282fa","Type":"ContainerStarted","Data":"794aa651a7d3e6fb3aba0cc19a527dd131949ca1e44faf7b806312cb2d5d0833"} Feb 27 17:24:14 crc kubenswrapper[4700]: I0227 17:24:14.715399 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Feb 27 17:24:14 crc kubenswrapper[4700]: I0227 17:24:14.755288 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 27 17:24:14 crc kubenswrapper[4700]: I0227 17:24:14.806615 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-api-0" podStartSLOduration=2.806593264 podStartE2EDuration="2.806593264s" podCreationTimestamp="2026-02-27 17:24:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:24:14.745202994 +0000 UTC m=+1414.730515741" watchObservedRunningTime="2026-02-27 17:24:14.806593264 +0000 UTC m=+1414.791906011" Feb 27 17:24:14 crc kubenswrapper[4700]: I0227 17:24:14.808300 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b6dff65bc-t5czk" Feb 27 17:24:14 crc kubenswrapper[4700]: I0227 17:24:14.900264 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/18c2be27-ec7d-4d27-a61b-8c0c5d332ebe-ovsdbserver-sb\") pod \"18c2be27-ec7d-4d27-a61b-8c0c5d332ebe\" (UID: \"18c2be27-ec7d-4d27-a61b-8c0c5d332ebe\") " Feb 27 17:24:14 crc kubenswrapper[4700]: I0227 17:24:14.900330 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18c2be27-ec7d-4d27-a61b-8c0c5d332ebe-config\") pod \"18c2be27-ec7d-4d27-a61b-8c0c5d332ebe\" (UID: \"18c2be27-ec7d-4d27-a61b-8c0c5d332ebe\") " Feb 27 17:24:14 crc kubenswrapper[4700]: I0227 17:24:14.900374 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kxtp\" (UniqueName: \"kubernetes.io/projected/18c2be27-ec7d-4d27-a61b-8c0c5d332ebe-kube-api-access-2kxtp\") pod \"18c2be27-ec7d-4d27-a61b-8c0c5d332ebe\" (UID: \"18c2be27-ec7d-4d27-a61b-8c0c5d332ebe\") " Feb 27 17:24:14 crc kubenswrapper[4700]: I0227 17:24:14.900405 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/18c2be27-ec7d-4d27-a61b-8c0c5d332ebe-ovsdbserver-nb\") pod \"18c2be27-ec7d-4d27-a61b-8c0c5d332ebe\" (UID: \"18c2be27-ec7d-4d27-a61b-8c0c5d332ebe\") " Feb 27 17:24:14 crc kubenswrapper[4700]: I0227 17:24:14.900497 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/18c2be27-ec7d-4d27-a61b-8c0c5d332ebe-dns-swift-storage-0\") pod \"18c2be27-ec7d-4d27-a61b-8c0c5d332ebe\" (UID: \"18c2be27-ec7d-4d27-a61b-8c0c5d332ebe\") " Feb 27 17:24:14 crc kubenswrapper[4700]: I0227 17:24:14.900545 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/18c2be27-ec7d-4d27-a61b-8c0c5d332ebe-dns-svc\") pod \"18c2be27-ec7d-4d27-a61b-8c0c5d332ebe\" (UID: \"18c2be27-ec7d-4d27-a61b-8c0c5d332ebe\") " Feb 27 17:24:14 crc kubenswrapper[4700]: I0227 17:24:14.909080 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18c2be27-ec7d-4d27-a61b-8c0c5d332ebe-kube-api-access-2kxtp" (OuterVolumeSpecName: "kube-api-access-2kxtp") pod "18c2be27-ec7d-4d27-a61b-8c0c5d332ebe" (UID: "18c2be27-ec7d-4d27-a61b-8c0c5d332ebe"). InnerVolumeSpecName "kube-api-access-2kxtp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:24:14 crc kubenswrapper[4700]: I0227 17:24:14.969037 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18c2be27-ec7d-4d27-a61b-8c0c5d332ebe-config" (OuterVolumeSpecName: "config") pod "18c2be27-ec7d-4d27-a61b-8c0c5d332ebe" (UID: "18c2be27-ec7d-4d27-a61b-8c0c5d332ebe"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:24:14 crc kubenswrapper[4700]: I0227 17:24:14.984177 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18c2be27-ec7d-4d27-a61b-8c0c5d332ebe-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "18c2be27-ec7d-4d27-a61b-8c0c5d332ebe" (UID: "18c2be27-ec7d-4d27-a61b-8c0c5d332ebe"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:24:14 crc kubenswrapper[4700]: I0227 17:24:14.994379 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4" path="/var/lib/kubelet/pods/95206c9a-7b01-4efd-bc0a-58d5b7fb5bc4/volumes" Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.002578 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18c2be27-ec7d-4d27-a61b-8c0c5d332ebe-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "18c2be27-ec7d-4d27-a61b-8c0c5d332ebe" (UID: "18c2be27-ec7d-4d27-a61b-8c0c5d332ebe"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.003259 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5eecbe2a-5069-402b-a74f-f38cf0c8b48a" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.214:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.003888 4700 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/18c2be27-ec7d-4d27-a61b-8c0c5d332ebe-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.009318 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18c2be27-ec7d-4d27-a61b-8c0c5d332ebe-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.009692 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kxtp\" (UniqueName: \"kubernetes.io/projected/18c2be27-ec7d-4d27-a61b-8c0c5d332ebe-kube-api-access-2kxtp\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.009714 4700 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/18c2be27-ec7d-4d27-a61b-8c0c5d332ebe-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.011622 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5eecbe2a-5069-402b-a74f-f38cf0c8b48a" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.214:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.013021 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18c2be27-ec7d-4d27-a61b-8c0c5d332ebe-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "18c2be27-ec7d-4d27-a61b-8c0c5d332ebe" (UID: "18c2be27-ec7d-4d27-a61b-8c0c5d332ebe"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.025030 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18c2be27-ec7d-4d27-a61b-8c0c5d332ebe-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "18c2be27-ec7d-4d27-a61b-8c0c5d332ebe" (UID: "18c2be27-ec7d-4d27-a61b-8c0c5d332ebe"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.111897 4700 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/18c2be27-ec7d-4d27-a61b-8c0c5d332ebe-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.111930 4700 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/18c2be27-ec7d-4d27-a61b-8c0c5d332ebe-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.195269 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-xc2m6" Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.213548 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e412b441-8c49-41fe-b795-7c50d4ef09d0-combined-ca-bundle\") pod \"e412b441-8c49-41fe-b795-7c50d4ef09d0\" (UID: \"e412b441-8c49-41fe-b795-7c50d4ef09d0\") " Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.213640 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ldld\" (UniqueName: \"kubernetes.io/projected/e412b441-8c49-41fe-b795-7c50d4ef09d0-kube-api-access-7ldld\") pod \"e412b441-8c49-41fe-b795-7c50d4ef09d0\" (UID: \"e412b441-8c49-41fe-b795-7c50d4ef09d0\") " Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.213723 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e412b441-8c49-41fe-b795-7c50d4ef09d0-scripts\") pod \"e412b441-8c49-41fe-b795-7c50d4ef09d0\" (UID: \"e412b441-8c49-41fe-b795-7c50d4ef09d0\") " Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.213768 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e412b441-8c49-41fe-b795-7c50d4ef09d0-config-data\") pod \"e412b441-8c49-41fe-b795-7c50d4ef09d0\" (UID: \"e412b441-8c49-41fe-b795-7c50d4ef09d0\") " Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.270453 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e412b441-8c49-41fe-b795-7c50d4ef09d0-scripts" (OuterVolumeSpecName: "scripts") pod "e412b441-8c49-41fe-b795-7c50d4ef09d0" (UID: "e412b441-8c49-41fe-b795-7c50d4ef09d0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.270751 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e412b441-8c49-41fe-b795-7c50d4ef09d0-kube-api-access-7ldld" (OuterVolumeSpecName: "kube-api-access-7ldld") pod "e412b441-8c49-41fe-b795-7c50d4ef09d0" (UID: "e412b441-8c49-41fe-b795-7c50d4ef09d0"). InnerVolumeSpecName "kube-api-access-7ldld". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.291588 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e412b441-8c49-41fe-b795-7c50d4ef09d0-config-data" (OuterVolumeSpecName: "config-data") pod "e412b441-8c49-41fe-b795-7c50d4ef09d0" (UID: "e412b441-8c49-41fe-b795-7c50d4ef09d0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.315966 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ldld\" (UniqueName: \"kubernetes.io/projected/e412b441-8c49-41fe-b795-7c50d4ef09d0-kube-api-access-7ldld\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.316215 4700 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e412b441-8c49-41fe-b795-7c50d4ef09d0-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.316225 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e412b441-8c49-41fe-b795-7c50d4ef09d0-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.318615 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e412b441-8c49-41fe-b795-7c50d4ef09d0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e412b441-8c49-41fe-b795-7c50d4ef09d0" (UID: "e412b441-8c49-41fe-b795-7c50d4ef09d0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.418254 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e412b441-8c49-41fe-b795-7c50d4ef09d0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.723656 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-xc2m6" Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.724021 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-xc2m6" event={"ID":"e412b441-8c49-41fe-b795-7c50d4ef09d0","Type":"ContainerDied","Data":"54a2226ca79913fded1290e53c9afbd65e0c1f311bff3241c910439c79fbb0f1"} Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.724059 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54a2226ca79913fded1290e53c9afbd65e0c1f311bff3241c910439c79fbb0f1" Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.726024 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b6dff65bc-t5czk" Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.726052 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b6dff65bc-t5czk" event={"ID":"18c2be27-ec7d-4d27-a61b-8c0c5d332ebe","Type":"ContainerDied","Data":"0543c218861ea5696488062d6e28b6e04cc2f07a613db947fafc831bbf8eeabb"} Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.726113 4700 scope.go:117] "RemoveContainer" containerID="e2a8677fd655199d727e66834fa84ee4f4ce05a81fe53e4fa222115722e3e594" Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.727697 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"e3bb7b7a-9fbd-47ed-99f5-ac0dd93ecceb","Type":"ContainerStarted","Data":"e6801d3742c0c59687329212f50355ec011f7f2e403970564fc06a7e9ec0f4fb"} Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.761017 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-applier-0" podStartSLOduration=2.760995403 podStartE2EDuration="2.760995403s" podCreationTimestamp="2026-02-27 17:24:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:24:15.751643256 +0000 UTC m=+1415.736956003" watchObservedRunningTime="2026-02-27 17:24:15.760995403 +0000 UTC m=+1415.746308150" Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.770269 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b6dff65bc-t5czk"] Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.778684 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7b6dff65bc-t5czk"] Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.779759 4700 scope.go:117] "RemoveContainer" containerID="7d103bbbca70d6c1b6a20a31114f7bfcfee1cce65fafad6dd2b8e21440fffdd9" Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.859764 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.860287 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5eecbe2a-5069-402b-a74f-f38cf0c8b48a" containerName="nova-api-log" containerID="cri-o://d6625309a4014a2de803d7a5de878f2a92784bd7a91a306dcbde3232203d1f24" gracePeriod=30 Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.860782 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5eecbe2a-5069-402b-a74f-f38cf0c8b48a" containerName="nova-api-api" containerID="cri-o://7f5fba4bd94dc76296228ff84afffd2e2939936211b46ddee429ae242376daed" gracePeriod=30 Feb 27 17:24:15 crc kubenswrapper[4700]: I0227 17:24:15.868731 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.400749 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.453253 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f1ae433-28e4-45cf-9635-dc6edbf16f6e-logs\") pod \"0f1ae433-28e4-45cf-9635-dc6edbf16f6e\" (UID: \"0f1ae433-28e4-45cf-9635-dc6edbf16f6e\") " Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.453363 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f1ae433-28e4-45cf-9635-dc6edbf16f6e-combined-ca-bundle\") pod \"0f1ae433-28e4-45cf-9635-dc6edbf16f6e\" (UID: \"0f1ae433-28e4-45cf-9635-dc6edbf16f6e\") " Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.453469 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ntfbz\" (UniqueName: \"kubernetes.io/projected/0f1ae433-28e4-45cf-9635-dc6edbf16f6e-kube-api-access-ntfbz\") pod \"0f1ae433-28e4-45cf-9635-dc6edbf16f6e\" (UID: \"0f1ae433-28e4-45cf-9635-dc6edbf16f6e\") " Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.453496 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/0f1ae433-28e4-45cf-9635-dc6edbf16f6e-custom-prometheus-ca\") pod \"0f1ae433-28e4-45cf-9635-dc6edbf16f6e\" (UID: \"0f1ae433-28e4-45cf-9635-dc6edbf16f6e\") " Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.453587 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f1ae433-28e4-45cf-9635-dc6edbf16f6e-config-data\") pod \"0f1ae433-28e4-45cf-9635-dc6edbf16f6e\" (UID: \"0f1ae433-28e4-45cf-9635-dc6edbf16f6e\") " Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.454110 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f1ae433-28e4-45cf-9635-dc6edbf16f6e-logs" (OuterVolumeSpecName: "logs") pod "0f1ae433-28e4-45cf-9635-dc6edbf16f6e" (UID: "0f1ae433-28e4-45cf-9635-dc6edbf16f6e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.471660 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f1ae433-28e4-45cf-9635-dc6edbf16f6e-kube-api-access-ntfbz" (OuterVolumeSpecName: "kube-api-access-ntfbz") pod "0f1ae433-28e4-45cf-9635-dc6edbf16f6e" (UID: "0f1ae433-28e4-45cf-9635-dc6edbf16f6e"). InnerVolumeSpecName "kube-api-access-ntfbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.490551 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f1ae433-28e4-45cf-9635-dc6edbf16f6e-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "0f1ae433-28e4-45cf-9635-dc6edbf16f6e" (UID: "0f1ae433-28e4-45cf-9635-dc6edbf16f6e"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.491658 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f1ae433-28e4-45cf-9635-dc6edbf16f6e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0f1ae433-28e4-45cf-9635-dc6edbf16f6e" (UID: "0f1ae433-28e4-45cf-9635-dc6edbf16f6e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.539831 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f1ae433-28e4-45cf-9635-dc6edbf16f6e-config-data" (OuterVolumeSpecName: "config-data") pod "0f1ae433-28e4-45cf-9635-dc6edbf16f6e" (UID: "0f1ae433-28e4-45cf-9635-dc6edbf16f6e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.556572 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f1ae433-28e4-45cf-9635-dc6edbf16f6e-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.556604 4700 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f1ae433-28e4-45cf-9635-dc6edbf16f6e-logs\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.556613 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f1ae433-28e4-45cf-9635-dc6edbf16f6e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.556624 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ntfbz\" (UniqueName: \"kubernetes.io/projected/0f1ae433-28e4-45cf-9635-dc6edbf16f6e-kube-api-access-ntfbz\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.556633 4700 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/0f1ae433-28e4-45cf-9635-dc6edbf16f6e-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.739386 4700 generic.go:334] "Generic (PLEG): container finished" podID="0f1ae433-28e4-45cf-9635-dc6edbf16f6e" containerID="fd78392ba78e151375f967806c2c7011c5e3b2bd6832596a30adb5d8da70a937" exitCode=0 Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.739481 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.739508 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"0f1ae433-28e4-45cf-9635-dc6edbf16f6e","Type":"ContainerDied","Data":"fd78392ba78e151375f967806c2c7011c5e3b2bd6832596a30adb5d8da70a937"} Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.739932 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"0f1ae433-28e4-45cf-9635-dc6edbf16f6e","Type":"ContainerDied","Data":"97555bdd71416290ebc1bd8a3f9aa7230fd3e06b7abf8dddadb69b7494e10c96"} Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.739953 4700 scope.go:117] "RemoveContainer" containerID="fd78392ba78e151375f967806c2c7011c5e3b2bd6832596a30adb5d8da70a937" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.749564 4700 generic.go:334] "Generic (PLEG): container finished" podID="5eecbe2a-5069-402b-a74f-f38cf0c8b48a" containerID="d6625309a4014a2de803d7a5de878f2a92784bd7a91a306dcbde3232203d1f24" exitCode=143 Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.749787 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="04cedaeb-c4e6-4287-9826-29690ab72dce" containerName="nova-scheduler-scheduler" containerID="cri-o://a760d630bf8db09e3b132243d30e1d2e128ba2de8394fde4e801ef125df76d73" gracePeriod=30 Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.749901 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5eecbe2a-5069-402b-a74f-f38cf0c8b48a","Type":"ContainerDied","Data":"d6625309a4014a2de803d7a5de878f2a92784bd7a91a306dcbde3232203d1f24"} Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.749979 4700 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.777121 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-decision-engine-0"] Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.795828 4700 scope.go:117] "RemoveContainer" containerID="2d0da9f450ceed9ed3cba22dade279b2f7018efcb70068e35e804ea9219036d5" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.802632 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-decision-engine-0"] Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.815728 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-decision-engine-0"] Feb 27 17:24:16 crc kubenswrapper[4700]: E0227 17:24:16.816472 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f1ae433-28e4-45cf-9635-dc6edbf16f6e" containerName="watcher-decision-engine" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.816497 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f1ae433-28e4-45cf-9635-dc6edbf16f6e" containerName="watcher-decision-engine" Feb 27 17:24:16 crc kubenswrapper[4700]: E0227 17:24:16.816519 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18c2be27-ec7d-4d27-a61b-8c0c5d332ebe" containerName="init" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.816530 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="18c2be27-ec7d-4d27-a61b-8c0c5d332ebe" containerName="init" Feb 27 17:24:16 crc kubenswrapper[4700]: E0227 17:24:16.816548 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f1ae433-28e4-45cf-9635-dc6edbf16f6e" containerName="watcher-decision-engine" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.816556 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f1ae433-28e4-45cf-9635-dc6edbf16f6e" containerName="watcher-decision-engine" Feb 27 17:24:16 crc kubenswrapper[4700]: E0227 17:24:16.816566 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18c2be27-ec7d-4d27-a61b-8c0c5d332ebe" containerName="dnsmasq-dns" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.816574 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="18c2be27-ec7d-4d27-a61b-8c0c5d332ebe" containerName="dnsmasq-dns" Feb 27 17:24:16 crc kubenswrapper[4700]: E0227 17:24:16.816590 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e412b441-8c49-41fe-b795-7c50d4ef09d0" containerName="nova-manage" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.816627 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="e412b441-8c49-41fe-b795-7c50d4ef09d0" containerName="nova-manage" Feb 27 17:24:16 crc kubenswrapper[4700]: E0227 17:24:16.816646 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f1ae433-28e4-45cf-9635-dc6edbf16f6e" containerName="watcher-decision-engine" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.816655 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f1ae433-28e4-45cf-9635-dc6edbf16f6e" containerName="watcher-decision-engine" Feb 27 17:24:16 crc kubenswrapper[4700]: E0227 17:24:16.816667 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f1ae433-28e4-45cf-9635-dc6edbf16f6e" containerName="watcher-decision-engine" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.816675 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f1ae433-28e4-45cf-9635-dc6edbf16f6e" containerName="watcher-decision-engine" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.817082 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="18c2be27-ec7d-4d27-a61b-8c0c5d332ebe" containerName="dnsmasq-dns" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.817140 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f1ae433-28e4-45cf-9635-dc6edbf16f6e" containerName="watcher-decision-engine" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.817154 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="e412b441-8c49-41fe-b795-7c50d4ef09d0" containerName="nova-manage" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.817165 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f1ae433-28e4-45cf-9635-dc6edbf16f6e" containerName="watcher-decision-engine" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.817178 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f1ae433-28e4-45cf-9635-dc6edbf16f6e" containerName="watcher-decision-engine" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.817215 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f1ae433-28e4-45cf-9635-dc6edbf16f6e" containerName="watcher-decision-engine" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.817227 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f1ae433-28e4-45cf-9635-dc6edbf16f6e" containerName="watcher-decision-engine" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.818128 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.821913 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-decision-engine-config-data" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.837839 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.845730 4700 scope.go:117] "RemoveContainer" containerID="fd78392ba78e151375f967806c2c7011c5e3b2bd6832596a30adb5d8da70a937" Feb 27 17:24:16 crc kubenswrapper[4700]: E0227 17:24:16.846236 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd78392ba78e151375f967806c2c7011c5e3b2bd6832596a30adb5d8da70a937\": container with ID starting with fd78392ba78e151375f967806c2c7011c5e3b2bd6832596a30adb5d8da70a937 not found: ID does not exist" containerID="fd78392ba78e151375f967806c2c7011c5e3b2bd6832596a30adb5d8da70a937" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.846286 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd78392ba78e151375f967806c2c7011c5e3b2bd6832596a30adb5d8da70a937"} err="failed to get container status \"fd78392ba78e151375f967806c2c7011c5e3b2bd6832596a30adb5d8da70a937\": rpc error: code = NotFound desc = could not find container \"fd78392ba78e151375f967806c2c7011c5e3b2bd6832596a30adb5d8da70a937\": container with ID starting with fd78392ba78e151375f967806c2c7011c5e3b2bd6832596a30adb5d8da70a937 not found: ID does not exist" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.846314 4700 scope.go:117] "RemoveContainer" containerID="2d0da9f450ceed9ed3cba22dade279b2f7018efcb70068e35e804ea9219036d5" Feb 27 17:24:16 crc kubenswrapper[4700]: E0227 17:24:16.846826 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d0da9f450ceed9ed3cba22dade279b2f7018efcb70068e35e804ea9219036d5\": container with ID starting with 2d0da9f450ceed9ed3cba22dade279b2f7018efcb70068e35e804ea9219036d5 not found: ID does not exist" containerID="2d0da9f450ceed9ed3cba22dade279b2f7018efcb70068e35e804ea9219036d5" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.846858 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d0da9f450ceed9ed3cba22dade279b2f7018efcb70068e35e804ea9219036d5"} err="failed to get container status \"2d0da9f450ceed9ed3cba22dade279b2f7018efcb70068e35e804ea9219036d5\": rpc error: code = NotFound desc = could not find container \"2d0da9f450ceed9ed3cba22dade279b2f7018efcb70068e35e804ea9219036d5\": container with ID starting with 2d0da9f450ceed9ed3cba22dade279b2f7018efcb70068e35e804ea9219036d5 not found: ID does not exist" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.862679 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7f47383-ff4b-4d9b-912a-9fe2f5ec26da-config-data\") pod \"watcher-decision-engine-0\" (UID: \"e7f47383-ff4b-4d9b-912a-9fe2f5ec26da\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.862724 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfvbm\" (UniqueName: \"kubernetes.io/projected/e7f47383-ff4b-4d9b-912a-9fe2f5ec26da-kube-api-access-lfvbm\") pod \"watcher-decision-engine-0\" (UID: \"e7f47383-ff4b-4d9b-912a-9fe2f5ec26da\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.862837 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/e7f47383-ff4b-4d9b-912a-9fe2f5ec26da-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"e7f47383-ff4b-4d9b-912a-9fe2f5ec26da\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.862905 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7f47383-ff4b-4d9b-912a-9fe2f5ec26da-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"e7f47383-ff4b-4d9b-912a-9fe2f5ec26da\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.862972 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7f47383-ff4b-4d9b-912a-9fe2f5ec26da-logs\") pod \"watcher-decision-engine-0\" (UID: \"e7f47383-ff4b-4d9b-912a-9fe2f5ec26da\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.993972 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7f47383-ff4b-4d9b-912a-9fe2f5ec26da-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"e7f47383-ff4b-4d9b-912a-9fe2f5ec26da\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.994085 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7f47383-ff4b-4d9b-912a-9fe2f5ec26da-logs\") pod \"watcher-decision-engine-0\" (UID: \"e7f47383-ff4b-4d9b-912a-9fe2f5ec26da\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.994139 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7f47383-ff4b-4d9b-912a-9fe2f5ec26da-config-data\") pod \"watcher-decision-engine-0\" (UID: \"e7f47383-ff4b-4d9b-912a-9fe2f5ec26da\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.994162 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfvbm\" (UniqueName: \"kubernetes.io/projected/e7f47383-ff4b-4d9b-912a-9fe2f5ec26da-kube-api-access-lfvbm\") pod \"watcher-decision-engine-0\" (UID: \"e7f47383-ff4b-4d9b-912a-9fe2f5ec26da\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.994337 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/e7f47383-ff4b-4d9b-912a-9fe2f5ec26da-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"e7f47383-ff4b-4d9b-912a-9fe2f5ec26da\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:24:16 crc kubenswrapper[4700]: I0227 17:24:16.994736 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7f47383-ff4b-4d9b-912a-9fe2f5ec26da-logs\") pod \"watcher-decision-engine-0\" (UID: \"e7f47383-ff4b-4d9b-912a-9fe2f5ec26da\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:24:17 crc kubenswrapper[4700]: I0227 17:24:17.004308 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/e7f47383-ff4b-4d9b-912a-9fe2f5ec26da-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"e7f47383-ff4b-4d9b-912a-9fe2f5ec26da\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:24:17 crc kubenswrapper[4700]: I0227 17:24:17.004690 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f1ae433-28e4-45cf-9635-dc6edbf16f6e" path="/var/lib/kubelet/pods/0f1ae433-28e4-45cf-9635-dc6edbf16f6e/volumes" Feb 27 17:24:17 crc kubenswrapper[4700]: I0227 17:24:17.005200 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7f47383-ff4b-4d9b-912a-9fe2f5ec26da-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"e7f47383-ff4b-4d9b-912a-9fe2f5ec26da\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:24:17 crc kubenswrapper[4700]: I0227 17:24:17.005301 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18c2be27-ec7d-4d27-a61b-8c0c5d332ebe" path="/var/lib/kubelet/pods/18c2be27-ec7d-4d27-a61b-8c0c5d332ebe/volumes" Feb 27 17:24:17 crc kubenswrapper[4700]: I0227 17:24:17.007966 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7f47383-ff4b-4d9b-912a-9fe2f5ec26da-config-data\") pod \"watcher-decision-engine-0\" (UID: \"e7f47383-ff4b-4d9b-912a-9fe2f5ec26da\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:24:17 crc kubenswrapper[4700]: I0227 17:24:17.011987 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfvbm\" (UniqueName: \"kubernetes.io/projected/e7f47383-ff4b-4d9b-912a-9fe2f5ec26da-kube-api-access-lfvbm\") pod \"watcher-decision-engine-0\" (UID: \"e7f47383-ff4b-4d9b-912a-9fe2f5ec26da\") " pod="openstack/watcher-decision-engine-0" Feb 27 17:24:17 crc kubenswrapper[4700]: I0227 17:24:17.138354 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Feb 27 17:24:17 crc kubenswrapper[4700]: I0227 17:24:17.229034 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Feb 27 17:24:17 crc kubenswrapper[4700]: I0227 17:24:17.466551 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Feb 27 17:24:17 crc kubenswrapper[4700]: I0227 17:24:17.759065 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"e7f47383-ff4b-4d9b-912a-9fe2f5ec26da","Type":"ContainerStarted","Data":"e5f1fdcfc9ae27a67c0eeb162f0a1efcec06bf3a529507abc8645c141d56201f"} Feb 27 17:24:17 crc kubenswrapper[4700]: I0227 17:24:17.759288 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"e7f47383-ff4b-4d9b-912a-9fe2f5ec26da","Type":"ContainerStarted","Data":"cb707324c463338e17b94faf35477317f97ff90f736673308858bec5336e96c9"} Feb 27 17:24:17 crc kubenswrapper[4700]: I0227 17:24:17.781401 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-decision-engine-0" podStartSLOduration=1.781379244 podStartE2EDuration="1.781379244s" podCreationTimestamp="2026-02-27 17:24:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:24:17.777196924 +0000 UTC m=+1417.762509681" watchObservedRunningTime="2026-02-27 17:24:17.781379244 +0000 UTC m=+1417.766692011" Feb 27 17:24:17 crc kubenswrapper[4700]: I0227 17:24:17.918302 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Feb 27 17:24:18 crc kubenswrapper[4700]: I0227 17:24:18.773974 4700 generic.go:334] "Generic (PLEG): container finished" podID="c828f754-e6cd-46ff-9c5b-b3db50639317" containerID="8009591760c8c226bb147d170c36329029be33c21271f32af0d70481c9d9cc6e" exitCode=0 Feb 27 17:24:18 crc kubenswrapper[4700]: I0227 17:24:18.774057 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8djkd" event={"ID":"c828f754-e6cd-46ff-9c5b-b3db50639317","Type":"ContainerDied","Data":"8009591760c8c226bb147d170c36329029be33c21271f32af0d70481c9d9cc6e"} Feb 27 17:24:18 crc kubenswrapper[4700]: E0227 17:24:18.931120 4700 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a760d630bf8db09e3b132243d30e1d2e128ba2de8394fde4e801ef125df76d73" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 27 17:24:18 crc kubenswrapper[4700]: E0227 17:24:18.934593 4700 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a760d630bf8db09e3b132243d30e1d2e128ba2de8394fde4e801ef125df76d73" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 27 17:24:18 crc kubenswrapper[4700]: E0227 17:24:18.937228 4700 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a760d630bf8db09e3b132243d30e1d2e128ba2de8394fde4e801ef125df76d73" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 27 17:24:18 crc kubenswrapper[4700]: E0227 17:24:18.937292 4700 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="04cedaeb-c4e6-4287-9826-29690ab72dce" containerName="nova-scheduler-scheduler" Feb 27 17:24:19 crc kubenswrapper[4700]: I0227 17:24:19.066921 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-applier-0" Feb 27 17:24:19 crc kubenswrapper[4700]: I0227 17:24:19.554446 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 17:24:19 crc kubenswrapper[4700]: I0227 17:24:19.649956 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5eecbe2a-5069-402b-a74f-f38cf0c8b48a-config-data\") pod \"5eecbe2a-5069-402b-a74f-f38cf0c8b48a\" (UID: \"5eecbe2a-5069-402b-a74f-f38cf0c8b48a\") " Feb 27 17:24:19 crc kubenswrapper[4700]: I0227 17:24:19.650166 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5eecbe2a-5069-402b-a74f-f38cf0c8b48a-logs\") pod \"5eecbe2a-5069-402b-a74f-f38cf0c8b48a\" (UID: \"5eecbe2a-5069-402b-a74f-f38cf0c8b48a\") " Feb 27 17:24:19 crc kubenswrapper[4700]: I0227 17:24:19.650346 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5tvj\" (UniqueName: \"kubernetes.io/projected/5eecbe2a-5069-402b-a74f-f38cf0c8b48a-kube-api-access-h5tvj\") pod \"5eecbe2a-5069-402b-a74f-f38cf0c8b48a\" (UID: \"5eecbe2a-5069-402b-a74f-f38cf0c8b48a\") " Feb 27 17:24:19 crc kubenswrapper[4700]: I0227 17:24:19.650376 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5eecbe2a-5069-402b-a74f-f38cf0c8b48a-combined-ca-bundle\") pod \"5eecbe2a-5069-402b-a74f-f38cf0c8b48a\" (UID: \"5eecbe2a-5069-402b-a74f-f38cf0c8b48a\") " Feb 27 17:24:19 crc kubenswrapper[4700]: I0227 17:24:19.650796 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5eecbe2a-5069-402b-a74f-f38cf0c8b48a-logs" (OuterVolumeSpecName: "logs") pod "5eecbe2a-5069-402b-a74f-f38cf0c8b48a" (UID: "5eecbe2a-5069-402b-a74f-f38cf0c8b48a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:24:19 crc kubenswrapper[4700]: I0227 17:24:19.656336 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5eecbe2a-5069-402b-a74f-f38cf0c8b48a-kube-api-access-h5tvj" (OuterVolumeSpecName: "kube-api-access-h5tvj") pod "5eecbe2a-5069-402b-a74f-f38cf0c8b48a" (UID: "5eecbe2a-5069-402b-a74f-f38cf0c8b48a"). InnerVolumeSpecName "kube-api-access-h5tvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:24:19 crc kubenswrapper[4700]: I0227 17:24:19.679282 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5eecbe2a-5069-402b-a74f-f38cf0c8b48a-config-data" (OuterVolumeSpecName: "config-data") pod "5eecbe2a-5069-402b-a74f-f38cf0c8b48a" (UID: "5eecbe2a-5069-402b-a74f-f38cf0c8b48a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:19 crc kubenswrapper[4700]: I0227 17:24:19.695627 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5eecbe2a-5069-402b-a74f-f38cf0c8b48a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5eecbe2a-5069-402b-a74f-f38cf0c8b48a" (UID: "5eecbe2a-5069-402b-a74f-f38cf0c8b48a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:19 crc kubenswrapper[4700]: I0227 17:24:19.752795 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5tvj\" (UniqueName: \"kubernetes.io/projected/5eecbe2a-5069-402b-a74f-f38cf0c8b48a-kube-api-access-h5tvj\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:19 crc kubenswrapper[4700]: I0227 17:24:19.752834 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5eecbe2a-5069-402b-a74f-f38cf0c8b48a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:19 crc kubenswrapper[4700]: I0227 17:24:19.752848 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5eecbe2a-5069-402b-a74f-f38cf0c8b48a-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:19 crc kubenswrapper[4700]: I0227 17:24:19.752860 4700 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5eecbe2a-5069-402b-a74f-f38cf0c8b48a-logs\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:19 crc kubenswrapper[4700]: I0227 17:24:19.797962 4700 generic.go:334] "Generic (PLEG): container finished" podID="5eecbe2a-5069-402b-a74f-f38cf0c8b48a" containerID="7f5fba4bd94dc76296228ff84afffd2e2939936211b46ddee429ae242376daed" exitCode=0 Feb 27 17:24:19 crc kubenswrapper[4700]: I0227 17:24:19.798334 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 17:24:19 crc kubenswrapper[4700]: I0227 17:24:19.798355 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5eecbe2a-5069-402b-a74f-f38cf0c8b48a","Type":"ContainerDied","Data":"7f5fba4bd94dc76296228ff84afffd2e2939936211b46ddee429ae242376daed"} Feb 27 17:24:19 crc kubenswrapper[4700]: I0227 17:24:19.798429 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5eecbe2a-5069-402b-a74f-f38cf0c8b48a","Type":"ContainerDied","Data":"516de32a2a22bda02aa72a4101902c855b52f3d369f039e29d193e020d91969f"} Feb 27 17:24:19 crc kubenswrapper[4700]: I0227 17:24:19.798449 4700 scope.go:117] "RemoveContainer" containerID="7f5fba4bd94dc76296228ff84afffd2e2939936211b46ddee429ae242376daed" Feb 27 17:24:19 crc kubenswrapper[4700]: I0227 17:24:19.851598 4700 scope.go:117] "RemoveContainer" containerID="d6625309a4014a2de803d7a5de878f2a92784bd7a91a306dcbde3232203d1f24" Feb 27 17:24:19 crc kubenswrapper[4700]: I0227 17:24:19.900328 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 27 17:24:19 crc kubenswrapper[4700]: I0227 17:24:19.914803 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 27 17:24:19 crc kubenswrapper[4700]: I0227 17:24:19.926853 4700 scope.go:117] "RemoveContainer" containerID="7f5fba4bd94dc76296228ff84afffd2e2939936211b46ddee429ae242376daed" Feb 27 17:24:19 crc kubenswrapper[4700]: E0227 17:24:19.927336 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f5fba4bd94dc76296228ff84afffd2e2939936211b46ddee429ae242376daed\": container with ID starting with 7f5fba4bd94dc76296228ff84afffd2e2939936211b46ddee429ae242376daed not found: ID does not exist" containerID="7f5fba4bd94dc76296228ff84afffd2e2939936211b46ddee429ae242376daed" Feb 27 17:24:19 crc kubenswrapper[4700]: I0227 17:24:19.927375 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f5fba4bd94dc76296228ff84afffd2e2939936211b46ddee429ae242376daed"} err="failed to get container status \"7f5fba4bd94dc76296228ff84afffd2e2939936211b46ddee429ae242376daed\": rpc error: code = NotFound desc = could not find container \"7f5fba4bd94dc76296228ff84afffd2e2939936211b46ddee429ae242376daed\": container with ID starting with 7f5fba4bd94dc76296228ff84afffd2e2939936211b46ddee429ae242376daed not found: ID does not exist" Feb 27 17:24:19 crc kubenswrapper[4700]: I0227 17:24:19.927402 4700 scope.go:117] "RemoveContainer" containerID="d6625309a4014a2de803d7a5de878f2a92784bd7a91a306dcbde3232203d1f24" Feb 27 17:24:19 crc kubenswrapper[4700]: I0227 17:24:19.927462 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 27 17:24:19 crc kubenswrapper[4700]: E0227 17:24:19.927932 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5eecbe2a-5069-402b-a74f-f38cf0c8b48a" containerName="nova-api-api" Feb 27 17:24:19 crc kubenswrapper[4700]: I0227 17:24:19.927957 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="5eecbe2a-5069-402b-a74f-f38cf0c8b48a" containerName="nova-api-api" Feb 27 17:24:19 crc kubenswrapper[4700]: E0227 17:24:19.928010 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f1ae433-28e4-45cf-9635-dc6edbf16f6e" containerName="watcher-decision-engine" Feb 27 17:24:19 crc kubenswrapper[4700]: I0227 17:24:19.928020 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f1ae433-28e4-45cf-9635-dc6edbf16f6e" containerName="watcher-decision-engine" Feb 27 17:24:19 crc kubenswrapper[4700]: E0227 17:24:19.928045 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5eecbe2a-5069-402b-a74f-f38cf0c8b48a" containerName="nova-api-log" Feb 27 17:24:19 crc kubenswrapper[4700]: I0227 17:24:19.928054 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="5eecbe2a-5069-402b-a74f-f38cf0c8b48a" containerName="nova-api-log" Feb 27 17:24:19 crc kubenswrapper[4700]: I0227 17:24:19.928315 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="5eecbe2a-5069-402b-a74f-f38cf0c8b48a" containerName="nova-api-log" Feb 27 17:24:19 crc kubenswrapper[4700]: I0227 17:24:19.928336 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="5eecbe2a-5069-402b-a74f-f38cf0c8b48a" containerName="nova-api-api" Feb 27 17:24:19 crc kubenswrapper[4700]: I0227 17:24:19.929567 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 17:24:19 crc kubenswrapper[4700]: E0227 17:24:19.930984 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6625309a4014a2de803d7a5de878f2a92784bd7a91a306dcbde3232203d1f24\": container with ID starting with d6625309a4014a2de803d7a5de878f2a92784bd7a91a306dcbde3232203d1f24 not found: ID does not exist" containerID="d6625309a4014a2de803d7a5de878f2a92784bd7a91a306dcbde3232203d1f24" Feb 27 17:24:19 crc kubenswrapper[4700]: I0227 17:24:19.931020 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6625309a4014a2de803d7a5de878f2a92784bd7a91a306dcbde3232203d1f24"} err="failed to get container status \"d6625309a4014a2de803d7a5de878f2a92784bd7a91a306dcbde3232203d1f24\": rpc error: code = NotFound desc = could not find container \"d6625309a4014a2de803d7a5de878f2a92784bd7a91a306dcbde3232203d1f24\": container with ID starting with d6625309a4014a2de803d7a5de878f2a92784bd7a91a306dcbde3232203d1f24 not found: ID does not exist" Feb 27 17:24:19 crc kubenswrapper[4700]: I0227 17:24:19.935161 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 27 17:24:19 crc kubenswrapper[4700]: I0227 17:24:19.939852 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.059335 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea091df4-f7f3-4580-b20e-215922e9bbed-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ea091df4-f7f3-4580-b20e-215922e9bbed\") " pod="openstack/nova-api-0" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.059562 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea091df4-f7f3-4580-b20e-215922e9bbed-config-data\") pod \"nova-api-0\" (UID: \"ea091df4-f7f3-4580-b20e-215922e9bbed\") " pod="openstack/nova-api-0" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.059588 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea091df4-f7f3-4580-b20e-215922e9bbed-logs\") pod \"nova-api-0\" (UID: \"ea091df4-f7f3-4580-b20e-215922e9bbed\") " pod="openstack/nova-api-0" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.059657 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtjjb\" (UniqueName: \"kubernetes.io/projected/ea091df4-f7f3-4580-b20e-215922e9bbed-kube-api-access-qtjjb\") pod \"nova-api-0\" (UID: \"ea091df4-f7f3-4580-b20e-215922e9bbed\") " pod="openstack/nova-api-0" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.161727 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea091df4-f7f3-4580-b20e-215922e9bbed-config-data\") pod \"nova-api-0\" (UID: \"ea091df4-f7f3-4580-b20e-215922e9bbed\") " pod="openstack/nova-api-0" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.161773 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea091df4-f7f3-4580-b20e-215922e9bbed-logs\") pod \"nova-api-0\" (UID: \"ea091df4-f7f3-4580-b20e-215922e9bbed\") " pod="openstack/nova-api-0" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.161843 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtjjb\" (UniqueName: \"kubernetes.io/projected/ea091df4-f7f3-4580-b20e-215922e9bbed-kube-api-access-qtjjb\") pod \"nova-api-0\" (UID: \"ea091df4-f7f3-4580-b20e-215922e9bbed\") " pod="openstack/nova-api-0" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.161869 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea091df4-f7f3-4580-b20e-215922e9bbed-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ea091df4-f7f3-4580-b20e-215922e9bbed\") " pod="openstack/nova-api-0" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.162364 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea091df4-f7f3-4580-b20e-215922e9bbed-logs\") pod \"nova-api-0\" (UID: \"ea091df4-f7f3-4580-b20e-215922e9bbed\") " pod="openstack/nova-api-0" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.169735 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea091df4-f7f3-4580-b20e-215922e9bbed-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ea091df4-f7f3-4580-b20e-215922e9bbed\") " pod="openstack/nova-api-0" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.170058 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea091df4-f7f3-4580-b20e-215922e9bbed-config-data\") pod \"nova-api-0\" (UID: \"ea091df4-f7f3-4580-b20e-215922e9bbed\") " pod="openstack/nova-api-0" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.203025 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtjjb\" (UniqueName: \"kubernetes.io/projected/ea091df4-f7f3-4580-b20e-215922e9bbed-kube-api-access-qtjjb\") pod \"nova-api-0\" (UID: \"ea091df4-f7f3-4580-b20e-215922e9bbed\") " pod="openstack/nova-api-0" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.254601 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.406571 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8djkd" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.457958 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.467864 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04cedaeb-c4e6-4287-9826-29690ab72dce-config-data\") pod \"04cedaeb-c4e6-4287-9826-29690ab72dce\" (UID: \"04cedaeb-c4e6-4287-9826-29690ab72dce\") " Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.468017 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfdt2\" (UniqueName: \"kubernetes.io/projected/04cedaeb-c4e6-4287-9826-29690ab72dce-kube-api-access-gfdt2\") pod \"04cedaeb-c4e6-4287-9826-29690ab72dce\" (UID: \"04cedaeb-c4e6-4287-9826-29690ab72dce\") " Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.468147 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c828f754-e6cd-46ff-9c5b-b3db50639317-config-data\") pod \"c828f754-e6cd-46ff-9c5b-b3db50639317\" (UID: \"c828f754-e6cd-46ff-9c5b-b3db50639317\") " Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.468266 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c828f754-e6cd-46ff-9c5b-b3db50639317-scripts\") pod \"c828f754-e6cd-46ff-9c5b-b3db50639317\" (UID: \"c828f754-e6cd-46ff-9c5b-b3db50639317\") " Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.468334 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04cedaeb-c4e6-4287-9826-29690ab72dce-combined-ca-bundle\") pod \"04cedaeb-c4e6-4287-9826-29690ab72dce\" (UID: \"04cedaeb-c4e6-4287-9826-29690ab72dce\") " Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.468363 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tk2c\" (UniqueName: \"kubernetes.io/projected/c828f754-e6cd-46ff-9c5b-b3db50639317-kube-api-access-2tk2c\") pod \"c828f754-e6cd-46ff-9c5b-b3db50639317\" (UID: \"c828f754-e6cd-46ff-9c5b-b3db50639317\") " Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.468419 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c828f754-e6cd-46ff-9c5b-b3db50639317-combined-ca-bundle\") pod \"c828f754-e6cd-46ff-9c5b-b3db50639317\" (UID: \"c828f754-e6cd-46ff-9c5b-b3db50639317\") " Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.473108 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c828f754-e6cd-46ff-9c5b-b3db50639317-scripts" (OuterVolumeSpecName: "scripts") pod "c828f754-e6cd-46ff-9c5b-b3db50639317" (UID: "c828f754-e6cd-46ff-9c5b-b3db50639317"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.478105 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04cedaeb-c4e6-4287-9826-29690ab72dce-kube-api-access-gfdt2" (OuterVolumeSpecName: "kube-api-access-gfdt2") pod "04cedaeb-c4e6-4287-9826-29690ab72dce" (UID: "04cedaeb-c4e6-4287-9826-29690ab72dce"). InnerVolumeSpecName "kube-api-access-gfdt2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.491773 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c828f754-e6cd-46ff-9c5b-b3db50639317-kube-api-access-2tk2c" (OuterVolumeSpecName: "kube-api-access-2tk2c") pod "c828f754-e6cd-46ff-9c5b-b3db50639317" (UID: "c828f754-e6cd-46ff-9c5b-b3db50639317"). InnerVolumeSpecName "kube-api-access-2tk2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.502020 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c828f754-e6cd-46ff-9c5b-b3db50639317-config-data" (OuterVolumeSpecName: "config-data") pod "c828f754-e6cd-46ff-9c5b-b3db50639317" (UID: "c828f754-e6cd-46ff-9c5b-b3db50639317"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.520722 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04cedaeb-c4e6-4287-9826-29690ab72dce-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "04cedaeb-c4e6-4287-9826-29690ab72dce" (UID: "04cedaeb-c4e6-4287-9826-29690ab72dce"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.525218 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04cedaeb-c4e6-4287-9826-29690ab72dce-config-data" (OuterVolumeSpecName: "config-data") pod "04cedaeb-c4e6-4287-9826-29690ab72dce" (UID: "04cedaeb-c4e6-4287-9826-29690ab72dce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.526862 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c828f754-e6cd-46ff-9c5b-b3db50639317-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c828f754-e6cd-46ff-9c5b-b3db50639317" (UID: "c828f754-e6cd-46ff-9c5b-b3db50639317"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.570795 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tk2c\" (UniqueName: \"kubernetes.io/projected/c828f754-e6cd-46ff-9c5b-b3db50639317-kube-api-access-2tk2c\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.570830 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c828f754-e6cd-46ff-9c5b-b3db50639317-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.570840 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04cedaeb-c4e6-4287-9826-29690ab72dce-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.570849 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfdt2\" (UniqueName: \"kubernetes.io/projected/04cedaeb-c4e6-4287-9826-29690ab72dce-kube-api-access-gfdt2\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.570858 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c828f754-e6cd-46ff-9c5b-b3db50639317-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.570868 4700 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c828f754-e6cd-46ff-9c5b-b3db50639317-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.570879 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04cedaeb-c4e6-4287-9826-29690ab72dce-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.811385 4700 generic.go:334] "Generic (PLEG): container finished" podID="04cedaeb-c4e6-4287-9826-29690ab72dce" containerID="a760d630bf8db09e3b132243d30e1d2e128ba2de8394fde4e801ef125df76d73" exitCode=0 Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.811496 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"04cedaeb-c4e6-4287-9826-29690ab72dce","Type":"ContainerDied","Data":"a760d630bf8db09e3b132243d30e1d2e128ba2de8394fde4e801ef125df76d73"} Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.811536 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"04cedaeb-c4e6-4287-9826-29690ab72dce","Type":"ContainerDied","Data":"b44439856568de28db2c973f63928b3cf97368a4fa707560c8574ae3f88757c6"} Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.811564 4700 scope.go:117] "RemoveContainer" containerID="a760d630bf8db09e3b132243d30e1d2e128ba2de8394fde4e801ef125df76d73" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.811687 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.823414 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8djkd" event={"ID":"c828f754-e6cd-46ff-9c5b-b3db50639317","Type":"ContainerDied","Data":"0b863820a179f4779b2fa0a6ad29853dbcb0ecc56cf0ee13dddd0531f98a960d"} Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.823493 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b863820a179f4779b2fa0a6ad29853dbcb0ecc56cf0ee13dddd0531f98a960d" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.823577 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8djkd" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.844776 4700 scope.go:117] "RemoveContainer" containerID="a760d630bf8db09e3b132243d30e1d2e128ba2de8394fde4e801ef125df76d73" Feb 27 17:24:20 crc kubenswrapper[4700]: E0227 17:24:20.845283 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a760d630bf8db09e3b132243d30e1d2e128ba2de8394fde4e801ef125df76d73\": container with ID starting with a760d630bf8db09e3b132243d30e1d2e128ba2de8394fde4e801ef125df76d73 not found: ID does not exist" containerID="a760d630bf8db09e3b132243d30e1d2e128ba2de8394fde4e801ef125df76d73" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.845343 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a760d630bf8db09e3b132243d30e1d2e128ba2de8394fde4e801ef125df76d73"} err="failed to get container status \"a760d630bf8db09e3b132243d30e1d2e128ba2de8394fde4e801ef125df76d73\": rpc error: code = NotFound desc = could not find container \"a760d630bf8db09e3b132243d30e1d2e128ba2de8394fde4e801ef125df76d73\": container with ID starting with a760d630bf8db09e3b132243d30e1d2e128ba2de8394fde4e801ef125df76d73 not found: ID does not exist" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.846713 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.889826 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.905548 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.915826 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 17:24:20 crc kubenswrapper[4700]: E0227 17:24:20.922363 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c828f754-e6cd-46ff-9c5b-b3db50639317" containerName="nova-cell1-conductor-db-sync" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.922406 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="c828f754-e6cd-46ff-9c5b-b3db50639317" containerName="nova-cell1-conductor-db-sync" Feb 27 17:24:20 crc kubenswrapper[4700]: E0227 17:24:20.922457 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04cedaeb-c4e6-4287-9826-29690ab72dce" containerName="nova-scheduler-scheduler" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.922468 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="04cedaeb-c4e6-4287-9826-29690ab72dce" containerName="nova-scheduler-scheduler" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.922802 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="c828f754-e6cd-46ff-9c5b-b3db50639317" containerName="nova-cell1-conductor-db-sync" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.922832 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="04cedaeb-c4e6-4287-9826-29690ab72dce" containerName="nova-scheduler-scheduler" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.923908 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.927110 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.935616 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.937550 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.944921 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.956143 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.969807 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.979982 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7602b1c6-8920-4dfd-a379-a008447edb85-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"7602b1c6-8920-4dfd-a379-a008447edb85\") " pod="openstack/nova-cell1-conductor-0" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.980055 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7602b1c6-8920-4dfd-a379-a008447edb85-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"7602b1c6-8920-4dfd-a379-a008447edb85\") " pod="openstack/nova-cell1-conductor-0" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.980084 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd4be9bd-89e3-4e6c-b5a9-6ced858cb974-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"fd4be9bd-89e3-4e6c-b5a9-6ced858cb974\") " pod="openstack/nova-scheduler-0" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.980364 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpwtf\" (UniqueName: \"kubernetes.io/projected/fd4be9bd-89e3-4e6c-b5a9-6ced858cb974-kube-api-access-kpwtf\") pod \"nova-scheduler-0\" (UID: \"fd4be9bd-89e3-4e6c-b5a9-6ced858cb974\") " pod="openstack/nova-scheduler-0" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.980639 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vrm7\" (UniqueName: \"kubernetes.io/projected/7602b1c6-8920-4dfd-a379-a008447edb85-kube-api-access-9vrm7\") pod \"nova-cell1-conductor-0\" (UID: \"7602b1c6-8920-4dfd-a379-a008447edb85\") " pod="openstack/nova-cell1-conductor-0" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.980794 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd4be9bd-89e3-4e6c-b5a9-6ced858cb974-config-data\") pod \"nova-scheduler-0\" (UID: \"fd4be9bd-89e3-4e6c-b5a9-6ced858cb974\") " pod="openstack/nova-scheduler-0" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.997866 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04cedaeb-c4e6-4287-9826-29690ab72dce" path="/var/lib/kubelet/pods/04cedaeb-c4e6-4287-9826-29690ab72dce/volumes" Feb 27 17:24:20 crc kubenswrapper[4700]: I0227 17:24:20.998647 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5eecbe2a-5069-402b-a74f-f38cf0c8b48a" path="/var/lib/kubelet/pods/5eecbe2a-5069-402b-a74f-f38cf0c8b48a/volumes" Feb 27 17:24:21 crc kubenswrapper[4700]: I0227 17:24:21.082703 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd4be9bd-89e3-4e6c-b5a9-6ced858cb974-config-data\") pod \"nova-scheduler-0\" (UID: \"fd4be9bd-89e3-4e6c-b5a9-6ced858cb974\") " pod="openstack/nova-scheduler-0" Feb 27 17:24:21 crc kubenswrapper[4700]: I0227 17:24:21.082772 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7602b1c6-8920-4dfd-a379-a008447edb85-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"7602b1c6-8920-4dfd-a379-a008447edb85\") " pod="openstack/nova-cell1-conductor-0" Feb 27 17:24:21 crc kubenswrapper[4700]: I0227 17:24:21.082826 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7602b1c6-8920-4dfd-a379-a008447edb85-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"7602b1c6-8920-4dfd-a379-a008447edb85\") " pod="openstack/nova-cell1-conductor-0" Feb 27 17:24:21 crc kubenswrapper[4700]: I0227 17:24:21.082852 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd4be9bd-89e3-4e6c-b5a9-6ced858cb974-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"fd4be9bd-89e3-4e6c-b5a9-6ced858cb974\") " pod="openstack/nova-scheduler-0" Feb 27 17:24:21 crc kubenswrapper[4700]: I0227 17:24:21.082923 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpwtf\" (UniqueName: \"kubernetes.io/projected/fd4be9bd-89e3-4e6c-b5a9-6ced858cb974-kube-api-access-kpwtf\") pod \"nova-scheduler-0\" (UID: \"fd4be9bd-89e3-4e6c-b5a9-6ced858cb974\") " pod="openstack/nova-scheduler-0" Feb 27 17:24:21 crc kubenswrapper[4700]: I0227 17:24:21.082946 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vrm7\" (UniqueName: \"kubernetes.io/projected/7602b1c6-8920-4dfd-a379-a008447edb85-kube-api-access-9vrm7\") pod \"nova-cell1-conductor-0\" (UID: \"7602b1c6-8920-4dfd-a379-a008447edb85\") " pod="openstack/nova-cell1-conductor-0" Feb 27 17:24:21 crc kubenswrapper[4700]: I0227 17:24:21.092030 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7602b1c6-8920-4dfd-a379-a008447edb85-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"7602b1c6-8920-4dfd-a379-a008447edb85\") " pod="openstack/nova-cell1-conductor-0" Feb 27 17:24:21 crc kubenswrapper[4700]: I0227 17:24:21.095014 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd4be9bd-89e3-4e6c-b5a9-6ced858cb974-config-data\") pod \"nova-scheduler-0\" (UID: \"fd4be9bd-89e3-4e6c-b5a9-6ced858cb974\") " pod="openstack/nova-scheduler-0" Feb 27 17:24:21 crc kubenswrapper[4700]: I0227 17:24:21.097842 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd4be9bd-89e3-4e6c-b5a9-6ced858cb974-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"fd4be9bd-89e3-4e6c-b5a9-6ced858cb974\") " pod="openstack/nova-scheduler-0" Feb 27 17:24:21 crc kubenswrapper[4700]: I0227 17:24:21.098110 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7602b1c6-8920-4dfd-a379-a008447edb85-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"7602b1c6-8920-4dfd-a379-a008447edb85\") " pod="openstack/nova-cell1-conductor-0" Feb 27 17:24:21 crc kubenswrapper[4700]: I0227 17:24:21.101863 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vrm7\" (UniqueName: \"kubernetes.io/projected/7602b1c6-8920-4dfd-a379-a008447edb85-kube-api-access-9vrm7\") pod \"nova-cell1-conductor-0\" (UID: \"7602b1c6-8920-4dfd-a379-a008447edb85\") " pod="openstack/nova-cell1-conductor-0" Feb 27 17:24:21 crc kubenswrapper[4700]: I0227 17:24:21.105038 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpwtf\" (UniqueName: \"kubernetes.io/projected/fd4be9bd-89e3-4e6c-b5a9-6ced858cb974-kube-api-access-kpwtf\") pod \"nova-scheduler-0\" (UID: \"fd4be9bd-89e3-4e6c-b5a9-6ced858cb974\") " pod="openstack/nova-scheduler-0" Feb 27 17:24:21 crc kubenswrapper[4700]: I0227 17:24:21.273971 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 27 17:24:21 crc kubenswrapper[4700]: I0227 17:24:21.288014 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 27 17:24:21 crc kubenswrapper[4700]: W0227 17:24:21.773573 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd4be9bd_89e3_4e6c_b5a9_6ced858cb974.slice/crio-6c8c2b139534e5bb14ca0c8bd2dee47022a7332b82a8b4775b8ec4a421ad39ee WatchSource:0}: Error finding container 6c8c2b139534e5bb14ca0c8bd2dee47022a7332b82a8b4775b8ec4a421ad39ee: Status 404 returned error can't find the container with id 6c8c2b139534e5bb14ca0c8bd2dee47022a7332b82a8b4775b8ec4a421ad39ee Feb 27 17:24:21 crc kubenswrapper[4700]: I0227 17:24:21.780674 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 17:24:21 crc kubenswrapper[4700]: I0227 17:24:21.833579 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"fd4be9bd-89e3-4e6c-b5a9-6ced858cb974","Type":"ContainerStarted","Data":"6c8c2b139534e5bb14ca0c8bd2dee47022a7332b82a8b4775b8ec4a421ad39ee"} Feb 27 17:24:21 crc kubenswrapper[4700]: I0227 17:24:21.836962 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ea091df4-f7f3-4580-b20e-215922e9bbed","Type":"ContainerStarted","Data":"9abc4287de101622f56dcf027eadf4227af7525ef8bffdcd76e13612ac66591f"} Feb 27 17:24:21 crc kubenswrapper[4700]: I0227 17:24:21.837012 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ea091df4-f7f3-4580-b20e-215922e9bbed","Type":"ContainerStarted","Data":"2e18d66beced46445840352f57cc57d4534058c5afafc4fa9a24a0845342f796"} Feb 27 17:24:21 crc kubenswrapper[4700]: I0227 17:24:21.837027 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ea091df4-f7f3-4580-b20e-215922e9bbed","Type":"ContainerStarted","Data":"8ac15a89c5fa8e552ce59e15666209fdebb0576c8b016e8810bf5725e84fc019"} Feb 27 17:24:21 crc kubenswrapper[4700]: I0227 17:24:21.866563 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.866540838 podStartE2EDuration="2.866540838s" podCreationTimestamp="2026-02-27 17:24:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:24:21.854995723 +0000 UTC m=+1421.840308470" watchObservedRunningTime="2026-02-27 17:24:21.866540838 +0000 UTC m=+1421.851853585" Feb 27 17:24:21 crc kubenswrapper[4700]: I0227 17:24:21.877960 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 27 17:24:21 crc kubenswrapper[4700]: W0227 17:24:21.881492 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7602b1c6_8920_4dfd_a379_a008447edb85.slice/crio-1bbdf6b92c8127c1bc6fe8115cf5357063a3fb868c69a0e9dd230795a9f1d709 WatchSource:0}: Error finding container 1bbdf6b92c8127c1bc6fe8115cf5357063a3fb868c69a0e9dd230795a9f1d709: Status 404 returned error can't find the container with id 1bbdf6b92c8127c1bc6fe8115cf5357063a3fb868c69a0e9dd230795a9f1d709 Feb 27 17:24:22 crc kubenswrapper[4700]: I0227 17:24:22.613782 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 27 17:24:22 crc kubenswrapper[4700]: I0227 17:24:22.849680 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"fd4be9bd-89e3-4e6c-b5a9-6ced858cb974","Type":"ContainerStarted","Data":"edb119d383758b51fad5555188e081aba816283333b85f5639fad1cfbaf98c58"} Feb 27 17:24:22 crc kubenswrapper[4700]: I0227 17:24:22.852716 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"7602b1c6-8920-4dfd-a379-a008447edb85","Type":"ContainerStarted","Data":"ce1121d5707809610c23199368e379c0831bd126895065a5a9e494594e51bb03"} Feb 27 17:24:22 crc kubenswrapper[4700]: I0227 17:24:22.852766 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"7602b1c6-8920-4dfd-a379-a008447edb85","Type":"ContainerStarted","Data":"1bbdf6b92c8127c1bc6fe8115cf5357063a3fb868c69a0e9dd230795a9f1d709"} Feb 27 17:24:22 crc kubenswrapper[4700]: I0227 17:24:22.853490 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Feb 27 17:24:22 crc kubenswrapper[4700]: I0227 17:24:22.866329 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.866312353 podStartE2EDuration="2.866312353s" podCreationTimestamp="2026-02-27 17:24:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:24:22.863795767 +0000 UTC m=+1422.849108554" watchObservedRunningTime="2026-02-27 17:24:22.866312353 +0000 UTC m=+1422.851625100" Feb 27 17:24:22 crc kubenswrapper[4700]: I0227 17:24:22.885289 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.885268814 podStartE2EDuration="2.885268814s" podCreationTimestamp="2026-02-27 17:24:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:24:22.883707223 +0000 UTC m=+1422.869019990" watchObservedRunningTime="2026-02-27 17:24:22.885268814 +0000 UTC m=+1422.870581561" Feb 27 17:24:22 crc kubenswrapper[4700]: I0227 17:24:22.918622 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-api-0" Feb 27 17:24:22 crc kubenswrapper[4700]: I0227 17:24:22.929414 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-api-0" Feb 27 17:24:23 crc kubenswrapper[4700]: I0227 17:24:23.878123 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Feb 27 17:24:24 crc kubenswrapper[4700]: I0227 17:24:24.066651 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-applier-0" Feb 27 17:24:24 crc kubenswrapper[4700]: I0227 17:24:24.191343 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-applier-0" Feb 27 17:24:24 crc kubenswrapper[4700]: I0227 17:24:24.920305 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-applier-0" Feb 27 17:24:26 crc kubenswrapper[4700]: I0227 17:24:26.274313 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 27 17:24:27 crc kubenswrapper[4700]: I0227 17:24:27.138923 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Feb 27 17:24:27 crc kubenswrapper[4700]: I0227 17:24:27.171952 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-decision-engine-0" Feb 27 17:24:27 crc kubenswrapper[4700]: I0227 17:24:27.222943 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 27 17:24:27 crc kubenswrapper[4700]: I0227 17:24:27.223259 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="d204fe1c-3288-4837-992f-a04ddea6dbd3" containerName="kube-state-metrics" containerID="cri-o://faca510343c0aafdfbd58f2b0053ee3d479c24c17c45411fbcb42ad20c39d844" gracePeriod=30 Feb 27 17:24:27 crc kubenswrapper[4700]: I0227 17:24:27.908639 4700 generic.go:334] "Generic (PLEG): container finished" podID="d204fe1c-3288-4837-992f-a04ddea6dbd3" containerID="faca510343c0aafdfbd58f2b0053ee3d479c24c17c45411fbcb42ad20c39d844" exitCode=2 Feb 27 17:24:27 crc kubenswrapper[4700]: I0227 17:24:27.908735 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d204fe1c-3288-4837-992f-a04ddea6dbd3","Type":"ContainerDied","Data":"faca510343c0aafdfbd58f2b0053ee3d479c24c17c45411fbcb42ad20c39d844"} Feb 27 17:24:27 crc kubenswrapper[4700]: I0227 17:24:27.909604 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d204fe1c-3288-4837-992f-a04ddea6dbd3","Type":"ContainerDied","Data":"dd085c8bf25a8c9a5becee29a8178c0afe5712b089d9d7fffd72f63deae2b02f"} Feb 27 17:24:27 crc kubenswrapper[4700]: I0227 17:24:27.909624 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd085c8bf25a8c9a5becee29a8178c0afe5712b089d9d7fffd72f63deae2b02f" Feb 27 17:24:27 crc kubenswrapper[4700]: I0227 17:24:27.909836 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Feb 27 17:24:27 crc kubenswrapper[4700]: I0227 17:24:27.920570 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 27 17:24:27 crc kubenswrapper[4700]: I0227 17:24:27.954747 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-decision-engine-0" Feb 27 17:24:28 crc kubenswrapper[4700]: I0227 17:24:28.036201 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bpmhn\" (UniqueName: \"kubernetes.io/projected/d204fe1c-3288-4837-992f-a04ddea6dbd3-kube-api-access-bpmhn\") pod \"d204fe1c-3288-4837-992f-a04ddea6dbd3\" (UID: \"d204fe1c-3288-4837-992f-a04ddea6dbd3\") " Feb 27 17:24:28 crc kubenswrapper[4700]: I0227 17:24:28.047427 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d204fe1c-3288-4837-992f-a04ddea6dbd3-kube-api-access-bpmhn" (OuterVolumeSpecName: "kube-api-access-bpmhn") pod "d204fe1c-3288-4837-992f-a04ddea6dbd3" (UID: "d204fe1c-3288-4837-992f-a04ddea6dbd3"). InnerVolumeSpecName "kube-api-access-bpmhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:24:28 crc kubenswrapper[4700]: I0227 17:24:28.139298 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bpmhn\" (UniqueName: \"kubernetes.io/projected/d204fe1c-3288-4837-992f-a04ddea6dbd3-kube-api-access-bpmhn\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:28 crc kubenswrapper[4700]: I0227 17:24:28.918593 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 27 17:24:28 crc kubenswrapper[4700]: I0227 17:24:28.954881 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 27 17:24:28 crc kubenswrapper[4700]: I0227 17:24:28.964183 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 27 17:24:28 crc kubenswrapper[4700]: I0227 17:24:28.977133 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 27 17:24:28 crc kubenswrapper[4700]: E0227 17:24:28.977525 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d204fe1c-3288-4837-992f-a04ddea6dbd3" containerName="kube-state-metrics" Feb 27 17:24:28 crc kubenswrapper[4700]: I0227 17:24:28.977542 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="d204fe1c-3288-4837-992f-a04ddea6dbd3" containerName="kube-state-metrics" Feb 27 17:24:28 crc kubenswrapper[4700]: I0227 17:24:28.977719 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="d204fe1c-3288-4837-992f-a04ddea6dbd3" containerName="kube-state-metrics" Feb 27 17:24:28 crc kubenswrapper[4700]: I0227 17:24:28.978347 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 27 17:24:28 crc kubenswrapper[4700]: I0227 17:24:28.981911 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Feb 27 17:24:28 crc kubenswrapper[4700]: I0227 17:24:28.987178 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Feb 27 17:24:28 crc kubenswrapper[4700]: I0227 17:24:28.996760 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d204fe1c-3288-4837-992f-a04ddea6dbd3" path="/var/lib/kubelet/pods/d204fe1c-3288-4837-992f-a04ddea6dbd3/volumes" Feb 27 17:24:28 crc kubenswrapper[4700]: I0227 17:24:28.999199 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 27 17:24:29 crc kubenswrapper[4700]: I0227 17:24:29.161644 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f100325f-1efc-423e-9e23-e964c5547135-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"f100325f-1efc-423e-9e23-e964c5547135\") " pod="openstack/kube-state-metrics-0" Feb 27 17:24:29 crc kubenswrapper[4700]: I0227 17:24:29.161777 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/f100325f-1efc-423e-9e23-e964c5547135-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"f100325f-1efc-423e-9e23-e964c5547135\") " pod="openstack/kube-state-metrics-0" Feb 27 17:24:29 crc kubenswrapper[4700]: I0227 17:24:29.161900 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/f100325f-1efc-423e-9e23-e964c5547135-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"f100325f-1efc-423e-9e23-e964c5547135\") " pod="openstack/kube-state-metrics-0" Feb 27 17:24:29 crc kubenswrapper[4700]: I0227 17:24:29.161984 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzfn4\" (UniqueName: \"kubernetes.io/projected/f100325f-1efc-423e-9e23-e964c5547135-kube-api-access-kzfn4\") pod \"kube-state-metrics-0\" (UID: \"f100325f-1efc-423e-9e23-e964c5547135\") " pod="openstack/kube-state-metrics-0" Feb 27 17:24:29 crc kubenswrapper[4700]: I0227 17:24:29.262839 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f100325f-1efc-423e-9e23-e964c5547135-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"f100325f-1efc-423e-9e23-e964c5547135\") " pod="openstack/kube-state-metrics-0" Feb 27 17:24:29 crc kubenswrapper[4700]: I0227 17:24:29.262910 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/f100325f-1efc-423e-9e23-e964c5547135-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"f100325f-1efc-423e-9e23-e964c5547135\") " pod="openstack/kube-state-metrics-0" Feb 27 17:24:29 crc kubenswrapper[4700]: I0227 17:24:29.262971 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/f100325f-1efc-423e-9e23-e964c5547135-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"f100325f-1efc-423e-9e23-e964c5547135\") " pod="openstack/kube-state-metrics-0" Feb 27 17:24:29 crc kubenswrapper[4700]: I0227 17:24:29.263016 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzfn4\" (UniqueName: \"kubernetes.io/projected/f100325f-1efc-423e-9e23-e964c5547135-kube-api-access-kzfn4\") pod \"kube-state-metrics-0\" (UID: \"f100325f-1efc-423e-9e23-e964c5547135\") " pod="openstack/kube-state-metrics-0" Feb 27 17:24:29 crc kubenswrapper[4700]: I0227 17:24:29.268660 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f100325f-1efc-423e-9e23-e964c5547135-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"f100325f-1efc-423e-9e23-e964c5547135\") " pod="openstack/kube-state-metrics-0" Feb 27 17:24:29 crc kubenswrapper[4700]: I0227 17:24:29.269244 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/f100325f-1efc-423e-9e23-e964c5547135-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"f100325f-1efc-423e-9e23-e964c5547135\") " pod="openstack/kube-state-metrics-0" Feb 27 17:24:29 crc kubenswrapper[4700]: I0227 17:24:29.269930 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/f100325f-1efc-423e-9e23-e964c5547135-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"f100325f-1efc-423e-9e23-e964c5547135\") " pod="openstack/kube-state-metrics-0" Feb 27 17:24:29 crc kubenswrapper[4700]: I0227 17:24:29.286345 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzfn4\" (UniqueName: \"kubernetes.io/projected/f100325f-1efc-423e-9e23-e964c5547135-kube-api-access-kzfn4\") pod \"kube-state-metrics-0\" (UID: \"f100325f-1efc-423e-9e23-e964c5547135\") " pod="openstack/kube-state-metrics-0" Feb 27 17:24:29 crc kubenswrapper[4700]: I0227 17:24:29.294812 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 27 17:24:29 crc kubenswrapper[4700]: I0227 17:24:29.429758 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:24:29 crc kubenswrapper[4700]: I0227 17:24:29.430229 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7ce63ee0-bc69-4649-b644-34854c003845" containerName="ceilometer-central-agent" containerID="cri-o://986968e0c42c3b559f64654abac2a6fc19896236cf264f9c9cf307667e85b157" gracePeriod=30 Feb 27 17:24:29 crc kubenswrapper[4700]: I0227 17:24:29.430609 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7ce63ee0-bc69-4649-b644-34854c003845" containerName="proxy-httpd" containerID="cri-o://35d6afc4239d670644f97c066dc5c882b65e5365dfc28b389ce805563136bf19" gracePeriod=30 Feb 27 17:24:29 crc kubenswrapper[4700]: I0227 17:24:29.430653 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7ce63ee0-bc69-4649-b644-34854c003845" containerName="sg-core" containerID="cri-o://2ca94d3bd4e956f51cb9ac69bbc45b31e426962d22261e81abab6037c2e03fa8" gracePeriod=30 Feb 27 17:24:29 crc kubenswrapper[4700]: I0227 17:24:29.430684 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7ce63ee0-bc69-4649-b644-34854c003845" containerName="ceilometer-notification-agent" containerID="cri-o://42ec230b3f62902d00c3fd43df8526b7308de78bc5499217ccae8575e86dc256" gracePeriod=30 Feb 27 17:24:29 crc kubenswrapper[4700]: I0227 17:24:29.763518 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 27 17:24:29 crc kubenswrapper[4700]: W0227 17:24:29.773939 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf100325f_1efc_423e_9e23_e964c5547135.slice/crio-f58022d6cefebc22148206fe08f51a728b1539b65e234cfacd13d4c8a71f3248 WatchSource:0}: Error finding container f58022d6cefebc22148206fe08f51a728b1539b65e234cfacd13d4c8a71f3248: Status 404 returned error can't find the container with id f58022d6cefebc22148206fe08f51a728b1539b65e234cfacd13d4c8a71f3248 Feb 27 17:24:29 crc kubenswrapper[4700]: I0227 17:24:29.930501 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f100325f-1efc-423e-9e23-e964c5547135","Type":"ContainerStarted","Data":"f58022d6cefebc22148206fe08f51a728b1539b65e234cfacd13d4c8a71f3248"} Feb 27 17:24:29 crc kubenswrapper[4700]: I0227 17:24:29.934113 4700 generic.go:334] "Generic (PLEG): container finished" podID="7ce63ee0-bc69-4649-b644-34854c003845" containerID="35d6afc4239d670644f97c066dc5c882b65e5365dfc28b389ce805563136bf19" exitCode=0 Feb 27 17:24:29 crc kubenswrapper[4700]: I0227 17:24:29.934142 4700 generic.go:334] "Generic (PLEG): container finished" podID="7ce63ee0-bc69-4649-b644-34854c003845" containerID="2ca94d3bd4e956f51cb9ac69bbc45b31e426962d22261e81abab6037c2e03fa8" exitCode=2 Feb 27 17:24:29 crc kubenswrapper[4700]: I0227 17:24:29.934153 4700 generic.go:334] "Generic (PLEG): container finished" podID="7ce63ee0-bc69-4649-b644-34854c003845" containerID="986968e0c42c3b559f64654abac2a6fc19896236cf264f9c9cf307667e85b157" exitCode=0 Feb 27 17:24:29 crc kubenswrapper[4700]: I0227 17:24:29.934161 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ce63ee0-bc69-4649-b644-34854c003845","Type":"ContainerDied","Data":"35d6afc4239d670644f97c066dc5c882b65e5365dfc28b389ce805563136bf19"} Feb 27 17:24:29 crc kubenswrapper[4700]: I0227 17:24:29.934224 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ce63ee0-bc69-4649-b644-34854c003845","Type":"ContainerDied","Data":"2ca94d3bd4e956f51cb9ac69bbc45b31e426962d22261e81abab6037c2e03fa8"} Feb 27 17:24:29 crc kubenswrapper[4700]: I0227 17:24:29.934248 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ce63ee0-bc69-4649-b644-34854c003845","Type":"ContainerDied","Data":"986968e0c42c3b559f64654abac2a6fc19896236cf264f9c9cf307667e85b157"} Feb 27 17:24:30 crc kubenswrapper[4700]: I0227 17:24:30.256259 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 27 17:24:30 crc kubenswrapper[4700]: I0227 17:24:30.256314 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 27 17:24:30 crc kubenswrapper[4700]: I0227 17:24:30.954062 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f100325f-1efc-423e-9e23-e964c5547135","Type":"ContainerStarted","Data":"d2ade8f595b7555aba5ac8cc2af53d0264d5045d9b91c6429257d1e8e14282b0"} Feb 27 17:24:30 crc kubenswrapper[4700]: I0227 17:24:30.956031 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 27 17:24:30 crc kubenswrapper[4700]: I0227 17:24:30.990542 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.577079776 podStartE2EDuration="2.990523595s" podCreationTimestamp="2026-02-27 17:24:28 +0000 UTC" firstStartedPulling="2026-02-27 17:24:29.777293992 +0000 UTC m=+1429.762606749" lastFinishedPulling="2026-02-27 17:24:30.190737791 +0000 UTC m=+1430.176050568" observedRunningTime="2026-02-27 17:24:30.973081875 +0000 UTC m=+1430.958394622" watchObservedRunningTime="2026-02-27 17:24:30.990523595 +0000 UTC m=+1430.975836342" Feb 27 17:24:31 crc kubenswrapper[4700]: I0227 17:24:31.275030 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 27 17:24:31 crc kubenswrapper[4700]: I0227 17:24:31.336404 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Feb 27 17:24:31 crc kubenswrapper[4700]: I0227 17:24:31.336593 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 27 17:24:31 crc kubenswrapper[4700]: I0227 17:24:31.338786 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ea091df4-f7f3-4580-b20e-215922e9bbed" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.222:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 27 17:24:31 crc kubenswrapper[4700]: I0227 17:24:31.339154 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ea091df4-f7f3-4580-b20e-215922e9bbed" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.222:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 27 17:24:32 crc kubenswrapper[4700]: I0227 17:24:32.008768 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 27 17:24:35 crc kubenswrapper[4700]: I0227 17:24:35.746558 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 17:24:35 crc kubenswrapper[4700]: I0227 17:24:35.900660 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ce63ee0-bc69-4649-b644-34854c003845-scripts\") pod \"7ce63ee0-bc69-4649-b644-34854c003845\" (UID: \"7ce63ee0-bc69-4649-b644-34854c003845\") " Feb 27 17:24:35 crc kubenswrapper[4700]: I0227 17:24:35.900825 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ce63ee0-bc69-4649-b644-34854c003845-run-httpd\") pod \"7ce63ee0-bc69-4649-b644-34854c003845\" (UID: \"7ce63ee0-bc69-4649-b644-34854c003845\") " Feb 27 17:24:35 crc kubenswrapper[4700]: I0227 17:24:35.900888 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68z8n\" (UniqueName: \"kubernetes.io/projected/7ce63ee0-bc69-4649-b644-34854c003845-kube-api-access-68z8n\") pod \"7ce63ee0-bc69-4649-b644-34854c003845\" (UID: \"7ce63ee0-bc69-4649-b644-34854c003845\") " Feb 27 17:24:35 crc kubenswrapper[4700]: I0227 17:24:35.900904 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ce63ee0-bc69-4649-b644-34854c003845-config-data\") pod \"7ce63ee0-bc69-4649-b644-34854c003845\" (UID: \"7ce63ee0-bc69-4649-b644-34854c003845\") " Feb 27 17:24:35 crc kubenswrapper[4700]: I0227 17:24:35.900967 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ce63ee0-bc69-4649-b644-34854c003845-combined-ca-bundle\") pod \"7ce63ee0-bc69-4649-b644-34854c003845\" (UID: \"7ce63ee0-bc69-4649-b644-34854c003845\") " Feb 27 17:24:35 crc kubenswrapper[4700]: I0227 17:24:35.901009 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7ce63ee0-bc69-4649-b644-34854c003845-sg-core-conf-yaml\") pod \"7ce63ee0-bc69-4649-b644-34854c003845\" (UID: \"7ce63ee0-bc69-4649-b644-34854c003845\") " Feb 27 17:24:35 crc kubenswrapper[4700]: I0227 17:24:35.901028 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ce63ee0-bc69-4649-b644-34854c003845-log-httpd\") pod \"7ce63ee0-bc69-4649-b644-34854c003845\" (UID: \"7ce63ee0-bc69-4649-b644-34854c003845\") " Feb 27 17:24:35 crc kubenswrapper[4700]: I0227 17:24:35.901961 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ce63ee0-bc69-4649-b644-34854c003845-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7ce63ee0-bc69-4649-b644-34854c003845" (UID: "7ce63ee0-bc69-4649-b644-34854c003845"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:24:35 crc kubenswrapper[4700]: I0227 17:24:35.907545 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ce63ee0-bc69-4649-b644-34854c003845-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7ce63ee0-bc69-4649-b644-34854c003845" (UID: "7ce63ee0-bc69-4649-b644-34854c003845"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:24:35 crc kubenswrapper[4700]: I0227 17:24:35.908175 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ce63ee0-bc69-4649-b644-34854c003845-scripts" (OuterVolumeSpecName: "scripts") pod "7ce63ee0-bc69-4649-b644-34854c003845" (UID: "7ce63ee0-bc69-4649-b644-34854c003845"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:35 crc kubenswrapper[4700]: I0227 17:24:35.916062 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ce63ee0-bc69-4649-b644-34854c003845-kube-api-access-68z8n" (OuterVolumeSpecName: "kube-api-access-68z8n") pod "7ce63ee0-bc69-4649-b644-34854c003845" (UID: "7ce63ee0-bc69-4649-b644-34854c003845"). InnerVolumeSpecName "kube-api-access-68z8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:24:35 crc kubenswrapper[4700]: I0227 17:24:35.943157 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ce63ee0-bc69-4649-b644-34854c003845-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "7ce63ee0-bc69-4649-b644-34854c003845" (UID: "7ce63ee0-bc69-4649-b644-34854c003845"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.003171 4700 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ce63ee0-bc69-4649-b644-34854c003845-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.003196 4700 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ce63ee0-bc69-4649-b644-34854c003845-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.003205 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68z8n\" (UniqueName: \"kubernetes.io/projected/7ce63ee0-bc69-4649-b644-34854c003845-kube-api-access-68z8n\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.003214 4700 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7ce63ee0-bc69-4649-b644-34854c003845-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.003222 4700 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ce63ee0-bc69-4649-b644-34854c003845-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.017935 4700 generic.go:334] "Generic (PLEG): container finished" podID="7ce63ee0-bc69-4649-b644-34854c003845" containerID="42ec230b3f62902d00c3fd43df8526b7308de78bc5499217ccae8575e86dc256" exitCode=0 Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.017970 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ce63ee0-bc69-4649-b644-34854c003845","Type":"ContainerDied","Data":"42ec230b3f62902d00c3fd43df8526b7308de78bc5499217ccae8575e86dc256"} Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.017993 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ce63ee0-bc69-4649-b644-34854c003845","Type":"ContainerDied","Data":"44b24c6e8002529a4461c4d4cb652bfe388826c1770a837cc27fb733c2604652"} Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.018009 4700 scope.go:117] "RemoveContainer" containerID="35d6afc4239d670644f97c066dc5c882b65e5365dfc28b389ce805563136bf19" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.019197 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.023190 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ce63ee0-bc69-4649-b644-34854c003845-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7ce63ee0-bc69-4649-b644-34854c003845" (UID: "7ce63ee0-bc69-4649-b644-34854c003845"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.049623 4700 scope.go:117] "RemoveContainer" containerID="2ca94d3bd4e956f51cb9ac69bbc45b31e426962d22261e81abab6037c2e03fa8" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.069630 4700 scope.go:117] "RemoveContainer" containerID="42ec230b3f62902d00c3fd43df8526b7308de78bc5499217ccae8575e86dc256" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.083145 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ce63ee0-bc69-4649-b644-34854c003845-config-data" (OuterVolumeSpecName: "config-data") pod "7ce63ee0-bc69-4649-b644-34854c003845" (UID: "7ce63ee0-bc69-4649-b644-34854c003845"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.088782 4700 scope.go:117] "RemoveContainer" containerID="986968e0c42c3b559f64654abac2a6fc19896236cf264f9c9cf307667e85b157" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.105138 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ce63ee0-bc69-4649-b644-34854c003845-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.105161 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ce63ee0-bc69-4649-b644-34854c003845-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.109081 4700 scope.go:117] "RemoveContainer" containerID="35d6afc4239d670644f97c066dc5c882b65e5365dfc28b389ce805563136bf19" Feb 27 17:24:36 crc kubenswrapper[4700]: E0227 17:24:36.109502 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35d6afc4239d670644f97c066dc5c882b65e5365dfc28b389ce805563136bf19\": container with ID starting with 35d6afc4239d670644f97c066dc5c882b65e5365dfc28b389ce805563136bf19 not found: ID does not exist" containerID="35d6afc4239d670644f97c066dc5c882b65e5365dfc28b389ce805563136bf19" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.109536 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35d6afc4239d670644f97c066dc5c882b65e5365dfc28b389ce805563136bf19"} err="failed to get container status \"35d6afc4239d670644f97c066dc5c882b65e5365dfc28b389ce805563136bf19\": rpc error: code = NotFound desc = could not find container \"35d6afc4239d670644f97c066dc5c882b65e5365dfc28b389ce805563136bf19\": container with ID starting with 35d6afc4239d670644f97c066dc5c882b65e5365dfc28b389ce805563136bf19 not found: ID does not exist" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.109563 4700 scope.go:117] "RemoveContainer" containerID="2ca94d3bd4e956f51cb9ac69bbc45b31e426962d22261e81abab6037c2e03fa8" Feb 27 17:24:36 crc kubenswrapper[4700]: E0227 17:24:36.109927 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ca94d3bd4e956f51cb9ac69bbc45b31e426962d22261e81abab6037c2e03fa8\": container with ID starting with 2ca94d3bd4e956f51cb9ac69bbc45b31e426962d22261e81abab6037c2e03fa8 not found: ID does not exist" containerID="2ca94d3bd4e956f51cb9ac69bbc45b31e426962d22261e81abab6037c2e03fa8" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.109960 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ca94d3bd4e956f51cb9ac69bbc45b31e426962d22261e81abab6037c2e03fa8"} err="failed to get container status \"2ca94d3bd4e956f51cb9ac69bbc45b31e426962d22261e81abab6037c2e03fa8\": rpc error: code = NotFound desc = could not find container \"2ca94d3bd4e956f51cb9ac69bbc45b31e426962d22261e81abab6037c2e03fa8\": container with ID starting with 2ca94d3bd4e956f51cb9ac69bbc45b31e426962d22261e81abab6037c2e03fa8 not found: ID does not exist" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.109980 4700 scope.go:117] "RemoveContainer" containerID="42ec230b3f62902d00c3fd43df8526b7308de78bc5499217ccae8575e86dc256" Feb 27 17:24:36 crc kubenswrapper[4700]: E0227 17:24:36.110234 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42ec230b3f62902d00c3fd43df8526b7308de78bc5499217ccae8575e86dc256\": container with ID starting with 42ec230b3f62902d00c3fd43df8526b7308de78bc5499217ccae8575e86dc256 not found: ID does not exist" containerID="42ec230b3f62902d00c3fd43df8526b7308de78bc5499217ccae8575e86dc256" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.110269 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42ec230b3f62902d00c3fd43df8526b7308de78bc5499217ccae8575e86dc256"} err="failed to get container status \"42ec230b3f62902d00c3fd43df8526b7308de78bc5499217ccae8575e86dc256\": rpc error: code = NotFound desc = could not find container \"42ec230b3f62902d00c3fd43df8526b7308de78bc5499217ccae8575e86dc256\": container with ID starting with 42ec230b3f62902d00c3fd43df8526b7308de78bc5499217ccae8575e86dc256 not found: ID does not exist" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.110296 4700 scope.go:117] "RemoveContainer" containerID="986968e0c42c3b559f64654abac2a6fc19896236cf264f9c9cf307667e85b157" Feb 27 17:24:36 crc kubenswrapper[4700]: E0227 17:24:36.110568 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"986968e0c42c3b559f64654abac2a6fc19896236cf264f9c9cf307667e85b157\": container with ID starting with 986968e0c42c3b559f64654abac2a6fc19896236cf264f9c9cf307667e85b157 not found: ID does not exist" containerID="986968e0c42c3b559f64654abac2a6fc19896236cf264f9c9cf307667e85b157" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.110592 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"986968e0c42c3b559f64654abac2a6fc19896236cf264f9c9cf307667e85b157"} err="failed to get container status \"986968e0c42c3b559f64654abac2a6fc19896236cf264f9c9cf307667e85b157\": rpc error: code = NotFound desc = could not find container \"986968e0c42c3b559f64654abac2a6fc19896236cf264f9c9cf307667e85b157\": container with ID starting with 986968e0c42c3b559f64654abac2a6fc19896236cf264f9c9cf307667e85b157 not found: ID does not exist" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.392731 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.412451 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.412589 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.417817 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.430782 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:24:36 crc kubenswrapper[4700]: E0227 17:24:36.431301 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ce63ee0-bc69-4649-b644-34854c003845" containerName="ceilometer-notification-agent" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.431323 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ce63ee0-bc69-4649-b644-34854c003845" containerName="ceilometer-notification-agent" Feb 27 17:24:36 crc kubenswrapper[4700]: E0227 17:24:36.431356 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ce63ee0-bc69-4649-b644-34854c003845" containerName="ceilometer-central-agent" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.431364 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ce63ee0-bc69-4649-b644-34854c003845" containerName="ceilometer-central-agent" Feb 27 17:24:36 crc kubenswrapper[4700]: E0227 17:24:36.431383 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ce63ee0-bc69-4649-b644-34854c003845" containerName="sg-core" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.431392 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ce63ee0-bc69-4649-b644-34854c003845" containerName="sg-core" Feb 27 17:24:36 crc kubenswrapper[4700]: E0227 17:24:36.431416 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ce63ee0-bc69-4649-b644-34854c003845" containerName="proxy-httpd" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.431424 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ce63ee0-bc69-4649-b644-34854c003845" containerName="proxy-httpd" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.431753 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ce63ee0-bc69-4649-b644-34854c003845" containerName="proxy-httpd" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.431774 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ce63ee0-bc69-4649-b644-34854c003845" containerName="sg-core" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.431792 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ce63ee0-bc69-4649-b644-34854c003845" containerName="ceilometer-notification-agent" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.431808 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ce63ee0-bc69-4649-b644-34854c003845" containerName="ceilometer-central-agent" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.434414 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.437036 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.437062 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.438016 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.438561 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.618216 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/095a6539-e7fd-461b-9f54-8598f8d3629d-config-data\") pod \"ceilometer-0\" (UID: \"095a6539-e7fd-461b-9f54-8598f8d3629d\") " pod="openstack/ceilometer-0" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.618357 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/095a6539-e7fd-461b-9f54-8598f8d3629d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"095a6539-e7fd-461b-9f54-8598f8d3629d\") " pod="openstack/ceilometer-0" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.618450 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/095a6539-e7fd-461b-9f54-8598f8d3629d-scripts\") pod \"ceilometer-0\" (UID: \"095a6539-e7fd-461b-9f54-8598f8d3629d\") " pod="openstack/ceilometer-0" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.618538 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/095a6539-e7fd-461b-9f54-8598f8d3629d-run-httpd\") pod \"ceilometer-0\" (UID: \"095a6539-e7fd-461b-9f54-8598f8d3629d\") " pod="openstack/ceilometer-0" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.618581 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/095a6539-e7fd-461b-9f54-8598f8d3629d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"095a6539-e7fd-461b-9f54-8598f8d3629d\") " pod="openstack/ceilometer-0" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.618680 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/095a6539-e7fd-461b-9f54-8598f8d3629d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"095a6539-e7fd-461b-9f54-8598f8d3629d\") " pod="openstack/ceilometer-0" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.618845 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/095a6539-e7fd-461b-9f54-8598f8d3629d-log-httpd\") pod \"ceilometer-0\" (UID: \"095a6539-e7fd-461b-9f54-8598f8d3629d\") " pod="openstack/ceilometer-0" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.618965 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4bwb\" (UniqueName: \"kubernetes.io/projected/095a6539-e7fd-461b-9f54-8598f8d3629d-kube-api-access-k4bwb\") pod \"ceilometer-0\" (UID: \"095a6539-e7fd-461b-9f54-8598f8d3629d\") " pod="openstack/ceilometer-0" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.720997 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/095a6539-e7fd-461b-9f54-8598f8d3629d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"095a6539-e7fd-461b-9f54-8598f8d3629d\") " pod="openstack/ceilometer-0" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.721202 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/095a6539-e7fd-461b-9f54-8598f8d3629d-log-httpd\") pod \"ceilometer-0\" (UID: \"095a6539-e7fd-461b-9f54-8598f8d3629d\") " pod="openstack/ceilometer-0" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.721303 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4bwb\" (UniqueName: \"kubernetes.io/projected/095a6539-e7fd-461b-9f54-8598f8d3629d-kube-api-access-k4bwb\") pod \"ceilometer-0\" (UID: \"095a6539-e7fd-461b-9f54-8598f8d3629d\") " pod="openstack/ceilometer-0" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.721366 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/095a6539-e7fd-461b-9f54-8598f8d3629d-config-data\") pod \"ceilometer-0\" (UID: \"095a6539-e7fd-461b-9f54-8598f8d3629d\") " pod="openstack/ceilometer-0" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.721446 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/095a6539-e7fd-461b-9f54-8598f8d3629d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"095a6539-e7fd-461b-9f54-8598f8d3629d\") " pod="openstack/ceilometer-0" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.721550 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/095a6539-e7fd-461b-9f54-8598f8d3629d-scripts\") pod \"ceilometer-0\" (UID: \"095a6539-e7fd-461b-9f54-8598f8d3629d\") " pod="openstack/ceilometer-0" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.721600 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/095a6539-e7fd-461b-9f54-8598f8d3629d-run-httpd\") pod \"ceilometer-0\" (UID: \"095a6539-e7fd-461b-9f54-8598f8d3629d\") " pod="openstack/ceilometer-0" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.721633 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/095a6539-e7fd-461b-9f54-8598f8d3629d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"095a6539-e7fd-461b-9f54-8598f8d3629d\") " pod="openstack/ceilometer-0" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.721885 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/095a6539-e7fd-461b-9f54-8598f8d3629d-log-httpd\") pod \"ceilometer-0\" (UID: \"095a6539-e7fd-461b-9f54-8598f8d3629d\") " pod="openstack/ceilometer-0" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.722154 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/095a6539-e7fd-461b-9f54-8598f8d3629d-run-httpd\") pod \"ceilometer-0\" (UID: \"095a6539-e7fd-461b-9f54-8598f8d3629d\") " pod="openstack/ceilometer-0" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.726680 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/095a6539-e7fd-461b-9f54-8598f8d3629d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"095a6539-e7fd-461b-9f54-8598f8d3629d\") " pod="openstack/ceilometer-0" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.727497 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/095a6539-e7fd-461b-9f54-8598f8d3629d-scripts\") pod \"ceilometer-0\" (UID: \"095a6539-e7fd-461b-9f54-8598f8d3629d\") " pod="openstack/ceilometer-0" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.729337 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/095a6539-e7fd-461b-9f54-8598f8d3629d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"095a6539-e7fd-461b-9f54-8598f8d3629d\") " pod="openstack/ceilometer-0" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.745453 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/095a6539-e7fd-461b-9f54-8598f8d3629d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"095a6539-e7fd-461b-9f54-8598f8d3629d\") " pod="openstack/ceilometer-0" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.748416 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/095a6539-e7fd-461b-9f54-8598f8d3629d-config-data\") pod \"ceilometer-0\" (UID: \"095a6539-e7fd-461b-9f54-8598f8d3629d\") " pod="openstack/ceilometer-0" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.753201 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4bwb\" (UniqueName: \"kubernetes.io/projected/095a6539-e7fd-461b-9f54-8598f8d3629d-kube-api-access-k4bwb\") pod \"ceilometer-0\" (UID: \"095a6539-e7fd-461b-9f54-8598f8d3629d\") " pod="openstack/ceilometer-0" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.759238 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 17:24:36 crc kubenswrapper[4700]: I0227 17:24:36.992989 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ce63ee0-bc69-4649-b644-34854c003845" path="/var/lib/kubelet/pods/7ce63ee0-bc69-4649-b644-34854c003845/volumes" Feb 27 17:24:37 crc kubenswrapper[4700]: I0227 17:24:37.154642 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:24:37 crc kubenswrapper[4700]: W0227 17:24:37.184647 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod095a6539_e7fd_461b_9f54_8598f8d3629d.slice/crio-90bb3577ee808421701f78ac212a5f545b9753906a757383904bbfb706c80a87 WatchSource:0}: Error finding container 90bb3577ee808421701f78ac212a5f545b9753906a757383904bbfb706c80a87: Status 404 returned error can't find the container with id 90bb3577ee808421701f78ac212a5f545b9753906a757383904bbfb706c80a87 Feb 27 17:24:38 crc kubenswrapper[4700]: I0227 17:24:38.040498 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"095a6539-e7fd-461b-9f54-8598f8d3629d","Type":"ContainerStarted","Data":"c2f7f6f3efcb17f0b73cecd62e10125cb8895e576f71bf116e0dd8435870a0d6"} Feb 27 17:24:38 crc kubenswrapper[4700]: I0227 17:24:38.040743 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"095a6539-e7fd-461b-9f54-8598f8d3629d","Type":"ContainerStarted","Data":"8d0b0a95cec7cc0e1f33eb767754c2ecf9feec24a623a31e2a6558c41f8fe978"} Feb 27 17:24:38 crc kubenswrapper[4700]: I0227 17:24:38.040755 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"095a6539-e7fd-461b-9f54-8598f8d3629d","Type":"ContainerStarted","Data":"90bb3577ee808421701f78ac212a5f545b9753906a757383904bbfb706c80a87"} Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.038116 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.049546 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.053988 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"095a6539-e7fd-461b-9f54-8598f8d3629d","Type":"ContainerStarted","Data":"4c06622be61d03967d319c5c2a56f892dbb1ee843a9880787240e7021b92cf17"} Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.080080 4700 generic.go:334] "Generic (PLEG): container finished" podID="121fbb5a-aad2-4a23-bee5-79a13beebcc7" containerID="5332ac22fe4901ba743f6617ea0e18ccfad50da9adf328dd75043d3078428062" exitCode=137 Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.080166 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"121fbb5a-aad2-4a23-bee5-79a13beebcc7","Type":"ContainerDied","Data":"5332ac22fe4901ba743f6617ea0e18ccfad50da9adf328dd75043d3078428062"} Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.080191 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"121fbb5a-aad2-4a23-bee5-79a13beebcc7","Type":"ContainerDied","Data":"bd475a1378478065ed09b931231a301ad986f3f7ecde3b45984e96a8e81c3c44"} Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.080208 4700 scope.go:117] "RemoveContainer" containerID="5332ac22fe4901ba743f6617ea0e18ccfad50da9adf328dd75043d3078428062" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.080370 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.105216 4700 generic.go:334] "Generic (PLEG): container finished" podID="4b296eaf-d6d2-4028-bfac-2573f32841c5" containerID="3a64e23bc32ed82280a4c504ee2a7f4ceffd3f339803744211f6d83907e043c2" exitCode=137 Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.105266 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.105301 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4b296eaf-d6d2-4028-bfac-2573f32841c5","Type":"ContainerDied","Data":"3a64e23bc32ed82280a4c504ee2a7f4ceffd3f339803744211f6d83907e043c2"} Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.105603 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4b296eaf-d6d2-4028-bfac-2573f32841c5","Type":"ContainerDied","Data":"74d08da09a7eb33950cb58a98bd8e41e77dc92b8c82041b7024e8328f7db2d53"} Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.124328 4700 scope.go:117] "RemoveContainer" containerID="5332ac22fe4901ba743f6617ea0e18ccfad50da9adf328dd75043d3078428062" Feb 27 17:24:39 crc kubenswrapper[4700]: E0227 17:24:39.124758 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5332ac22fe4901ba743f6617ea0e18ccfad50da9adf328dd75043d3078428062\": container with ID starting with 5332ac22fe4901ba743f6617ea0e18ccfad50da9adf328dd75043d3078428062 not found: ID does not exist" containerID="5332ac22fe4901ba743f6617ea0e18ccfad50da9adf328dd75043d3078428062" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.124789 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5332ac22fe4901ba743f6617ea0e18ccfad50da9adf328dd75043d3078428062"} err="failed to get container status \"5332ac22fe4901ba743f6617ea0e18ccfad50da9adf328dd75043d3078428062\": rpc error: code = NotFound desc = could not find container \"5332ac22fe4901ba743f6617ea0e18ccfad50da9adf328dd75043d3078428062\": container with ID starting with 5332ac22fe4901ba743f6617ea0e18ccfad50da9adf328dd75043d3078428062 not found: ID does not exist" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.124813 4700 scope.go:117] "RemoveContainer" containerID="3a64e23bc32ed82280a4c504ee2a7f4ceffd3f339803744211f6d83907e043c2" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.144066 4700 scope.go:117] "RemoveContainer" containerID="0d953a3f45063d782dbc5280789fecc41036ec232071b3c63c22625c59b99b55" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.168475 4700 scope.go:117] "RemoveContainer" containerID="3a64e23bc32ed82280a4c504ee2a7f4ceffd3f339803744211f6d83907e043c2" Feb 27 17:24:39 crc kubenswrapper[4700]: E0227 17:24:39.169023 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a64e23bc32ed82280a4c504ee2a7f4ceffd3f339803744211f6d83907e043c2\": container with ID starting with 3a64e23bc32ed82280a4c504ee2a7f4ceffd3f339803744211f6d83907e043c2 not found: ID does not exist" containerID="3a64e23bc32ed82280a4c504ee2a7f4ceffd3f339803744211f6d83907e043c2" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.169056 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a64e23bc32ed82280a4c504ee2a7f4ceffd3f339803744211f6d83907e043c2"} err="failed to get container status \"3a64e23bc32ed82280a4c504ee2a7f4ceffd3f339803744211f6d83907e043c2\": rpc error: code = NotFound desc = could not find container \"3a64e23bc32ed82280a4c504ee2a7f4ceffd3f339803744211f6d83907e043c2\": container with ID starting with 3a64e23bc32ed82280a4c504ee2a7f4ceffd3f339803744211f6d83907e043c2 not found: ID does not exist" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.169079 4700 scope.go:117] "RemoveContainer" containerID="0d953a3f45063d782dbc5280789fecc41036ec232071b3c63c22625c59b99b55" Feb 27 17:24:39 crc kubenswrapper[4700]: E0227 17:24:39.171113 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d953a3f45063d782dbc5280789fecc41036ec232071b3c63c22625c59b99b55\": container with ID starting with 0d953a3f45063d782dbc5280789fecc41036ec232071b3c63c22625c59b99b55 not found: ID does not exist" containerID="0d953a3f45063d782dbc5280789fecc41036ec232071b3c63c22625c59b99b55" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.171163 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d953a3f45063d782dbc5280789fecc41036ec232071b3c63c22625c59b99b55"} err="failed to get container status \"0d953a3f45063d782dbc5280789fecc41036ec232071b3c63c22625c59b99b55\": rpc error: code = NotFound desc = could not find container \"0d953a3f45063d782dbc5280789fecc41036ec232071b3c63c22625c59b99b55\": container with ID starting with 0d953a3f45063d782dbc5280789fecc41036ec232071b3c63c22625c59b99b55 not found: ID does not exist" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.176878 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/121fbb5a-aad2-4a23-bee5-79a13beebcc7-config-data\") pod \"121fbb5a-aad2-4a23-bee5-79a13beebcc7\" (UID: \"121fbb5a-aad2-4a23-bee5-79a13beebcc7\") " Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.177007 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqcw9\" (UniqueName: \"kubernetes.io/projected/121fbb5a-aad2-4a23-bee5-79a13beebcc7-kube-api-access-lqcw9\") pod \"121fbb5a-aad2-4a23-bee5-79a13beebcc7\" (UID: \"121fbb5a-aad2-4a23-bee5-79a13beebcc7\") " Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.177037 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fcww\" (UniqueName: \"kubernetes.io/projected/4b296eaf-d6d2-4028-bfac-2573f32841c5-kube-api-access-4fcww\") pod \"4b296eaf-d6d2-4028-bfac-2573f32841c5\" (UID: \"4b296eaf-d6d2-4028-bfac-2573f32841c5\") " Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.177084 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b296eaf-d6d2-4028-bfac-2573f32841c5-logs\") pod \"4b296eaf-d6d2-4028-bfac-2573f32841c5\" (UID: \"4b296eaf-d6d2-4028-bfac-2573f32841c5\") " Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.177104 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/121fbb5a-aad2-4a23-bee5-79a13beebcc7-combined-ca-bundle\") pod \"121fbb5a-aad2-4a23-bee5-79a13beebcc7\" (UID: \"121fbb5a-aad2-4a23-bee5-79a13beebcc7\") " Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.177127 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b296eaf-d6d2-4028-bfac-2573f32841c5-config-data\") pod \"4b296eaf-d6d2-4028-bfac-2573f32841c5\" (UID: \"4b296eaf-d6d2-4028-bfac-2573f32841c5\") " Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.177152 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b296eaf-d6d2-4028-bfac-2573f32841c5-combined-ca-bundle\") pod \"4b296eaf-d6d2-4028-bfac-2573f32841c5\" (UID: \"4b296eaf-d6d2-4028-bfac-2573f32841c5\") " Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.183768 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b296eaf-d6d2-4028-bfac-2573f32841c5-logs" (OuterVolumeSpecName: "logs") pod "4b296eaf-d6d2-4028-bfac-2573f32841c5" (UID: "4b296eaf-d6d2-4028-bfac-2573f32841c5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.184071 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b296eaf-d6d2-4028-bfac-2573f32841c5-kube-api-access-4fcww" (OuterVolumeSpecName: "kube-api-access-4fcww") pod "4b296eaf-d6d2-4028-bfac-2573f32841c5" (UID: "4b296eaf-d6d2-4028-bfac-2573f32841c5"). InnerVolumeSpecName "kube-api-access-4fcww". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.189696 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/121fbb5a-aad2-4a23-bee5-79a13beebcc7-kube-api-access-lqcw9" (OuterVolumeSpecName: "kube-api-access-lqcw9") pod "121fbb5a-aad2-4a23-bee5-79a13beebcc7" (UID: "121fbb5a-aad2-4a23-bee5-79a13beebcc7"). InnerVolumeSpecName "kube-api-access-lqcw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.214293 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/121fbb5a-aad2-4a23-bee5-79a13beebcc7-config-data" (OuterVolumeSpecName: "config-data") pod "121fbb5a-aad2-4a23-bee5-79a13beebcc7" (UID: "121fbb5a-aad2-4a23-bee5-79a13beebcc7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.228007 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/121fbb5a-aad2-4a23-bee5-79a13beebcc7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "121fbb5a-aad2-4a23-bee5-79a13beebcc7" (UID: "121fbb5a-aad2-4a23-bee5-79a13beebcc7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.229556 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b296eaf-d6d2-4028-bfac-2573f32841c5-config-data" (OuterVolumeSpecName: "config-data") pod "4b296eaf-d6d2-4028-bfac-2573f32841c5" (UID: "4b296eaf-d6d2-4028-bfac-2573f32841c5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.232818 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b296eaf-d6d2-4028-bfac-2573f32841c5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4b296eaf-d6d2-4028-bfac-2573f32841c5" (UID: "4b296eaf-d6d2-4028-bfac-2573f32841c5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.279853 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/121fbb5a-aad2-4a23-bee5-79a13beebcc7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.279894 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b296eaf-d6d2-4028-bfac-2573f32841c5-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.279909 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b296eaf-d6d2-4028-bfac-2573f32841c5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.279920 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/121fbb5a-aad2-4a23-bee5-79a13beebcc7-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.279934 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqcw9\" (UniqueName: \"kubernetes.io/projected/121fbb5a-aad2-4a23-bee5-79a13beebcc7-kube-api-access-lqcw9\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.279947 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fcww\" (UniqueName: \"kubernetes.io/projected/4b296eaf-d6d2-4028-bfac-2573f32841c5-kube-api-access-4fcww\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.279958 4700 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b296eaf-d6d2-4028-bfac-2573f32841c5-logs\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.304302 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.503386 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.512824 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.522001 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.530345 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 27 17:24:39 crc kubenswrapper[4700]: E0227 17:24:39.530820 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b296eaf-d6d2-4028-bfac-2573f32841c5" containerName="nova-metadata-log" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.530837 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b296eaf-d6d2-4028-bfac-2573f32841c5" containerName="nova-metadata-log" Feb 27 17:24:39 crc kubenswrapper[4700]: E0227 17:24:39.530856 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b296eaf-d6d2-4028-bfac-2573f32841c5" containerName="nova-metadata-metadata" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.530863 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b296eaf-d6d2-4028-bfac-2573f32841c5" containerName="nova-metadata-metadata" Feb 27 17:24:39 crc kubenswrapper[4700]: E0227 17:24:39.530879 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="121fbb5a-aad2-4a23-bee5-79a13beebcc7" containerName="nova-cell1-novncproxy-novncproxy" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.530885 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="121fbb5a-aad2-4a23-bee5-79a13beebcc7" containerName="nova-cell1-novncproxy-novncproxy" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.531074 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="121fbb5a-aad2-4a23-bee5-79a13beebcc7" containerName="nova-cell1-novncproxy-novncproxy" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.531093 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b296eaf-d6d2-4028-bfac-2573f32841c5" containerName="nova-metadata-log" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.531104 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b296eaf-d6d2-4028-bfac-2573f32841c5" containerName="nova-metadata-metadata" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.531766 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.537760 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.538317 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.539165 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.539288 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.548699 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.557397 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.559070 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.560678 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.560944 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.606735 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.687192 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb9499b9-4b5c-4f5a-9579-e3f7329f36a7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fb9499b9-4b5c-4f5a-9579-e3f7329f36a7\") " pod="openstack/nova-metadata-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.687562 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rn2bf\" (UniqueName: \"kubernetes.io/projected/fb9499b9-4b5c-4f5a-9579-e3f7329f36a7-kube-api-access-rn2bf\") pod \"nova-metadata-0\" (UID: \"fb9499b9-4b5c-4f5a-9579-e3f7329f36a7\") " pod="openstack/nova-metadata-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.687635 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e395b037-9e65-41ac-b615-bee1dbf5f0f9-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e395b037-9e65-41ac-b615-bee1dbf5f0f9\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.687709 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e395b037-9e65-41ac-b615-bee1dbf5f0f9-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e395b037-9e65-41ac-b615-bee1dbf5f0f9\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.687829 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e395b037-9e65-41ac-b615-bee1dbf5f0f9-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e395b037-9e65-41ac-b615-bee1dbf5f0f9\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.687898 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e395b037-9e65-41ac-b615-bee1dbf5f0f9-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e395b037-9e65-41ac-b615-bee1dbf5f0f9\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.688143 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb9499b9-4b5c-4f5a-9579-e3f7329f36a7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fb9499b9-4b5c-4f5a-9579-e3f7329f36a7\") " pod="openstack/nova-metadata-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.688195 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvrjh\" (UniqueName: \"kubernetes.io/projected/e395b037-9e65-41ac-b615-bee1dbf5f0f9-kube-api-access-rvrjh\") pod \"nova-cell1-novncproxy-0\" (UID: \"e395b037-9e65-41ac-b615-bee1dbf5f0f9\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.688269 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb9499b9-4b5c-4f5a-9579-e3f7329f36a7-logs\") pod \"nova-metadata-0\" (UID: \"fb9499b9-4b5c-4f5a-9579-e3f7329f36a7\") " pod="openstack/nova-metadata-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.688296 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb9499b9-4b5c-4f5a-9579-e3f7329f36a7-config-data\") pod \"nova-metadata-0\" (UID: \"fb9499b9-4b5c-4f5a-9579-e3f7329f36a7\") " pod="openstack/nova-metadata-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.790954 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb9499b9-4b5c-4f5a-9579-e3f7329f36a7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fb9499b9-4b5c-4f5a-9579-e3f7329f36a7\") " pod="openstack/nova-metadata-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.791050 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rn2bf\" (UniqueName: \"kubernetes.io/projected/fb9499b9-4b5c-4f5a-9579-e3f7329f36a7-kube-api-access-rn2bf\") pod \"nova-metadata-0\" (UID: \"fb9499b9-4b5c-4f5a-9579-e3f7329f36a7\") " pod="openstack/nova-metadata-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.791126 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e395b037-9e65-41ac-b615-bee1dbf5f0f9-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e395b037-9e65-41ac-b615-bee1dbf5f0f9\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.791177 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e395b037-9e65-41ac-b615-bee1dbf5f0f9-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e395b037-9e65-41ac-b615-bee1dbf5f0f9\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.791208 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e395b037-9e65-41ac-b615-bee1dbf5f0f9-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e395b037-9e65-41ac-b615-bee1dbf5f0f9\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.791261 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e395b037-9e65-41ac-b615-bee1dbf5f0f9-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e395b037-9e65-41ac-b615-bee1dbf5f0f9\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.791308 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb9499b9-4b5c-4f5a-9579-e3f7329f36a7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fb9499b9-4b5c-4f5a-9579-e3f7329f36a7\") " pod="openstack/nova-metadata-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.791352 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvrjh\" (UniqueName: \"kubernetes.io/projected/e395b037-9e65-41ac-b615-bee1dbf5f0f9-kube-api-access-rvrjh\") pod \"nova-cell1-novncproxy-0\" (UID: \"e395b037-9e65-41ac-b615-bee1dbf5f0f9\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.791399 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb9499b9-4b5c-4f5a-9579-e3f7329f36a7-logs\") pod \"nova-metadata-0\" (UID: \"fb9499b9-4b5c-4f5a-9579-e3f7329f36a7\") " pod="openstack/nova-metadata-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.791431 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb9499b9-4b5c-4f5a-9579-e3f7329f36a7-config-data\") pod \"nova-metadata-0\" (UID: \"fb9499b9-4b5c-4f5a-9579-e3f7329f36a7\") " pod="openstack/nova-metadata-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.792229 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb9499b9-4b5c-4f5a-9579-e3f7329f36a7-logs\") pod \"nova-metadata-0\" (UID: \"fb9499b9-4b5c-4f5a-9579-e3f7329f36a7\") " pod="openstack/nova-metadata-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.795953 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e395b037-9e65-41ac-b615-bee1dbf5f0f9-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e395b037-9e65-41ac-b615-bee1dbf5f0f9\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.796169 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb9499b9-4b5c-4f5a-9579-e3f7329f36a7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fb9499b9-4b5c-4f5a-9579-e3f7329f36a7\") " pod="openstack/nova-metadata-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.798887 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e395b037-9e65-41ac-b615-bee1dbf5f0f9-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e395b037-9e65-41ac-b615-bee1dbf5f0f9\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.799048 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb9499b9-4b5c-4f5a-9579-e3f7329f36a7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fb9499b9-4b5c-4f5a-9579-e3f7329f36a7\") " pod="openstack/nova-metadata-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.799178 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e395b037-9e65-41ac-b615-bee1dbf5f0f9-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e395b037-9e65-41ac-b615-bee1dbf5f0f9\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.799553 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb9499b9-4b5c-4f5a-9579-e3f7329f36a7-config-data\") pod \"nova-metadata-0\" (UID: \"fb9499b9-4b5c-4f5a-9579-e3f7329f36a7\") " pod="openstack/nova-metadata-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.799619 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e395b037-9e65-41ac-b615-bee1dbf5f0f9-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e395b037-9e65-41ac-b615-bee1dbf5f0f9\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.813173 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rn2bf\" (UniqueName: \"kubernetes.io/projected/fb9499b9-4b5c-4f5a-9579-e3f7329f36a7-kube-api-access-rn2bf\") pod \"nova-metadata-0\" (UID: \"fb9499b9-4b5c-4f5a-9579-e3f7329f36a7\") " pod="openstack/nova-metadata-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.829058 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvrjh\" (UniqueName: \"kubernetes.io/projected/e395b037-9e65-41ac-b615-bee1dbf5f0f9-kube-api-access-rvrjh\") pod \"nova-cell1-novncproxy-0\" (UID: \"e395b037-9e65-41ac-b615-bee1dbf5f0f9\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.847580 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 27 17:24:39 crc kubenswrapper[4700]: I0227 17:24:39.881906 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 17:24:40 crc kubenswrapper[4700]: I0227 17:24:40.266205 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 27 17:24:40 crc kubenswrapper[4700]: I0227 17:24:40.267825 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 27 17:24:40 crc kubenswrapper[4700]: I0227 17:24:40.284669 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 27 17:24:40 crc kubenswrapper[4700]: I0227 17:24:40.286026 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 27 17:24:40 crc kubenswrapper[4700]: W0227 17:24:40.342130 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode395b037_9e65_41ac_b615_bee1dbf5f0f9.slice/crio-d696378e11160d2913aab09ecdc13b5dea9367758b8321b61ee2c4554e1e2bea WatchSource:0}: Error finding container d696378e11160d2913aab09ecdc13b5dea9367758b8321b61ee2c4554e1e2bea: Status 404 returned error can't find the container with id d696378e11160d2913aab09ecdc13b5dea9367758b8321b61ee2c4554e1e2bea Feb 27 17:24:40 crc kubenswrapper[4700]: I0227 17:24:40.344840 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 27 17:24:40 crc kubenswrapper[4700]: I0227 17:24:40.797440 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 17:24:40 crc kubenswrapper[4700]: W0227 17:24:40.799193 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb9499b9_4b5c_4f5a_9579_e3f7329f36a7.slice/crio-f811c5fcffa60355dc6283e3ed1086cd86b17dd5089ce28c46a64ad557bf98f2 WatchSource:0}: Error finding container f811c5fcffa60355dc6283e3ed1086cd86b17dd5089ce28c46a64ad557bf98f2: Status 404 returned error can't find the container with id f811c5fcffa60355dc6283e3ed1086cd86b17dd5089ce28c46a64ad557bf98f2 Feb 27 17:24:40 crc kubenswrapper[4700]: I0227 17:24:40.992141 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="121fbb5a-aad2-4a23-bee5-79a13beebcc7" path="/var/lib/kubelet/pods/121fbb5a-aad2-4a23-bee5-79a13beebcc7/volumes" Feb 27 17:24:40 crc kubenswrapper[4700]: I0227 17:24:40.993353 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b296eaf-d6d2-4028-bfac-2573f32841c5" path="/var/lib/kubelet/pods/4b296eaf-d6d2-4028-bfac-2573f32841c5/volumes" Feb 27 17:24:41 crc kubenswrapper[4700]: I0227 17:24:41.138930 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e395b037-9e65-41ac-b615-bee1dbf5f0f9","Type":"ContainerStarted","Data":"78f9bf44d068040556a18554382577b0cf5d935d687483fe71509d6a2eb31930"} Feb 27 17:24:41 crc kubenswrapper[4700]: I0227 17:24:41.139260 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e395b037-9e65-41ac-b615-bee1dbf5f0f9","Type":"ContainerStarted","Data":"d696378e11160d2913aab09ecdc13b5dea9367758b8321b61ee2c4554e1e2bea"} Feb 27 17:24:41 crc kubenswrapper[4700]: I0227 17:24:41.141437 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fb9499b9-4b5c-4f5a-9579-e3f7329f36a7","Type":"ContainerStarted","Data":"32ee692b1c18f2953e24371a5c20fd835dc7dce63f05a4c7c9136ba4edf02f31"} Feb 27 17:24:41 crc kubenswrapper[4700]: I0227 17:24:41.141506 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fb9499b9-4b5c-4f5a-9579-e3f7329f36a7","Type":"ContainerStarted","Data":"f811c5fcffa60355dc6283e3ed1086cd86b17dd5089ce28c46a64ad557bf98f2"} Feb 27 17:24:41 crc kubenswrapper[4700]: I0227 17:24:41.141761 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 27 17:24:41 crc kubenswrapper[4700]: I0227 17:24:41.160523 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 27 17:24:41 crc kubenswrapper[4700]: I0227 17:24:41.175297 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.175278397 podStartE2EDuration="2.175278397s" podCreationTimestamp="2026-02-27 17:24:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:24:41.159184643 +0000 UTC m=+1441.144497400" watchObservedRunningTime="2026-02-27 17:24:41.175278397 +0000 UTC m=+1441.160591154" Feb 27 17:24:41 crc kubenswrapper[4700]: I0227 17:24:41.335789 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d8cd4fdf7-74zlx"] Feb 27 17:24:41 crc kubenswrapper[4700]: I0227 17:24:41.337592 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d8cd4fdf7-74zlx" Feb 27 17:24:41 crc kubenswrapper[4700]: I0227 17:24:41.350685 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d8cd4fdf7-74zlx"] Feb 27 17:24:41 crc kubenswrapper[4700]: I0227 17:24:41.428418 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b896d6a2-a26c-4b15-adcf-bb73998922f2-config\") pod \"dnsmasq-dns-5d8cd4fdf7-74zlx\" (UID: \"b896d6a2-a26c-4b15-adcf-bb73998922f2\") " pod="openstack/dnsmasq-dns-5d8cd4fdf7-74zlx" Feb 27 17:24:41 crc kubenswrapper[4700]: I0227 17:24:41.428527 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvx5p\" (UniqueName: \"kubernetes.io/projected/b896d6a2-a26c-4b15-adcf-bb73998922f2-kube-api-access-nvx5p\") pod \"dnsmasq-dns-5d8cd4fdf7-74zlx\" (UID: \"b896d6a2-a26c-4b15-adcf-bb73998922f2\") " pod="openstack/dnsmasq-dns-5d8cd4fdf7-74zlx" Feb 27 17:24:41 crc kubenswrapper[4700]: I0227 17:24:41.428582 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b896d6a2-a26c-4b15-adcf-bb73998922f2-dns-svc\") pod \"dnsmasq-dns-5d8cd4fdf7-74zlx\" (UID: \"b896d6a2-a26c-4b15-adcf-bb73998922f2\") " pod="openstack/dnsmasq-dns-5d8cd4fdf7-74zlx" Feb 27 17:24:41 crc kubenswrapper[4700]: I0227 17:24:41.428608 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b896d6a2-a26c-4b15-adcf-bb73998922f2-dns-swift-storage-0\") pod \"dnsmasq-dns-5d8cd4fdf7-74zlx\" (UID: \"b896d6a2-a26c-4b15-adcf-bb73998922f2\") " pod="openstack/dnsmasq-dns-5d8cd4fdf7-74zlx" Feb 27 17:24:41 crc kubenswrapper[4700]: I0227 17:24:41.428668 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b896d6a2-a26c-4b15-adcf-bb73998922f2-ovsdbserver-sb\") pod \"dnsmasq-dns-5d8cd4fdf7-74zlx\" (UID: \"b896d6a2-a26c-4b15-adcf-bb73998922f2\") " pod="openstack/dnsmasq-dns-5d8cd4fdf7-74zlx" Feb 27 17:24:41 crc kubenswrapper[4700]: I0227 17:24:41.428689 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b896d6a2-a26c-4b15-adcf-bb73998922f2-ovsdbserver-nb\") pod \"dnsmasq-dns-5d8cd4fdf7-74zlx\" (UID: \"b896d6a2-a26c-4b15-adcf-bb73998922f2\") " pod="openstack/dnsmasq-dns-5d8cd4fdf7-74zlx" Feb 27 17:24:41 crc kubenswrapper[4700]: I0227 17:24:41.530730 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b896d6a2-a26c-4b15-adcf-bb73998922f2-dns-svc\") pod \"dnsmasq-dns-5d8cd4fdf7-74zlx\" (UID: \"b896d6a2-a26c-4b15-adcf-bb73998922f2\") " pod="openstack/dnsmasq-dns-5d8cd4fdf7-74zlx" Feb 27 17:24:41 crc kubenswrapper[4700]: I0227 17:24:41.531032 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b896d6a2-a26c-4b15-adcf-bb73998922f2-dns-swift-storage-0\") pod \"dnsmasq-dns-5d8cd4fdf7-74zlx\" (UID: \"b896d6a2-a26c-4b15-adcf-bb73998922f2\") " pod="openstack/dnsmasq-dns-5d8cd4fdf7-74zlx" Feb 27 17:24:41 crc kubenswrapper[4700]: I0227 17:24:41.531089 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b896d6a2-a26c-4b15-adcf-bb73998922f2-ovsdbserver-sb\") pod \"dnsmasq-dns-5d8cd4fdf7-74zlx\" (UID: \"b896d6a2-a26c-4b15-adcf-bb73998922f2\") " pod="openstack/dnsmasq-dns-5d8cd4fdf7-74zlx" Feb 27 17:24:41 crc kubenswrapper[4700]: I0227 17:24:41.531112 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b896d6a2-a26c-4b15-adcf-bb73998922f2-ovsdbserver-nb\") pod \"dnsmasq-dns-5d8cd4fdf7-74zlx\" (UID: \"b896d6a2-a26c-4b15-adcf-bb73998922f2\") " pod="openstack/dnsmasq-dns-5d8cd4fdf7-74zlx" Feb 27 17:24:41 crc kubenswrapper[4700]: I0227 17:24:41.531298 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b896d6a2-a26c-4b15-adcf-bb73998922f2-config\") pod \"dnsmasq-dns-5d8cd4fdf7-74zlx\" (UID: \"b896d6a2-a26c-4b15-adcf-bb73998922f2\") " pod="openstack/dnsmasq-dns-5d8cd4fdf7-74zlx" Feb 27 17:24:41 crc kubenswrapper[4700]: I0227 17:24:41.531327 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvx5p\" (UniqueName: \"kubernetes.io/projected/b896d6a2-a26c-4b15-adcf-bb73998922f2-kube-api-access-nvx5p\") pod \"dnsmasq-dns-5d8cd4fdf7-74zlx\" (UID: \"b896d6a2-a26c-4b15-adcf-bb73998922f2\") " pod="openstack/dnsmasq-dns-5d8cd4fdf7-74zlx" Feb 27 17:24:41 crc kubenswrapper[4700]: I0227 17:24:41.531730 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b896d6a2-a26c-4b15-adcf-bb73998922f2-dns-svc\") pod \"dnsmasq-dns-5d8cd4fdf7-74zlx\" (UID: \"b896d6a2-a26c-4b15-adcf-bb73998922f2\") " pod="openstack/dnsmasq-dns-5d8cd4fdf7-74zlx" Feb 27 17:24:41 crc kubenswrapper[4700]: I0227 17:24:41.532401 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b896d6a2-a26c-4b15-adcf-bb73998922f2-ovsdbserver-nb\") pod \"dnsmasq-dns-5d8cd4fdf7-74zlx\" (UID: \"b896d6a2-a26c-4b15-adcf-bb73998922f2\") " pod="openstack/dnsmasq-dns-5d8cd4fdf7-74zlx" Feb 27 17:24:41 crc kubenswrapper[4700]: I0227 17:24:41.532426 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b896d6a2-a26c-4b15-adcf-bb73998922f2-dns-swift-storage-0\") pod \"dnsmasq-dns-5d8cd4fdf7-74zlx\" (UID: \"b896d6a2-a26c-4b15-adcf-bb73998922f2\") " pod="openstack/dnsmasq-dns-5d8cd4fdf7-74zlx" Feb 27 17:24:41 crc kubenswrapper[4700]: I0227 17:24:41.532921 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b896d6a2-a26c-4b15-adcf-bb73998922f2-config\") pod \"dnsmasq-dns-5d8cd4fdf7-74zlx\" (UID: \"b896d6a2-a26c-4b15-adcf-bb73998922f2\") " pod="openstack/dnsmasq-dns-5d8cd4fdf7-74zlx" Feb 27 17:24:41 crc kubenswrapper[4700]: I0227 17:24:41.533427 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b896d6a2-a26c-4b15-adcf-bb73998922f2-ovsdbserver-sb\") pod \"dnsmasq-dns-5d8cd4fdf7-74zlx\" (UID: \"b896d6a2-a26c-4b15-adcf-bb73998922f2\") " pod="openstack/dnsmasq-dns-5d8cd4fdf7-74zlx" Feb 27 17:24:41 crc kubenswrapper[4700]: I0227 17:24:41.549753 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvx5p\" (UniqueName: \"kubernetes.io/projected/b896d6a2-a26c-4b15-adcf-bb73998922f2-kube-api-access-nvx5p\") pod \"dnsmasq-dns-5d8cd4fdf7-74zlx\" (UID: \"b896d6a2-a26c-4b15-adcf-bb73998922f2\") " pod="openstack/dnsmasq-dns-5d8cd4fdf7-74zlx" Feb 27 17:24:41 crc kubenswrapper[4700]: I0227 17:24:41.668147 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d8cd4fdf7-74zlx" Feb 27 17:24:42 crc kubenswrapper[4700]: I0227 17:24:42.156260 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d8cd4fdf7-74zlx"] Feb 27 17:24:42 crc kubenswrapper[4700]: W0227 17:24:42.165596 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb896d6a2_a26c_4b15_adcf_bb73998922f2.slice/crio-468bfd23b39aec2505dfb99e2b91f26af3a75f29bd3b2c5fdd537b97f41d6c82 WatchSource:0}: Error finding container 468bfd23b39aec2505dfb99e2b91f26af3a75f29bd3b2c5fdd537b97f41d6c82: Status 404 returned error can't find the container with id 468bfd23b39aec2505dfb99e2b91f26af3a75f29bd3b2c5fdd537b97f41d6c82 Feb 27 17:24:42 crc kubenswrapper[4700]: I0227 17:24:42.165774 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"095a6539-e7fd-461b-9f54-8598f8d3629d","Type":"ContainerStarted","Data":"2d8463c7e437670e8cbdad66d7b79629d957e872e6154426a17c45965ae114ce"} Feb 27 17:24:42 crc kubenswrapper[4700]: I0227 17:24:42.166517 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 27 17:24:42 crc kubenswrapper[4700]: I0227 17:24:42.179054 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fb9499b9-4b5c-4f5a-9579-e3f7329f36a7","Type":"ContainerStarted","Data":"9a9c8b97aaeb05640df58b406cf322ef021ffe0c9683755246a4bdd7f62ac10a"} Feb 27 17:24:42 crc kubenswrapper[4700]: I0227 17:24:42.207593 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.266665153 podStartE2EDuration="6.207570294s" podCreationTimestamp="2026-02-27 17:24:36 +0000 UTC" firstStartedPulling="2026-02-27 17:24:37.187290426 +0000 UTC m=+1437.172603173" lastFinishedPulling="2026-02-27 17:24:41.128195567 +0000 UTC m=+1441.113508314" observedRunningTime="2026-02-27 17:24:42.200905589 +0000 UTC m=+1442.186218336" watchObservedRunningTime="2026-02-27 17:24:42.207570294 +0000 UTC m=+1442.192883041" Feb 27 17:24:42 crc kubenswrapper[4700]: I0227 17:24:42.227245 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.227223802 podStartE2EDuration="3.227223802s" podCreationTimestamp="2026-02-27 17:24:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:24:42.217941177 +0000 UTC m=+1442.203253924" watchObservedRunningTime="2026-02-27 17:24:42.227223802 +0000 UTC m=+1442.212536549" Feb 27 17:24:43 crc kubenswrapper[4700]: I0227 17:24:43.188536 4700 generic.go:334] "Generic (PLEG): container finished" podID="b896d6a2-a26c-4b15-adcf-bb73998922f2" containerID="9d5028a387e275812362f6e533e459777ff21a701a1c370699589c2a8fa193ac" exitCode=0 Feb 27 17:24:43 crc kubenswrapper[4700]: I0227 17:24:43.188589 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d8cd4fdf7-74zlx" event={"ID":"b896d6a2-a26c-4b15-adcf-bb73998922f2","Type":"ContainerDied","Data":"9d5028a387e275812362f6e533e459777ff21a701a1c370699589c2a8fa193ac"} Feb 27 17:24:43 crc kubenswrapper[4700]: I0227 17:24:43.188933 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d8cd4fdf7-74zlx" event={"ID":"b896d6a2-a26c-4b15-adcf-bb73998922f2","Type":"ContainerStarted","Data":"468bfd23b39aec2505dfb99e2b91f26af3a75f29bd3b2c5fdd537b97f41d6c82"} Feb 27 17:24:43 crc kubenswrapper[4700]: I0227 17:24:43.864129 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 27 17:24:44 crc kubenswrapper[4700]: I0227 17:24:44.198787 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d8cd4fdf7-74zlx" event={"ID":"b896d6a2-a26c-4b15-adcf-bb73998922f2","Type":"ContainerStarted","Data":"2d1317865a2ce03538491b00c99808acf65eedb1a3994f3245a6fc4013fc05b4"} Feb 27 17:24:44 crc kubenswrapper[4700]: I0227 17:24:44.198934 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ea091df4-f7f3-4580-b20e-215922e9bbed" containerName="nova-api-log" containerID="cri-o://2e18d66beced46445840352f57cc57d4534058c5afafc4fa9a24a0845342f796" gracePeriod=30 Feb 27 17:24:44 crc kubenswrapper[4700]: I0227 17:24:44.199062 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ea091df4-f7f3-4580-b20e-215922e9bbed" containerName="nova-api-api" containerID="cri-o://9abc4287de101622f56dcf027eadf4227af7525ef8bffdcd76e13612ac66591f" gracePeriod=30 Feb 27 17:24:44 crc kubenswrapper[4700]: I0227 17:24:44.230689 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d8cd4fdf7-74zlx" podStartSLOduration=3.230669635 podStartE2EDuration="3.230669635s" podCreationTimestamp="2026-02-27 17:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:24:44.224439691 +0000 UTC m=+1444.209752458" watchObservedRunningTime="2026-02-27 17:24:44.230669635 +0000 UTC m=+1444.215982382" Feb 27 17:24:44 crc kubenswrapper[4700]: I0227 17:24:44.402876 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:24:44 crc kubenswrapper[4700]: I0227 17:24:44.403117 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="095a6539-e7fd-461b-9f54-8598f8d3629d" containerName="ceilometer-central-agent" containerID="cri-o://8d0b0a95cec7cc0e1f33eb767754c2ecf9feec24a623a31e2a6558c41f8fe978" gracePeriod=30 Feb 27 17:24:44 crc kubenswrapper[4700]: I0227 17:24:44.403251 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="095a6539-e7fd-461b-9f54-8598f8d3629d" containerName="proxy-httpd" containerID="cri-o://2d8463c7e437670e8cbdad66d7b79629d957e872e6154426a17c45965ae114ce" gracePeriod=30 Feb 27 17:24:44 crc kubenswrapper[4700]: I0227 17:24:44.403289 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="095a6539-e7fd-461b-9f54-8598f8d3629d" containerName="sg-core" containerID="cri-o://4c06622be61d03967d319c5c2a56f892dbb1ee843a9880787240e7021b92cf17" gracePeriod=30 Feb 27 17:24:44 crc kubenswrapper[4700]: I0227 17:24:44.403318 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="095a6539-e7fd-461b-9f54-8598f8d3629d" containerName="ceilometer-notification-agent" containerID="cri-o://c2f7f6f3efcb17f0b73cecd62e10125cb8895e576f71bf116e0dd8435870a0d6" gracePeriod=30 Feb 27 17:24:44 crc kubenswrapper[4700]: I0227 17:24:44.848174 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 27 17:24:44 crc kubenswrapper[4700]: I0227 17:24:44.883128 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 27 17:24:44 crc kubenswrapper[4700]: I0227 17:24:44.883314 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 27 17:24:45 crc kubenswrapper[4700]: I0227 17:24:45.212627 4700 generic.go:334] "Generic (PLEG): container finished" podID="095a6539-e7fd-461b-9f54-8598f8d3629d" containerID="2d8463c7e437670e8cbdad66d7b79629d957e872e6154426a17c45965ae114ce" exitCode=0 Feb 27 17:24:45 crc kubenswrapper[4700]: I0227 17:24:45.212664 4700 generic.go:334] "Generic (PLEG): container finished" podID="095a6539-e7fd-461b-9f54-8598f8d3629d" containerID="4c06622be61d03967d319c5c2a56f892dbb1ee843a9880787240e7021b92cf17" exitCode=2 Feb 27 17:24:45 crc kubenswrapper[4700]: I0227 17:24:45.212676 4700 generic.go:334] "Generic (PLEG): container finished" podID="095a6539-e7fd-461b-9f54-8598f8d3629d" containerID="c2f7f6f3efcb17f0b73cecd62e10125cb8895e576f71bf116e0dd8435870a0d6" exitCode=0 Feb 27 17:24:45 crc kubenswrapper[4700]: I0227 17:24:45.212719 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"095a6539-e7fd-461b-9f54-8598f8d3629d","Type":"ContainerDied","Data":"2d8463c7e437670e8cbdad66d7b79629d957e872e6154426a17c45965ae114ce"} Feb 27 17:24:45 crc kubenswrapper[4700]: I0227 17:24:45.212748 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"095a6539-e7fd-461b-9f54-8598f8d3629d","Type":"ContainerDied","Data":"4c06622be61d03967d319c5c2a56f892dbb1ee843a9880787240e7021b92cf17"} Feb 27 17:24:45 crc kubenswrapper[4700]: I0227 17:24:45.212760 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"095a6539-e7fd-461b-9f54-8598f8d3629d","Type":"ContainerDied","Data":"c2f7f6f3efcb17f0b73cecd62e10125cb8895e576f71bf116e0dd8435870a0d6"} Feb 27 17:24:45 crc kubenswrapper[4700]: I0227 17:24:45.215046 4700 generic.go:334] "Generic (PLEG): container finished" podID="ea091df4-f7f3-4580-b20e-215922e9bbed" containerID="9abc4287de101622f56dcf027eadf4227af7525ef8bffdcd76e13612ac66591f" exitCode=0 Feb 27 17:24:45 crc kubenswrapper[4700]: I0227 17:24:45.215323 4700 generic.go:334] "Generic (PLEG): container finished" podID="ea091df4-f7f3-4580-b20e-215922e9bbed" containerID="2e18d66beced46445840352f57cc57d4534058c5afafc4fa9a24a0845342f796" exitCode=143 Feb 27 17:24:45 crc kubenswrapper[4700]: I0227 17:24:45.216530 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ea091df4-f7f3-4580-b20e-215922e9bbed","Type":"ContainerDied","Data":"9abc4287de101622f56dcf027eadf4227af7525ef8bffdcd76e13612ac66591f"} Feb 27 17:24:45 crc kubenswrapper[4700]: I0227 17:24:45.216575 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d8cd4fdf7-74zlx" Feb 27 17:24:45 crc kubenswrapper[4700]: I0227 17:24:45.216590 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ea091df4-f7f3-4580-b20e-215922e9bbed","Type":"ContainerDied","Data":"2e18d66beced46445840352f57cc57d4534058c5afafc4fa9a24a0845342f796"} Feb 27 17:24:45 crc kubenswrapper[4700]: I0227 17:24:45.617681 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 17:24:45 crc kubenswrapper[4700]: I0227 17:24:45.711949 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtjjb\" (UniqueName: \"kubernetes.io/projected/ea091df4-f7f3-4580-b20e-215922e9bbed-kube-api-access-qtjjb\") pod \"ea091df4-f7f3-4580-b20e-215922e9bbed\" (UID: \"ea091df4-f7f3-4580-b20e-215922e9bbed\") " Feb 27 17:24:45 crc kubenswrapper[4700]: I0227 17:24:45.712193 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea091df4-f7f3-4580-b20e-215922e9bbed-config-data\") pod \"ea091df4-f7f3-4580-b20e-215922e9bbed\" (UID: \"ea091df4-f7f3-4580-b20e-215922e9bbed\") " Feb 27 17:24:45 crc kubenswrapper[4700]: I0227 17:24:45.712253 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea091df4-f7f3-4580-b20e-215922e9bbed-combined-ca-bundle\") pod \"ea091df4-f7f3-4580-b20e-215922e9bbed\" (UID: \"ea091df4-f7f3-4580-b20e-215922e9bbed\") " Feb 27 17:24:45 crc kubenswrapper[4700]: I0227 17:24:45.712335 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea091df4-f7f3-4580-b20e-215922e9bbed-logs\") pod \"ea091df4-f7f3-4580-b20e-215922e9bbed\" (UID: \"ea091df4-f7f3-4580-b20e-215922e9bbed\") " Feb 27 17:24:45 crc kubenswrapper[4700]: I0227 17:24:45.713090 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea091df4-f7f3-4580-b20e-215922e9bbed-logs" (OuterVolumeSpecName: "logs") pod "ea091df4-f7f3-4580-b20e-215922e9bbed" (UID: "ea091df4-f7f3-4580-b20e-215922e9bbed"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:24:45 crc kubenswrapper[4700]: I0227 17:24:45.719493 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea091df4-f7f3-4580-b20e-215922e9bbed-kube-api-access-qtjjb" (OuterVolumeSpecName: "kube-api-access-qtjjb") pod "ea091df4-f7f3-4580-b20e-215922e9bbed" (UID: "ea091df4-f7f3-4580-b20e-215922e9bbed"). InnerVolumeSpecName "kube-api-access-qtjjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:24:45 crc kubenswrapper[4700]: I0227 17:24:45.741731 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea091df4-f7f3-4580-b20e-215922e9bbed-config-data" (OuterVolumeSpecName: "config-data") pod "ea091df4-f7f3-4580-b20e-215922e9bbed" (UID: "ea091df4-f7f3-4580-b20e-215922e9bbed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:45 crc kubenswrapper[4700]: I0227 17:24:45.770630 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea091df4-f7f3-4580-b20e-215922e9bbed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ea091df4-f7f3-4580-b20e-215922e9bbed" (UID: "ea091df4-f7f3-4580-b20e-215922e9bbed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:45 crc kubenswrapper[4700]: I0227 17:24:45.814107 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea091df4-f7f3-4580-b20e-215922e9bbed-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:45 crc kubenswrapper[4700]: I0227 17:24:45.814140 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea091df4-f7f3-4580-b20e-215922e9bbed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:45 crc kubenswrapper[4700]: I0227 17:24:45.814150 4700 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea091df4-f7f3-4580-b20e-215922e9bbed-logs\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:45 crc kubenswrapper[4700]: I0227 17:24:45.814161 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtjjb\" (UniqueName: \"kubernetes.io/projected/ea091df4-f7f3-4580-b20e-215922e9bbed-kube-api-access-qtjjb\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:46 crc kubenswrapper[4700]: I0227 17:24:46.228453 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ea091df4-f7f3-4580-b20e-215922e9bbed","Type":"ContainerDied","Data":"8ac15a89c5fa8e552ce59e15666209fdebb0576c8b016e8810bf5725e84fc019"} Feb 27 17:24:46 crc kubenswrapper[4700]: I0227 17:24:46.228517 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 17:24:46 crc kubenswrapper[4700]: I0227 17:24:46.228898 4700 scope.go:117] "RemoveContainer" containerID="9abc4287de101622f56dcf027eadf4227af7525ef8bffdcd76e13612ac66591f" Feb 27 17:24:46 crc kubenswrapper[4700]: I0227 17:24:46.251283 4700 scope.go:117] "RemoveContainer" containerID="2e18d66beced46445840352f57cc57d4534058c5afafc4fa9a24a0845342f796" Feb 27 17:24:46 crc kubenswrapper[4700]: I0227 17:24:46.280854 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 27 17:24:46 crc kubenswrapper[4700]: I0227 17:24:46.291148 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 27 17:24:46 crc kubenswrapper[4700]: I0227 17:24:46.325604 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 27 17:24:46 crc kubenswrapper[4700]: E0227 17:24:46.326079 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea091df4-f7f3-4580-b20e-215922e9bbed" containerName="nova-api-api" Feb 27 17:24:46 crc kubenswrapper[4700]: I0227 17:24:46.326096 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea091df4-f7f3-4580-b20e-215922e9bbed" containerName="nova-api-api" Feb 27 17:24:46 crc kubenswrapper[4700]: E0227 17:24:46.326113 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea091df4-f7f3-4580-b20e-215922e9bbed" containerName="nova-api-log" Feb 27 17:24:46 crc kubenswrapper[4700]: I0227 17:24:46.326120 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea091df4-f7f3-4580-b20e-215922e9bbed" containerName="nova-api-log" Feb 27 17:24:46 crc kubenswrapper[4700]: I0227 17:24:46.326328 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea091df4-f7f3-4580-b20e-215922e9bbed" containerName="nova-api-log" Feb 27 17:24:46 crc kubenswrapper[4700]: I0227 17:24:46.326353 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea091df4-f7f3-4580-b20e-215922e9bbed" containerName="nova-api-api" Feb 27 17:24:46 crc kubenswrapper[4700]: I0227 17:24:46.327479 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 17:24:46 crc kubenswrapper[4700]: I0227 17:24:46.332851 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 27 17:24:46 crc kubenswrapper[4700]: I0227 17:24:46.333382 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 27 17:24:46 crc kubenswrapper[4700]: I0227 17:24:46.333604 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 27 17:24:46 crc kubenswrapper[4700]: I0227 17:24:46.334287 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 27 17:24:46 crc kubenswrapper[4700]: I0227 17:24:46.437376 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa7db367-fd4d-4915-803f-bdd113ddc82b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"aa7db367-fd4d-4915-803f-bdd113ddc82b\") " pod="openstack/nova-api-0" Feb 27 17:24:46 crc kubenswrapper[4700]: I0227 17:24:46.437427 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa7db367-fd4d-4915-803f-bdd113ddc82b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"aa7db367-fd4d-4915-803f-bdd113ddc82b\") " pod="openstack/nova-api-0" Feb 27 17:24:46 crc kubenswrapper[4700]: I0227 17:24:46.437467 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa7db367-fd4d-4915-803f-bdd113ddc82b-config-data\") pod \"nova-api-0\" (UID: \"aa7db367-fd4d-4915-803f-bdd113ddc82b\") " pod="openstack/nova-api-0" Feb 27 17:24:46 crc kubenswrapper[4700]: I0227 17:24:46.437653 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa7db367-fd4d-4915-803f-bdd113ddc82b-public-tls-certs\") pod \"nova-api-0\" (UID: \"aa7db367-fd4d-4915-803f-bdd113ddc82b\") " pod="openstack/nova-api-0" Feb 27 17:24:46 crc kubenswrapper[4700]: I0227 17:24:46.437705 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmbmw\" (UniqueName: \"kubernetes.io/projected/aa7db367-fd4d-4915-803f-bdd113ddc82b-kube-api-access-mmbmw\") pod \"nova-api-0\" (UID: \"aa7db367-fd4d-4915-803f-bdd113ddc82b\") " pod="openstack/nova-api-0" Feb 27 17:24:46 crc kubenswrapper[4700]: I0227 17:24:46.437910 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa7db367-fd4d-4915-803f-bdd113ddc82b-logs\") pod \"nova-api-0\" (UID: \"aa7db367-fd4d-4915-803f-bdd113ddc82b\") " pod="openstack/nova-api-0" Feb 27 17:24:46 crc kubenswrapper[4700]: I0227 17:24:46.539804 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa7db367-fd4d-4915-803f-bdd113ddc82b-logs\") pod \"nova-api-0\" (UID: \"aa7db367-fd4d-4915-803f-bdd113ddc82b\") " pod="openstack/nova-api-0" Feb 27 17:24:46 crc kubenswrapper[4700]: I0227 17:24:46.539881 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa7db367-fd4d-4915-803f-bdd113ddc82b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"aa7db367-fd4d-4915-803f-bdd113ddc82b\") " pod="openstack/nova-api-0" Feb 27 17:24:46 crc kubenswrapper[4700]: I0227 17:24:46.539955 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa7db367-fd4d-4915-803f-bdd113ddc82b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"aa7db367-fd4d-4915-803f-bdd113ddc82b\") " pod="openstack/nova-api-0" Feb 27 17:24:46 crc kubenswrapper[4700]: I0227 17:24:46.539986 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa7db367-fd4d-4915-803f-bdd113ddc82b-config-data\") pod \"nova-api-0\" (UID: \"aa7db367-fd4d-4915-803f-bdd113ddc82b\") " pod="openstack/nova-api-0" Feb 27 17:24:46 crc kubenswrapper[4700]: I0227 17:24:46.540044 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa7db367-fd4d-4915-803f-bdd113ddc82b-public-tls-certs\") pod \"nova-api-0\" (UID: \"aa7db367-fd4d-4915-803f-bdd113ddc82b\") " pod="openstack/nova-api-0" Feb 27 17:24:46 crc kubenswrapper[4700]: I0227 17:24:46.540065 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmbmw\" (UniqueName: \"kubernetes.io/projected/aa7db367-fd4d-4915-803f-bdd113ddc82b-kube-api-access-mmbmw\") pod \"nova-api-0\" (UID: \"aa7db367-fd4d-4915-803f-bdd113ddc82b\") " pod="openstack/nova-api-0" Feb 27 17:24:46 crc kubenswrapper[4700]: I0227 17:24:46.541128 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa7db367-fd4d-4915-803f-bdd113ddc82b-logs\") pod \"nova-api-0\" (UID: \"aa7db367-fd4d-4915-803f-bdd113ddc82b\") " pod="openstack/nova-api-0" Feb 27 17:24:46 crc kubenswrapper[4700]: I0227 17:24:46.546208 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa7db367-fd4d-4915-803f-bdd113ddc82b-public-tls-certs\") pod \"nova-api-0\" (UID: \"aa7db367-fd4d-4915-803f-bdd113ddc82b\") " pod="openstack/nova-api-0" Feb 27 17:24:46 crc kubenswrapper[4700]: I0227 17:24:46.546880 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa7db367-fd4d-4915-803f-bdd113ddc82b-config-data\") pod \"nova-api-0\" (UID: \"aa7db367-fd4d-4915-803f-bdd113ddc82b\") " pod="openstack/nova-api-0" Feb 27 17:24:46 crc kubenswrapper[4700]: I0227 17:24:46.554984 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa7db367-fd4d-4915-803f-bdd113ddc82b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"aa7db367-fd4d-4915-803f-bdd113ddc82b\") " pod="openstack/nova-api-0" Feb 27 17:24:46 crc kubenswrapper[4700]: I0227 17:24:46.559595 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa7db367-fd4d-4915-803f-bdd113ddc82b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"aa7db367-fd4d-4915-803f-bdd113ddc82b\") " pod="openstack/nova-api-0" Feb 27 17:24:46 crc kubenswrapper[4700]: I0227 17:24:46.567109 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmbmw\" (UniqueName: \"kubernetes.io/projected/aa7db367-fd4d-4915-803f-bdd113ddc82b-kube-api-access-mmbmw\") pod \"nova-api-0\" (UID: \"aa7db367-fd4d-4915-803f-bdd113ddc82b\") " pod="openstack/nova-api-0" Feb 27 17:24:46 crc kubenswrapper[4700]: I0227 17:24:46.686781 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 17:24:47 crc kubenswrapper[4700]: I0227 17:24:47.029516 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea091df4-f7f3-4580-b20e-215922e9bbed" path="/var/lib/kubelet/pods/ea091df4-f7f3-4580-b20e-215922e9bbed/volumes" Feb 27 17:24:47 crc kubenswrapper[4700]: I0227 17:24:47.234207 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 27 17:24:47 crc kubenswrapper[4700]: I0227 17:24:47.247870 4700 generic.go:334] "Generic (PLEG): container finished" podID="095a6539-e7fd-461b-9f54-8598f8d3629d" containerID="8d0b0a95cec7cc0e1f33eb767754c2ecf9feec24a623a31e2a6558c41f8fe978" exitCode=0 Feb 27 17:24:47 crc kubenswrapper[4700]: I0227 17:24:47.247973 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"095a6539-e7fd-461b-9f54-8598f8d3629d","Type":"ContainerDied","Data":"8d0b0a95cec7cc0e1f33eb767754c2ecf9feec24a623a31e2a6558c41f8fe978"} Feb 27 17:24:47 crc kubenswrapper[4700]: I0227 17:24:47.691600 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 17:24:47 crc kubenswrapper[4700]: I0227 17:24:47.783778 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/095a6539-e7fd-461b-9f54-8598f8d3629d-ceilometer-tls-certs\") pod \"095a6539-e7fd-461b-9f54-8598f8d3629d\" (UID: \"095a6539-e7fd-461b-9f54-8598f8d3629d\") " Feb 27 17:24:47 crc kubenswrapper[4700]: I0227 17:24:47.783888 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/095a6539-e7fd-461b-9f54-8598f8d3629d-run-httpd\") pod \"095a6539-e7fd-461b-9f54-8598f8d3629d\" (UID: \"095a6539-e7fd-461b-9f54-8598f8d3629d\") " Feb 27 17:24:47 crc kubenswrapper[4700]: I0227 17:24:47.783924 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/095a6539-e7fd-461b-9f54-8598f8d3629d-config-data\") pod \"095a6539-e7fd-461b-9f54-8598f8d3629d\" (UID: \"095a6539-e7fd-461b-9f54-8598f8d3629d\") " Feb 27 17:24:47 crc kubenswrapper[4700]: I0227 17:24:47.783973 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/095a6539-e7fd-461b-9f54-8598f8d3629d-scripts\") pod \"095a6539-e7fd-461b-9f54-8598f8d3629d\" (UID: \"095a6539-e7fd-461b-9f54-8598f8d3629d\") " Feb 27 17:24:47 crc kubenswrapper[4700]: I0227 17:24:47.784007 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4bwb\" (UniqueName: \"kubernetes.io/projected/095a6539-e7fd-461b-9f54-8598f8d3629d-kube-api-access-k4bwb\") pod \"095a6539-e7fd-461b-9f54-8598f8d3629d\" (UID: \"095a6539-e7fd-461b-9f54-8598f8d3629d\") " Feb 27 17:24:47 crc kubenswrapper[4700]: I0227 17:24:47.784029 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/095a6539-e7fd-461b-9f54-8598f8d3629d-log-httpd\") pod \"095a6539-e7fd-461b-9f54-8598f8d3629d\" (UID: \"095a6539-e7fd-461b-9f54-8598f8d3629d\") " Feb 27 17:24:47 crc kubenswrapper[4700]: I0227 17:24:47.784144 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/095a6539-e7fd-461b-9f54-8598f8d3629d-combined-ca-bundle\") pod \"095a6539-e7fd-461b-9f54-8598f8d3629d\" (UID: \"095a6539-e7fd-461b-9f54-8598f8d3629d\") " Feb 27 17:24:47 crc kubenswrapper[4700]: I0227 17:24:47.784167 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/095a6539-e7fd-461b-9f54-8598f8d3629d-sg-core-conf-yaml\") pod \"095a6539-e7fd-461b-9f54-8598f8d3629d\" (UID: \"095a6539-e7fd-461b-9f54-8598f8d3629d\") " Feb 27 17:24:47 crc kubenswrapper[4700]: I0227 17:24:47.784586 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/095a6539-e7fd-461b-9f54-8598f8d3629d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "095a6539-e7fd-461b-9f54-8598f8d3629d" (UID: "095a6539-e7fd-461b-9f54-8598f8d3629d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:24:47 crc kubenswrapper[4700]: I0227 17:24:47.784606 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/095a6539-e7fd-461b-9f54-8598f8d3629d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "095a6539-e7fd-461b-9f54-8598f8d3629d" (UID: "095a6539-e7fd-461b-9f54-8598f8d3629d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:24:47 crc kubenswrapper[4700]: I0227 17:24:47.788526 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/095a6539-e7fd-461b-9f54-8598f8d3629d-scripts" (OuterVolumeSpecName: "scripts") pod "095a6539-e7fd-461b-9f54-8598f8d3629d" (UID: "095a6539-e7fd-461b-9f54-8598f8d3629d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:47 crc kubenswrapper[4700]: I0227 17:24:47.788593 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/095a6539-e7fd-461b-9f54-8598f8d3629d-kube-api-access-k4bwb" (OuterVolumeSpecName: "kube-api-access-k4bwb") pod "095a6539-e7fd-461b-9f54-8598f8d3629d" (UID: "095a6539-e7fd-461b-9f54-8598f8d3629d"). InnerVolumeSpecName "kube-api-access-k4bwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:24:47 crc kubenswrapper[4700]: I0227 17:24:47.830049 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/095a6539-e7fd-461b-9f54-8598f8d3629d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "095a6539-e7fd-461b-9f54-8598f8d3629d" (UID: "095a6539-e7fd-461b-9f54-8598f8d3629d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:47 crc kubenswrapper[4700]: I0227 17:24:47.852972 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/095a6539-e7fd-461b-9f54-8598f8d3629d-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "095a6539-e7fd-461b-9f54-8598f8d3629d" (UID: "095a6539-e7fd-461b-9f54-8598f8d3629d"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:47 crc kubenswrapper[4700]: I0227 17:24:47.885836 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/095a6539-e7fd-461b-9f54-8598f8d3629d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "095a6539-e7fd-461b-9f54-8598f8d3629d" (UID: "095a6539-e7fd-461b-9f54-8598f8d3629d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:47 crc kubenswrapper[4700]: I0227 17:24:47.886608 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/095a6539-e7fd-461b-9f54-8598f8d3629d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:47 crc kubenswrapper[4700]: I0227 17:24:47.886670 4700 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/095a6539-e7fd-461b-9f54-8598f8d3629d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:47 crc kubenswrapper[4700]: I0227 17:24:47.886679 4700 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/095a6539-e7fd-461b-9f54-8598f8d3629d-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:47 crc kubenswrapper[4700]: I0227 17:24:47.886687 4700 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/095a6539-e7fd-461b-9f54-8598f8d3629d-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:47 crc kubenswrapper[4700]: I0227 17:24:47.886696 4700 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/095a6539-e7fd-461b-9f54-8598f8d3629d-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:47 crc kubenswrapper[4700]: I0227 17:24:47.886704 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4bwb\" (UniqueName: \"kubernetes.io/projected/095a6539-e7fd-461b-9f54-8598f8d3629d-kube-api-access-k4bwb\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:47 crc kubenswrapper[4700]: I0227 17:24:47.886713 4700 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/095a6539-e7fd-461b-9f54-8598f8d3629d-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:47 crc kubenswrapper[4700]: I0227 17:24:47.892603 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/095a6539-e7fd-461b-9f54-8598f8d3629d-config-data" (OuterVolumeSpecName: "config-data") pod "095a6539-e7fd-461b-9f54-8598f8d3629d" (UID: "095a6539-e7fd-461b-9f54-8598f8d3629d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:47 crc kubenswrapper[4700]: I0227 17:24:47.988603 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/095a6539-e7fd-461b-9f54-8598f8d3629d-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.266168 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"aa7db367-fd4d-4915-803f-bdd113ddc82b","Type":"ContainerStarted","Data":"b49c8d5c91529dc1ecfa9452fcd09735ab3f7589725e3dc78a86b3ac9064a73a"} Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.266688 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"aa7db367-fd4d-4915-803f-bdd113ddc82b","Type":"ContainerStarted","Data":"6fb577ceb4ebed75979c2a5787d8e9babf6acb979630029fa2e18c525a970734"} Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.266713 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"aa7db367-fd4d-4915-803f-bdd113ddc82b","Type":"ContainerStarted","Data":"48a76d5ea18f0eed6ebc338a256fb081f461a39b6d50b87adde3d3a6e35c66a0"} Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.280912 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"095a6539-e7fd-461b-9f54-8598f8d3629d","Type":"ContainerDied","Data":"90bb3577ee808421701f78ac212a5f545b9753906a757383904bbfb706c80a87"} Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.281166 4700 scope.go:117] "RemoveContainer" containerID="2d8463c7e437670e8cbdad66d7b79629d957e872e6154426a17c45965ae114ce" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.280987 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.308848 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.30882072 podStartE2EDuration="2.30882072s" podCreationTimestamp="2026-02-27 17:24:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:24:48.294811401 +0000 UTC m=+1448.280124158" watchObservedRunningTime="2026-02-27 17:24:48.30882072 +0000 UTC m=+1448.294133467" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.321885 4700 scope.go:117] "RemoveContainer" containerID="4c06622be61d03967d319c5c2a56f892dbb1ee843a9880787240e7021b92cf17" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.350876 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.363806 4700 scope.go:117] "RemoveContainer" containerID="c2f7f6f3efcb17f0b73cecd62e10125cb8895e576f71bf116e0dd8435870a0d6" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.365879 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.375365 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:24:48 crc kubenswrapper[4700]: E0227 17:24:48.375889 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="095a6539-e7fd-461b-9f54-8598f8d3629d" containerName="ceilometer-notification-agent" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.375911 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="095a6539-e7fd-461b-9f54-8598f8d3629d" containerName="ceilometer-notification-agent" Feb 27 17:24:48 crc kubenswrapper[4700]: E0227 17:24:48.375932 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="095a6539-e7fd-461b-9f54-8598f8d3629d" containerName="proxy-httpd" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.375943 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="095a6539-e7fd-461b-9f54-8598f8d3629d" containerName="proxy-httpd" Feb 27 17:24:48 crc kubenswrapper[4700]: E0227 17:24:48.375972 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="095a6539-e7fd-461b-9f54-8598f8d3629d" containerName="sg-core" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.375980 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="095a6539-e7fd-461b-9f54-8598f8d3629d" containerName="sg-core" Feb 27 17:24:48 crc kubenswrapper[4700]: E0227 17:24:48.376000 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="095a6539-e7fd-461b-9f54-8598f8d3629d" containerName="ceilometer-central-agent" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.376010 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="095a6539-e7fd-461b-9f54-8598f8d3629d" containerName="ceilometer-central-agent" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.376225 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="095a6539-e7fd-461b-9f54-8598f8d3629d" containerName="ceilometer-central-agent" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.376264 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="095a6539-e7fd-461b-9f54-8598f8d3629d" containerName="sg-core" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.376275 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="095a6539-e7fd-461b-9f54-8598f8d3629d" containerName="proxy-httpd" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.376301 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="095a6539-e7fd-461b-9f54-8598f8d3629d" containerName="ceilometer-notification-agent" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.382660 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.382768 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.387940 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.388196 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.388228 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.391758 4700 scope.go:117] "RemoveContainer" containerID="8d0b0a95cec7cc0e1f33eb767754c2ecf9feec24a623a31e2a6558c41f8fe978" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.504773 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6c62q\" (UniqueName: \"kubernetes.io/projected/0a50deb6-9983-40f1-bf2c-f8ce7584e809-kube-api-access-6c62q\") pod \"ceilometer-0\" (UID: \"0a50deb6-9983-40f1-bf2c-f8ce7584e809\") " pod="openstack/ceilometer-0" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.504830 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a50deb6-9983-40f1-bf2c-f8ce7584e809-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0a50deb6-9983-40f1-bf2c-f8ce7584e809\") " pod="openstack/ceilometer-0" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.504873 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a50deb6-9983-40f1-bf2c-f8ce7584e809-log-httpd\") pod \"ceilometer-0\" (UID: \"0a50deb6-9983-40f1-bf2c-f8ce7584e809\") " pod="openstack/ceilometer-0" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.504914 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a50deb6-9983-40f1-bf2c-f8ce7584e809-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0a50deb6-9983-40f1-bf2c-f8ce7584e809\") " pod="openstack/ceilometer-0" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.504936 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a50deb6-9983-40f1-bf2c-f8ce7584e809-config-data\") pod \"ceilometer-0\" (UID: \"0a50deb6-9983-40f1-bf2c-f8ce7584e809\") " pod="openstack/ceilometer-0" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.504980 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a50deb6-9983-40f1-bf2c-f8ce7584e809-scripts\") pod \"ceilometer-0\" (UID: \"0a50deb6-9983-40f1-bf2c-f8ce7584e809\") " pod="openstack/ceilometer-0" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.505005 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a50deb6-9983-40f1-bf2c-f8ce7584e809-run-httpd\") pod \"ceilometer-0\" (UID: \"0a50deb6-9983-40f1-bf2c-f8ce7584e809\") " pod="openstack/ceilometer-0" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.505070 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a50deb6-9983-40f1-bf2c-f8ce7584e809-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0a50deb6-9983-40f1-bf2c-f8ce7584e809\") " pod="openstack/ceilometer-0" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.607047 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a50deb6-9983-40f1-bf2c-f8ce7584e809-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0a50deb6-9983-40f1-bf2c-f8ce7584e809\") " pod="openstack/ceilometer-0" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.607136 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6c62q\" (UniqueName: \"kubernetes.io/projected/0a50deb6-9983-40f1-bf2c-f8ce7584e809-kube-api-access-6c62q\") pod \"ceilometer-0\" (UID: \"0a50deb6-9983-40f1-bf2c-f8ce7584e809\") " pod="openstack/ceilometer-0" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.607166 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a50deb6-9983-40f1-bf2c-f8ce7584e809-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0a50deb6-9983-40f1-bf2c-f8ce7584e809\") " pod="openstack/ceilometer-0" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.607207 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a50deb6-9983-40f1-bf2c-f8ce7584e809-log-httpd\") pod \"ceilometer-0\" (UID: \"0a50deb6-9983-40f1-bf2c-f8ce7584e809\") " pod="openstack/ceilometer-0" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.607244 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a50deb6-9983-40f1-bf2c-f8ce7584e809-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0a50deb6-9983-40f1-bf2c-f8ce7584e809\") " pod="openstack/ceilometer-0" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.607263 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a50deb6-9983-40f1-bf2c-f8ce7584e809-config-data\") pod \"ceilometer-0\" (UID: \"0a50deb6-9983-40f1-bf2c-f8ce7584e809\") " pod="openstack/ceilometer-0" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.607309 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a50deb6-9983-40f1-bf2c-f8ce7584e809-scripts\") pod \"ceilometer-0\" (UID: \"0a50deb6-9983-40f1-bf2c-f8ce7584e809\") " pod="openstack/ceilometer-0" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.607335 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a50deb6-9983-40f1-bf2c-f8ce7584e809-run-httpd\") pod \"ceilometer-0\" (UID: \"0a50deb6-9983-40f1-bf2c-f8ce7584e809\") " pod="openstack/ceilometer-0" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.608313 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a50deb6-9983-40f1-bf2c-f8ce7584e809-log-httpd\") pod \"ceilometer-0\" (UID: \"0a50deb6-9983-40f1-bf2c-f8ce7584e809\") " pod="openstack/ceilometer-0" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.608322 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a50deb6-9983-40f1-bf2c-f8ce7584e809-run-httpd\") pod \"ceilometer-0\" (UID: \"0a50deb6-9983-40f1-bf2c-f8ce7584e809\") " pod="openstack/ceilometer-0" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.614564 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a50deb6-9983-40f1-bf2c-f8ce7584e809-config-data\") pod \"ceilometer-0\" (UID: \"0a50deb6-9983-40f1-bf2c-f8ce7584e809\") " pod="openstack/ceilometer-0" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.621730 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a50deb6-9983-40f1-bf2c-f8ce7584e809-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0a50deb6-9983-40f1-bf2c-f8ce7584e809\") " pod="openstack/ceilometer-0" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.622456 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a50deb6-9983-40f1-bf2c-f8ce7584e809-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0a50deb6-9983-40f1-bf2c-f8ce7584e809\") " pod="openstack/ceilometer-0" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.627037 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a50deb6-9983-40f1-bf2c-f8ce7584e809-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0a50deb6-9983-40f1-bf2c-f8ce7584e809\") " pod="openstack/ceilometer-0" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.635995 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a50deb6-9983-40f1-bf2c-f8ce7584e809-scripts\") pod \"ceilometer-0\" (UID: \"0a50deb6-9983-40f1-bf2c-f8ce7584e809\") " pod="openstack/ceilometer-0" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.644207 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6c62q\" (UniqueName: \"kubernetes.io/projected/0a50deb6-9983-40f1-bf2c-f8ce7584e809-kube-api-access-6c62q\") pod \"ceilometer-0\" (UID: \"0a50deb6-9983-40f1-bf2c-f8ce7584e809\") " pod="openstack/ceilometer-0" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.704241 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 17:24:48 crc kubenswrapper[4700]: I0227 17:24:48.991679 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="095a6539-e7fd-461b-9f54-8598f8d3629d" path="/var/lib/kubelet/pods/095a6539-e7fd-461b-9f54-8598f8d3629d/volumes" Feb 27 17:24:49 crc kubenswrapper[4700]: I0227 17:24:49.209992 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 17:24:49 crc kubenswrapper[4700]: W0227 17:24:49.210665 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a50deb6_9983_40f1_bf2c_f8ce7584e809.slice/crio-7ff937446246ea0e5d24417dc7db76138071af69236b1b5fc8d9adda346dca3d WatchSource:0}: Error finding container 7ff937446246ea0e5d24417dc7db76138071af69236b1b5fc8d9adda346dca3d: Status 404 returned error can't find the container with id 7ff937446246ea0e5d24417dc7db76138071af69236b1b5fc8d9adda346dca3d Feb 27 17:24:49 crc kubenswrapper[4700]: I0227 17:24:49.294162 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a50deb6-9983-40f1-bf2c-f8ce7584e809","Type":"ContainerStarted","Data":"7ff937446246ea0e5d24417dc7db76138071af69236b1b5fc8d9adda346dca3d"} Feb 27 17:24:49 crc kubenswrapper[4700]: I0227 17:24:49.848136 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Feb 27 17:24:49 crc kubenswrapper[4700]: I0227 17:24:49.871155 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Feb 27 17:24:49 crc kubenswrapper[4700]: I0227 17:24:49.883217 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 27 17:24:49 crc kubenswrapper[4700]: I0227 17:24:49.883272 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 27 17:24:50 crc kubenswrapper[4700]: I0227 17:24:50.310666 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a50deb6-9983-40f1-bf2c-f8ce7584e809","Type":"ContainerStarted","Data":"8a2eaeba8408e444f6fc8624878ea574e19cf802c026e7b69c93814e26028409"} Feb 27 17:24:50 crc kubenswrapper[4700]: I0227 17:24:50.310744 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a50deb6-9983-40f1-bf2c-f8ce7584e809","Type":"ContainerStarted","Data":"2c30263af4bf57a3c0b2c9a34c54fd66f9b9aeaca3163c644e80399eaa73104f"} Feb 27 17:24:50 crc kubenswrapper[4700]: I0227 17:24:50.326120 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Feb 27 17:24:50 crc kubenswrapper[4700]: I0227 17:24:50.553944 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-9s776"] Feb 27 17:24:50 crc kubenswrapper[4700]: I0227 17:24:50.555544 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-9s776" Feb 27 17:24:50 crc kubenswrapper[4700]: I0227 17:24:50.569916 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Feb 27 17:24:50 crc kubenswrapper[4700]: I0227 17:24:50.569916 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Feb 27 17:24:50 crc kubenswrapper[4700]: I0227 17:24:50.579080 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-9s776"] Feb 27 17:24:50 crc kubenswrapper[4700]: I0227 17:24:50.654653 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5e54324-9f69-4bc9-a6af-02b5ca6445e8-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-9s776\" (UID: \"e5e54324-9f69-4bc9-a6af-02b5ca6445e8\") " pod="openstack/nova-cell1-cell-mapping-9s776" Feb 27 17:24:50 crc kubenswrapper[4700]: I0227 17:24:50.654823 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rb9lz\" (UniqueName: \"kubernetes.io/projected/e5e54324-9f69-4bc9-a6af-02b5ca6445e8-kube-api-access-rb9lz\") pod \"nova-cell1-cell-mapping-9s776\" (UID: \"e5e54324-9f69-4bc9-a6af-02b5ca6445e8\") " pod="openstack/nova-cell1-cell-mapping-9s776" Feb 27 17:24:50 crc kubenswrapper[4700]: I0227 17:24:50.654916 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5e54324-9f69-4bc9-a6af-02b5ca6445e8-scripts\") pod \"nova-cell1-cell-mapping-9s776\" (UID: \"e5e54324-9f69-4bc9-a6af-02b5ca6445e8\") " pod="openstack/nova-cell1-cell-mapping-9s776" Feb 27 17:24:50 crc kubenswrapper[4700]: I0227 17:24:50.654955 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5e54324-9f69-4bc9-a6af-02b5ca6445e8-config-data\") pod \"nova-cell1-cell-mapping-9s776\" (UID: \"e5e54324-9f69-4bc9-a6af-02b5ca6445e8\") " pod="openstack/nova-cell1-cell-mapping-9s776" Feb 27 17:24:50 crc kubenswrapper[4700]: I0227 17:24:50.757391 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rb9lz\" (UniqueName: \"kubernetes.io/projected/e5e54324-9f69-4bc9-a6af-02b5ca6445e8-kube-api-access-rb9lz\") pod \"nova-cell1-cell-mapping-9s776\" (UID: \"e5e54324-9f69-4bc9-a6af-02b5ca6445e8\") " pod="openstack/nova-cell1-cell-mapping-9s776" Feb 27 17:24:50 crc kubenswrapper[4700]: I0227 17:24:50.757540 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5e54324-9f69-4bc9-a6af-02b5ca6445e8-scripts\") pod \"nova-cell1-cell-mapping-9s776\" (UID: \"e5e54324-9f69-4bc9-a6af-02b5ca6445e8\") " pod="openstack/nova-cell1-cell-mapping-9s776" Feb 27 17:24:50 crc kubenswrapper[4700]: I0227 17:24:50.757583 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5e54324-9f69-4bc9-a6af-02b5ca6445e8-config-data\") pod \"nova-cell1-cell-mapping-9s776\" (UID: \"e5e54324-9f69-4bc9-a6af-02b5ca6445e8\") " pod="openstack/nova-cell1-cell-mapping-9s776" Feb 27 17:24:50 crc kubenswrapper[4700]: I0227 17:24:50.757732 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5e54324-9f69-4bc9-a6af-02b5ca6445e8-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-9s776\" (UID: \"e5e54324-9f69-4bc9-a6af-02b5ca6445e8\") " pod="openstack/nova-cell1-cell-mapping-9s776" Feb 27 17:24:50 crc kubenswrapper[4700]: I0227 17:24:50.764196 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5e54324-9f69-4bc9-a6af-02b5ca6445e8-config-data\") pod \"nova-cell1-cell-mapping-9s776\" (UID: \"e5e54324-9f69-4bc9-a6af-02b5ca6445e8\") " pod="openstack/nova-cell1-cell-mapping-9s776" Feb 27 17:24:50 crc kubenswrapper[4700]: I0227 17:24:50.774166 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5e54324-9f69-4bc9-a6af-02b5ca6445e8-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-9s776\" (UID: \"e5e54324-9f69-4bc9-a6af-02b5ca6445e8\") " pod="openstack/nova-cell1-cell-mapping-9s776" Feb 27 17:24:50 crc kubenswrapper[4700]: I0227 17:24:50.774559 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5e54324-9f69-4bc9-a6af-02b5ca6445e8-scripts\") pod \"nova-cell1-cell-mapping-9s776\" (UID: \"e5e54324-9f69-4bc9-a6af-02b5ca6445e8\") " pod="openstack/nova-cell1-cell-mapping-9s776" Feb 27 17:24:50 crc kubenswrapper[4700]: I0227 17:24:50.793990 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rb9lz\" (UniqueName: \"kubernetes.io/projected/e5e54324-9f69-4bc9-a6af-02b5ca6445e8-kube-api-access-rb9lz\") pod \"nova-cell1-cell-mapping-9s776\" (UID: \"e5e54324-9f69-4bc9-a6af-02b5ca6445e8\") " pod="openstack/nova-cell1-cell-mapping-9s776" Feb 27 17:24:50 crc kubenswrapper[4700]: I0227 17:24:50.887718 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-9s776" Feb 27 17:24:50 crc kubenswrapper[4700]: I0227 17:24:50.898679 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="fb9499b9-4b5c-4f5a-9579-e3f7329f36a7" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.228:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 27 17:24:50 crc kubenswrapper[4700]: I0227 17:24:50.898687 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="fb9499b9-4b5c-4f5a-9579-e3f7329f36a7" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.228:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 27 17:24:51 crc kubenswrapper[4700]: I0227 17:24:51.330640 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a50deb6-9983-40f1-bf2c-f8ce7584e809","Type":"ContainerStarted","Data":"2bf189c424fa8af4383039c66fb9837821ed198e9cc95b7b7053e1ed1cfc166f"} Feb 27 17:24:51 crc kubenswrapper[4700]: I0227 17:24:51.380529 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-9s776"] Feb 27 17:24:51 crc kubenswrapper[4700]: W0227 17:24:51.382639 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5e54324_9f69_4bc9_a6af_02b5ca6445e8.slice/crio-72dd36b34bb6c86c99d9768649ad846189264bc5c5dd905119b3533b33eeb62a WatchSource:0}: Error finding container 72dd36b34bb6c86c99d9768649ad846189264bc5c5dd905119b3533b33eeb62a: Status 404 returned error can't find the container with id 72dd36b34bb6c86c99d9768649ad846189264bc5c5dd905119b3533b33eeb62a Feb 27 17:24:51 crc kubenswrapper[4700]: I0227 17:24:51.670664 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5d8cd4fdf7-74zlx" Feb 27 17:24:51 crc kubenswrapper[4700]: I0227 17:24:51.731591 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bb48d97cc-t2pb6"] Feb 27 17:24:51 crc kubenswrapper[4700]: I0227 17:24:51.731825 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bb48d97cc-t2pb6" podUID="884d838f-ded4-4ef7-92d5-7a6a1fcdecf2" containerName="dnsmasq-dns" containerID="cri-o://a18083759d87601fd4c2f06b18c7f74979cf39abb7274d034a6781bb0f2eebc4" gracePeriod=10 Feb 27 17:24:52 crc kubenswrapper[4700]: I0227 17:24:52.355944 4700 generic.go:334] "Generic (PLEG): container finished" podID="884d838f-ded4-4ef7-92d5-7a6a1fcdecf2" containerID="a18083759d87601fd4c2f06b18c7f74979cf39abb7274d034a6781bb0f2eebc4" exitCode=0 Feb 27 17:24:52 crc kubenswrapper[4700]: I0227 17:24:52.356286 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bb48d97cc-t2pb6" event={"ID":"884d838f-ded4-4ef7-92d5-7a6a1fcdecf2","Type":"ContainerDied","Data":"a18083759d87601fd4c2f06b18c7f74979cf39abb7274d034a6781bb0f2eebc4"} Feb 27 17:24:52 crc kubenswrapper[4700]: I0227 17:24:52.356310 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bb48d97cc-t2pb6" event={"ID":"884d838f-ded4-4ef7-92d5-7a6a1fcdecf2","Type":"ContainerDied","Data":"95a2721b3473ea4df6ef723bbaf8aede130791e2c0e1a34de3b67bf5106ad20d"} Feb 27 17:24:52 crc kubenswrapper[4700]: I0227 17:24:52.356323 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="95a2721b3473ea4df6ef723bbaf8aede130791e2c0e1a34de3b67bf5106ad20d" Feb 27 17:24:52 crc kubenswrapper[4700]: I0227 17:24:52.357723 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-9s776" event={"ID":"e5e54324-9f69-4bc9-a6af-02b5ca6445e8","Type":"ContainerStarted","Data":"76cd2335a654dbd228865b5835dd72e39a3fe261f701ccccfeed3f61f4f4fbf2"} Feb 27 17:24:52 crc kubenswrapper[4700]: I0227 17:24:52.357746 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-9s776" event={"ID":"e5e54324-9f69-4bc9-a6af-02b5ca6445e8","Type":"ContainerStarted","Data":"72dd36b34bb6c86c99d9768649ad846189264bc5c5dd905119b3533b33eeb62a"} Feb 27 17:24:52 crc kubenswrapper[4700]: I0227 17:24:52.376958 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-9s776" podStartSLOduration=2.3769434609999998 podStartE2EDuration="2.376943461s" podCreationTimestamp="2026-02-27 17:24:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:24:52.374837265 +0000 UTC m=+1452.360150012" watchObservedRunningTime="2026-02-27 17:24:52.376943461 +0000 UTC m=+1452.362256208" Feb 27 17:24:52 crc kubenswrapper[4700]: I0227 17:24:52.410489 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bb48d97cc-t2pb6" Feb 27 17:24:52 crc kubenswrapper[4700]: I0227 17:24:52.498091 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7c7j\" (UniqueName: \"kubernetes.io/projected/884d838f-ded4-4ef7-92d5-7a6a1fcdecf2-kube-api-access-c7c7j\") pod \"884d838f-ded4-4ef7-92d5-7a6a1fcdecf2\" (UID: \"884d838f-ded4-4ef7-92d5-7a6a1fcdecf2\") " Feb 27 17:24:52 crc kubenswrapper[4700]: I0227 17:24:52.498176 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/884d838f-ded4-4ef7-92d5-7a6a1fcdecf2-config\") pod \"884d838f-ded4-4ef7-92d5-7a6a1fcdecf2\" (UID: \"884d838f-ded4-4ef7-92d5-7a6a1fcdecf2\") " Feb 27 17:24:52 crc kubenswrapper[4700]: I0227 17:24:52.498311 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/884d838f-ded4-4ef7-92d5-7a6a1fcdecf2-dns-svc\") pod \"884d838f-ded4-4ef7-92d5-7a6a1fcdecf2\" (UID: \"884d838f-ded4-4ef7-92d5-7a6a1fcdecf2\") " Feb 27 17:24:52 crc kubenswrapper[4700]: I0227 17:24:52.498357 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/884d838f-ded4-4ef7-92d5-7a6a1fcdecf2-ovsdbserver-sb\") pod \"884d838f-ded4-4ef7-92d5-7a6a1fcdecf2\" (UID: \"884d838f-ded4-4ef7-92d5-7a6a1fcdecf2\") " Feb 27 17:24:52 crc kubenswrapper[4700]: I0227 17:24:52.498389 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/884d838f-ded4-4ef7-92d5-7a6a1fcdecf2-dns-swift-storage-0\") pod \"884d838f-ded4-4ef7-92d5-7a6a1fcdecf2\" (UID: \"884d838f-ded4-4ef7-92d5-7a6a1fcdecf2\") " Feb 27 17:24:52 crc kubenswrapper[4700]: I0227 17:24:52.498465 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/884d838f-ded4-4ef7-92d5-7a6a1fcdecf2-ovsdbserver-nb\") pod \"884d838f-ded4-4ef7-92d5-7a6a1fcdecf2\" (UID: \"884d838f-ded4-4ef7-92d5-7a6a1fcdecf2\") " Feb 27 17:24:52 crc kubenswrapper[4700]: I0227 17:24:52.514720 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/884d838f-ded4-4ef7-92d5-7a6a1fcdecf2-kube-api-access-c7c7j" (OuterVolumeSpecName: "kube-api-access-c7c7j") pod "884d838f-ded4-4ef7-92d5-7a6a1fcdecf2" (UID: "884d838f-ded4-4ef7-92d5-7a6a1fcdecf2"). InnerVolumeSpecName "kube-api-access-c7c7j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:24:52 crc kubenswrapper[4700]: I0227 17:24:52.549992 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/884d838f-ded4-4ef7-92d5-7a6a1fcdecf2-config" (OuterVolumeSpecName: "config") pod "884d838f-ded4-4ef7-92d5-7a6a1fcdecf2" (UID: "884d838f-ded4-4ef7-92d5-7a6a1fcdecf2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:24:52 crc kubenswrapper[4700]: I0227 17:24:52.551582 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/884d838f-ded4-4ef7-92d5-7a6a1fcdecf2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "884d838f-ded4-4ef7-92d5-7a6a1fcdecf2" (UID: "884d838f-ded4-4ef7-92d5-7a6a1fcdecf2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:24:52 crc kubenswrapper[4700]: I0227 17:24:52.571902 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/884d838f-ded4-4ef7-92d5-7a6a1fcdecf2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "884d838f-ded4-4ef7-92d5-7a6a1fcdecf2" (UID: "884d838f-ded4-4ef7-92d5-7a6a1fcdecf2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:24:52 crc kubenswrapper[4700]: I0227 17:24:52.579932 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/884d838f-ded4-4ef7-92d5-7a6a1fcdecf2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "884d838f-ded4-4ef7-92d5-7a6a1fcdecf2" (UID: "884d838f-ded4-4ef7-92d5-7a6a1fcdecf2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:24:52 crc kubenswrapper[4700]: I0227 17:24:52.580842 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/884d838f-ded4-4ef7-92d5-7a6a1fcdecf2-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "884d838f-ded4-4ef7-92d5-7a6a1fcdecf2" (UID: "884d838f-ded4-4ef7-92d5-7a6a1fcdecf2"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:24:52 crc kubenswrapper[4700]: I0227 17:24:52.601059 4700 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/884d838f-ded4-4ef7-92d5-7a6a1fcdecf2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:52 crc kubenswrapper[4700]: I0227 17:24:52.601094 4700 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/884d838f-ded4-4ef7-92d5-7a6a1fcdecf2-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:52 crc kubenswrapper[4700]: I0227 17:24:52.601104 4700 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/884d838f-ded4-4ef7-92d5-7a6a1fcdecf2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:52 crc kubenswrapper[4700]: I0227 17:24:52.601113 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7c7j\" (UniqueName: \"kubernetes.io/projected/884d838f-ded4-4ef7-92d5-7a6a1fcdecf2-kube-api-access-c7c7j\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:52 crc kubenswrapper[4700]: I0227 17:24:52.601124 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/884d838f-ded4-4ef7-92d5-7a6a1fcdecf2-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:52 crc kubenswrapper[4700]: I0227 17:24:52.601134 4700 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/884d838f-ded4-4ef7-92d5-7a6a1fcdecf2-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:53 crc kubenswrapper[4700]: I0227 17:24:53.381703 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a50deb6-9983-40f1-bf2c-f8ce7584e809","Type":"ContainerStarted","Data":"74ff301207577114be10705fe0d035002b8d8d0c03c3106671759b5a019a77cd"} Feb 27 17:24:53 crc kubenswrapper[4700]: I0227 17:24:53.382142 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bb48d97cc-t2pb6" Feb 27 17:24:53 crc kubenswrapper[4700]: I0227 17:24:53.382950 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 27 17:24:53 crc kubenswrapper[4700]: I0227 17:24:53.415174 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.8974665499999999 podStartE2EDuration="5.415154764s" podCreationTimestamp="2026-02-27 17:24:48 +0000 UTC" firstStartedPulling="2026-02-27 17:24:49.22235779 +0000 UTC m=+1449.207670537" lastFinishedPulling="2026-02-27 17:24:52.740045994 +0000 UTC m=+1452.725358751" observedRunningTime="2026-02-27 17:24:53.403792725 +0000 UTC m=+1453.389105472" watchObservedRunningTime="2026-02-27 17:24:53.415154764 +0000 UTC m=+1453.400467511" Feb 27 17:24:53 crc kubenswrapper[4700]: I0227 17:24:53.436828 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bb48d97cc-t2pb6"] Feb 27 17:24:53 crc kubenswrapper[4700]: I0227 17:24:53.450731 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bb48d97cc-t2pb6"] Feb 27 17:24:55 crc kubenswrapper[4700]: I0227 17:24:55.002434 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="884d838f-ded4-4ef7-92d5-7a6a1fcdecf2" path="/var/lib/kubelet/pods/884d838f-ded4-4ef7-92d5-7a6a1fcdecf2/volumes" Feb 27 17:24:56 crc kubenswrapper[4700]: I0227 17:24:56.688611 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 27 17:24:56 crc kubenswrapper[4700]: I0227 17:24:56.689107 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 27 17:24:57 crc kubenswrapper[4700]: I0227 17:24:57.436326 4700 generic.go:334] "Generic (PLEG): container finished" podID="e5e54324-9f69-4bc9-a6af-02b5ca6445e8" containerID="76cd2335a654dbd228865b5835dd72e39a3fe261f701ccccfeed3f61f4f4fbf2" exitCode=0 Feb 27 17:24:57 crc kubenswrapper[4700]: I0227 17:24:57.436394 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-9s776" event={"ID":"e5e54324-9f69-4bc9-a6af-02b5ca6445e8","Type":"ContainerDied","Data":"76cd2335a654dbd228865b5835dd72e39a3fe261f701ccccfeed3f61f4f4fbf2"} Feb 27 17:24:57 crc kubenswrapper[4700]: I0227 17:24:57.707673 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="aa7db367-fd4d-4915-803f-bdd113ddc82b" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.230:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 27 17:24:57 crc kubenswrapper[4700]: I0227 17:24:57.707698 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="aa7db367-fd4d-4915-803f-bdd113ddc82b" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.230:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 27 17:24:58 crc kubenswrapper[4700]: I0227 17:24:58.876711 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-9s776" Feb 27 17:24:58 crc kubenswrapper[4700]: I0227 17:24:58.970457 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rb9lz\" (UniqueName: \"kubernetes.io/projected/e5e54324-9f69-4bc9-a6af-02b5ca6445e8-kube-api-access-rb9lz\") pod \"e5e54324-9f69-4bc9-a6af-02b5ca6445e8\" (UID: \"e5e54324-9f69-4bc9-a6af-02b5ca6445e8\") " Feb 27 17:24:58 crc kubenswrapper[4700]: I0227 17:24:58.970707 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5e54324-9f69-4bc9-a6af-02b5ca6445e8-config-data\") pod \"e5e54324-9f69-4bc9-a6af-02b5ca6445e8\" (UID: \"e5e54324-9f69-4bc9-a6af-02b5ca6445e8\") " Feb 27 17:24:58 crc kubenswrapper[4700]: I0227 17:24:58.970757 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5e54324-9f69-4bc9-a6af-02b5ca6445e8-combined-ca-bundle\") pod \"e5e54324-9f69-4bc9-a6af-02b5ca6445e8\" (UID: \"e5e54324-9f69-4bc9-a6af-02b5ca6445e8\") " Feb 27 17:24:58 crc kubenswrapper[4700]: I0227 17:24:58.970783 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5e54324-9f69-4bc9-a6af-02b5ca6445e8-scripts\") pod \"e5e54324-9f69-4bc9-a6af-02b5ca6445e8\" (UID: \"e5e54324-9f69-4bc9-a6af-02b5ca6445e8\") " Feb 27 17:24:58 crc kubenswrapper[4700]: I0227 17:24:58.976965 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5e54324-9f69-4bc9-a6af-02b5ca6445e8-kube-api-access-rb9lz" (OuterVolumeSpecName: "kube-api-access-rb9lz") pod "e5e54324-9f69-4bc9-a6af-02b5ca6445e8" (UID: "e5e54324-9f69-4bc9-a6af-02b5ca6445e8"). InnerVolumeSpecName "kube-api-access-rb9lz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:24:58 crc kubenswrapper[4700]: I0227 17:24:58.988116 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5e54324-9f69-4bc9-a6af-02b5ca6445e8-scripts" (OuterVolumeSpecName: "scripts") pod "e5e54324-9f69-4bc9-a6af-02b5ca6445e8" (UID: "e5e54324-9f69-4bc9-a6af-02b5ca6445e8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:59 crc kubenswrapper[4700]: I0227 17:24:59.011978 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5e54324-9f69-4bc9-a6af-02b5ca6445e8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e5e54324-9f69-4bc9-a6af-02b5ca6445e8" (UID: "e5e54324-9f69-4bc9-a6af-02b5ca6445e8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:59 crc kubenswrapper[4700]: I0227 17:24:59.013780 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5e54324-9f69-4bc9-a6af-02b5ca6445e8-config-data" (OuterVolumeSpecName: "config-data") pod "e5e54324-9f69-4bc9-a6af-02b5ca6445e8" (UID: "e5e54324-9f69-4bc9-a6af-02b5ca6445e8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:24:59 crc kubenswrapper[4700]: I0227 17:24:59.073678 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5e54324-9f69-4bc9-a6af-02b5ca6445e8-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:59 crc kubenswrapper[4700]: I0227 17:24:59.073717 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5e54324-9f69-4bc9-a6af-02b5ca6445e8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:59 crc kubenswrapper[4700]: I0227 17:24:59.073733 4700 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5e54324-9f69-4bc9-a6af-02b5ca6445e8-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:59 crc kubenswrapper[4700]: I0227 17:24:59.073746 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rb9lz\" (UniqueName: \"kubernetes.io/projected/e5e54324-9f69-4bc9-a6af-02b5ca6445e8-kube-api-access-rb9lz\") on node \"crc\" DevicePath \"\"" Feb 27 17:24:59 crc kubenswrapper[4700]: I0227 17:24:59.487069 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-9s776" event={"ID":"e5e54324-9f69-4bc9-a6af-02b5ca6445e8","Type":"ContainerDied","Data":"72dd36b34bb6c86c99d9768649ad846189264bc5c5dd905119b3533b33eeb62a"} Feb 27 17:24:59 crc kubenswrapper[4700]: I0227 17:24:59.487408 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="72dd36b34bb6c86c99d9768649ad846189264bc5c5dd905119b3533b33eeb62a" Feb 27 17:24:59 crc kubenswrapper[4700]: I0227 17:24:59.487514 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-9s776" Feb 27 17:24:59 crc kubenswrapper[4700]: I0227 17:24:59.700437 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 17:24:59 crc kubenswrapper[4700]: I0227 17:24:59.700928 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="fd4be9bd-89e3-4e6c-b5a9-6ced858cb974" containerName="nova-scheduler-scheduler" containerID="cri-o://edb119d383758b51fad5555188e081aba816283333b85f5639fad1cfbaf98c58" gracePeriod=30 Feb 27 17:24:59 crc kubenswrapper[4700]: I0227 17:24:59.710644 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 27 17:24:59 crc kubenswrapper[4700]: I0227 17:24:59.710912 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="aa7db367-fd4d-4915-803f-bdd113ddc82b" containerName="nova-api-log" containerID="cri-o://6fb577ceb4ebed75979c2a5787d8e9babf6acb979630029fa2e18c525a970734" gracePeriod=30 Feb 27 17:24:59 crc kubenswrapper[4700]: I0227 17:24:59.711090 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="aa7db367-fd4d-4915-803f-bdd113ddc82b" containerName="nova-api-api" containerID="cri-o://b49c8d5c91529dc1ecfa9452fcd09735ab3f7589725e3dc78a86b3ac9064a73a" gracePeriod=30 Feb 27 17:24:59 crc kubenswrapper[4700]: I0227 17:24:59.728680 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 17:24:59 crc kubenswrapper[4700]: I0227 17:24:59.728926 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fb9499b9-4b5c-4f5a-9579-e3f7329f36a7" containerName="nova-metadata-log" containerID="cri-o://32ee692b1c18f2953e24371a5c20fd835dc7dce63f05a4c7c9136ba4edf02f31" gracePeriod=30 Feb 27 17:24:59 crc kubenswrapper[4700]: I0227 17:24:59.729008 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fb9499b9-4b5c-4f5a-9579-e3f7329f36a7" containerName="nova-metadata-metadata" containerID="cri-o://9a9c8b97aaeb05640df58b406cf322ef021ffe0c9683755246a4bdd7f62ac10a" gracePeriod=30 Feb 27 17:24:59 crc kubenswrapper[4700]: E0227 17:24:59.773928 4700 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb9499b9_4b5c_4f5a_9579_e3f7329f36a7.slice/crio-32ee692b1c18f2953e24371a5c20fd835dc7dce63f05a4c7c9136ba4edf02f31.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa7db367_fd4d_4915_803f_bdd113ddc82b.slice/crio-conmon-6fb577ceb4ebed75979c2a5787d8e9babf6acb979630029fa2e18c525a970734.scope\": RecentStats: unable to find data in memory cache]" Feb 27 17:25:00 crc kubenswrapper[4700]: I0227 17:25:00.496332 4700 generic.go:334] "Generic (PLEG): container finished" podID="fb9499b9-4b5c-4f5a-9579-e3f7329f36a7" containerID="32ee692b1c18f2953e24371a5c20fd835dc7dce63f05a4c7c9136ba4edf02f31" exitCode=143 Feb 27 17:25:00 crc kubenswrapper[4700]: I0227 17:25:00.496384 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fb9499b9-4b5c-4f5a-9579-e3f7329f36a7","Type":"ContainerDied","Data":"32ee692b1c18f2953e24371a5c20fd835dc7dce63f05a4c7c9136ba4edf02f31"} Feb 27 17:25:00 crc kubenswrapper[4700]: I0227 17:25:00.500106 4700 generic.go:334] "Generic (PLEG): container finished" podID="aa7db367-fd4d-4915-803f-bdd113ddc82b" containerID="6fb577ceb4ebed75979c2a5787d8e9babf6acb979630029fa2e18c525a970734" exitCode=143 Feb 27 17:25:00 crc kubenswrapper[4700]: I0227 17:25:00.500128 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"aa7db367-fd4d-4915-803f-bdd113ddc82b","Type":"ContainerDied","Data":"6fb577ceb4ebed75979c2a5787d8e9babf6acb979630029fa2e18c525a970734"} Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.002994 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.122597 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb9499b9-4b5c-4f5a-9579-e3f7329f36a7-config-data\") pod \"fb9499b9-4b5c-4f5a-9579-e3f7329f36a7\" (UID: \"fb9499b9-4b5c-4f5a-9579-e3f7329f36a7\") " Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.122740 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb9499b9-4b5c-4f5a-9579-e3f7329f36a7-combined-ca-bundle\") pod \"fb9499b9-4b5c-4f5a-9579-e3f7329f36a7\" (UID: \"fb9499b9-4b5c-4f5a-9579-e3f7329f36a7\") " Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.122923 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb9499b9-4b5c-4f5a-9579-e3f7329f36a7-logs\") pod \"fb9499b9-4b5c-4f5a-9579-e3f7329f36a7\" (UID: \"fb9499b9-4b5c-4f5a-9579-e3f7329f36a7\") " Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.123009 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rn2bf\" (UniqueName: \"kubernetes.io/projected/fb9499b9-4b5c-4f5a-9579-e3f7329f36a7-kube-api-access-rn2bf\") pod \"fb9499b9-4b5c-4f5a-9579-e3f7329f36a7\" (UID: \"fb9499b9-4b5c-4f5a-9579-e3f7329f36a7\") " Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.123043 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb9499b9-4b5c-4f5a-9579-e3f7329f36a7-nova-metadata-tls-certs\") pod \"fb9499b9-4b5c-4f5a-9579-e3f7329f36a7\" (UID: \"fb9499b9-4b5c-4f5a-9579-e3f7329f36a7\") " Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.123590 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb9499b9-4b5c-4f5a-9579-e3f7329f36a7-logs" (OuterVolumeSpecName: "logs") pod "fb9499b9-4b5c-4f5a-9579-e3f7329f36a7" (UID: "fb9499b9-4b5c-4f5a-9579-e3f7329f36a7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.125962 4700 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb9499b9-4b5c-4f5a-9579-e3f7329f36a7-logs\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.146501 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb9499b9-4b5c-4f5a-9579-e3f7329f36a7-kube-api-access-rn2bf" (OuterVolumeSpecName: "kube-api-access-rn2bf") pod "fb9499b9-4b5c-4f5a-9579-e3f7329f36a7" (UID: "fb9499b9-4b5c-4f5a-9579-e3f7329f36a7"). InnerVolumeSpecName "kube-api-access-rn2bf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.149166 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb9499b9-4b5c-4f5a-9579-e3f7329f36a7-config-data" (OuterVolumeSpecName: "config-data") pod "fb9499b9-4b5c-4f5a-9579-e3f7329f36a7" (UID: "fb9499b9-4b5c-4f5a-9579-e3f7329f36a7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.173632 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb9499b9-4b5c-4f5a-9579-e3f7329f36a7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fb9499b9-4b5c-4f5a-9579-e3f7329f36a7" (UID: "fb9499b9-4b5c-4f5a-9579-e3f7329f36a7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.195052 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb9499b9-4b5c-4f5a-9579-e3f7329f36a7-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "fb9499b9-4b5c-4f5a-9579-e3f7329f36a7" (UID: "fb9499b9-4b5c-4f5a-9579-e3f7329f36a7"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.227958 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rn2bf\" (UniqueName: \"kubernetes.io/projected/fb9499b9-4b5c-4f5a-9579-e3f7329f36a7-kube-api-access-rn2bf\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.227997 4700 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb9499b9-4b5c-4f5a-9579-e3f7329f36a7-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.228012 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb9499b9-4b5c-4f5a-9579-e3f7329f36a7-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.228025 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb9499b9-4b5c-4f5a-9579-e3f7329f36a7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:01 crc kubenswrapper[4700]: E0227 17:25:01.277374 4700 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="edb119d383758b51fad5555188e081aba816283333b85f5639fad1cfbaf98c58" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 27 17:25:01 crc kubenswrapper[4700]: E0227 17:25:01.280301 4700 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="edb119d383758b51fad5555188e081aba816283333b85f5639fad1cfbaf98c58" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 27 17:25:01 crc kubenswrapper[4700]: E0227 17:25:01.281577 4700 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="edb119d383758b51fad5555188e081aba816283333b85f5639fad1cfbaf98c58" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 27 17:25:01 crc kubenswrapper[4700]: E0227 17:25:01.281692 4700 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="fd4be9bd-89e3-4e6c-b5a9-6ced858cb974" containerName="nova-scheduler-scheduler" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.329721 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.430243 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa7db367-fd4d-4915-803f-bdd113ddc82b-internal-tls-certs\") pod \"aa7db367-fd4d-4915-803f-bdd113ddc82b\" (UID: \"aa7db367-fd4d-4915-803f-bdd113ddc82b\") " Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.430285 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmbmw\" (UniqueName: \"kubernetes.io/projected/aa7db367-fd4d-4915-803f-bdd113ddc82b-kube-api-access-mmbmw\") pod \"aa7db367-fd4d-4915-803f-bdd113ddc82b\" (UID: \"aa7db367-fd4d-4915-803f-bdd113ddc82b\") " Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.430366 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa7db367-fd4d-4915-803f-bdd113ddc82b-combined-ca-bundle\") pod \"aa7db367-fd4d-4915-803f-bdd113ddc82b\" (UID: \"aa7db367-fd4d-4915-803f-bdd113ddc82b\") " Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.430492 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa7db367-fd4d-4915-803f-bdd113ddc82b-config-data\") pod \"aa7db367-fd4d-4915-803f-bdd113ddc82b\" (UID: \"aa7db367-fd4d-4915-803f-bdd113ddc82b\") " Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.430535 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa7db367-fd4d-4915-803f-bdd113ddc82b-public-tls-certs\") pod \"aa7db367-fd4d-4915-803f-bdd113ddc82b\" (UID: \"aa7db367-fd4d-4915-803f-bdd113ddc82b\") " Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.430556 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa7db367-fd4d-4915-803f-bdd113ddc82b-logs\") pod \"aa7db367-fd4d-4915-803f-bdd113ddc82b\" (UID: \"aa7db367-fd4d-4915-803f-bdd113ddc82b\") " Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.431437 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa7db367-fd4d-4915-803f-bdd113ddc82b-logs" (OuterVolumeSpecName: "logs") pod "aa7db367-fd4d-4915-803f-bdd113ddc82b" (UID: "aa7db367-fd4d-4915-803f-bdd113ddc82b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.433137 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa7db367-fd4d-4915-803f-bdd113ddc82b-kube-api-access-mmbmw" (OuterVolumeSpecName: "kube-api-access-mmbmw") pod "aa7db367-fd4d-4915-803f-bdd113ddc82b" (UID: "aa7db367-fd4d-4915-803f-bdd113ddc82b"). InnerVolumeSpecName "kube-api-access-mmbmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.465203 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa7db367-fd4d-4915-803f-bdd113ddc82b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aa7db367-fd4d-4915-803f-bdd113ddc82b" (UID: "aa7db367-fd4d-4915-803f-bdd113ddc82b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.480973 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa7db367-fd4d-4915-803f-bdd113ddc82b-config-data" (OuterVolumeSpecName: "config-data") pod "aa7db367-fd4d-4915-803f-bdd113ddc82b" (UID: "aa7db367-fd4d-4915-803f-bdd113ddc82b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.490904 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa7db367-fd4d-4915-803f-bdd113ddc82b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "aa7db367-fd4d-4915-803f-bdd113ddc82b" (UID: "aa7db367-fd4d-4915-803f-bdd113ddc82b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.515650 4700 generic.go:334] "Generic (PLEG): container finished" podID="fb9499b9-4b5c-4f5a-9579-e3f7329f36a7" containerID="9a9c8b97aaeb05640df58b406cf322ef021ffe0c9683755246a4bdd7f62ac10a" exitCode=0 Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.515695 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fb9499b9-4b5c-4f5a-9579-e3f7329f36a7","Type":"ContainerDied","Data":"9a9c8b97aaeb05640df58b406cf322ef021ffe0c9683755246a4bdd7f62ac10a"} Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.515750 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fb9499b9-4b5c-4f5a-9579-e3f7329f36a7","Type":"ContainerDied","Data":"f811c5fcffa60355dc6283e3ed1086cd86b17dd5089ce28c46a64ad557bf98f2"} Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.515797 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.515818 4700 scope.go:117] "RemoveContainer" containerID="9a9c8b97aaeb05640df58b406cf322ef021ffe0c9683755246a4bdd7f62ac10a" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.521263 4700 generic.go:334] "Generic (PLEG): container finished" podID="aa7db367-fd4d-4915-803f-bdd113ddc82b" containerID="b49c8d5c91529dc1ecfa9452fcd09735ab3f7589725e3dc78a86b3ac9064a73a" exitCode=0 Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.521301 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"aa7db367-fd4d-4915-803f-bdd113ddc82b","Type":"ContainerDied","Data":"b49c8d5c91529dc1ecfa9452fcd09735ab3f7589725e3dc78a86b3ac9064a73a"} Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.521322 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"aa7db367-fd4d-4915-803f-bdd113ddc82b","Type":"ContainerDied","Data":"48a76d5ea18f0eed6ebc338a256fb081f461a39b6d50b87adde3d3a6e35c66a0"} Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.521825 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.531004 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa7db367-fd4d-4915-803f-bdd113ddc82b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "aa7db367-fd4d-4915-803f-bdd113ddc82b" (UID: "aa7db367-fd4d-4915-803f-bdd113ddc82b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.532951 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa7db367-fd4d-4915-803f-bdd113ddc82b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.533215 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa7db367-fd4d-4915-803f-bdd113ddc82b-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.533348 4700 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa7db367-fd4d-4915-803f-bdd113ddc82b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.533452 4700 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa7db367-fd4d-4915-803f-bdd113ddc82b-logs\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.533598 4700 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa7db367-fd4d-4915-803f-bdd113ddc82b-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.533704 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmbmw\" (UniqueName: \"kubernetes.io/projected/aa7db367-fd4d-4915-803f-bdd113ddc82b-kube-api-access-mmbmw\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.542223 4700 scope.go:117] "RemoveContainer" containerID="32ee692b1c18f2953e24371a5c20fd835dc7dce63f05a4c7c9136ba4edf02f31" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.580665 4700 scope.go:117] "RemoveContainer" containerID="9a9c8b97aaeb05640df58b406cf322ef021ffe0c9683755246a4bdd7f62ac10a" Feb 27 17:25:01 crc kubenswrapper[4700]: E0227 17:25:01.581275 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a9c8b97aaeb05640df58b406cf322ef021ffe0c9683755246a4bdd7f62ac10a\": container with ID starting with 9a9c8b97aaeb05640df58b406cf322ef021ffe0c9683755246a4bdd7f62ac10a not found: ID does not exist" containerID="9a9c8b97aaeb05640df58b406cf322ef021ffe0c9683755246a4bdd7f62ac10a" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.581408 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a9c8b97aaeb05640df58b406cf322ef021ffe0c9683755246a4bdd7f62ac10a"} err="failed to get container status \"9a9c8b97aaeb05640df58b406cf322ef021ffe0c9683755246a4bdd7f62ac10a\": rpc error: code = NotFound desc = could not find container \"9a9c8b97aaeb05640df58b406cf322ef021ffe0c9683755246a4bdd7f62ac10a\": container with ID starting with 9a9c8b97aaeb05640df58b406cf322ef021ffe0c9683755246a4bdd7f62ac10a not found: ID does not exist" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.581568 4700 scope.go:117] "RemoveContainer" containerID="32ee692b1c18f2953e24371a5c20fd835dc7dce63f05a4c7c9136ba4edf02f31" Feb 27 17:25:01 crc kubenswrapper[4700]: E0227 17:25:01.584709 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32ee692b1c18f2953e24371a5c20fd835dc7dce63f05a4c7c9136ba4edf02f31\": container with ID starting with 32ee692b1c18f2953e24371a5c20fd835dc7dce63f05a4c7c9136ba4edf02f31 not found: ID does not exist" containerID="32ee692b1c18f2953e24371a5c20fd835dc7dce63f05a4c7c9136ba4edf02f31" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.584762 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32ee692b1c18f2953e24371a5c20fd835dc7dce63f05a4c7c9136ba4edf02f31"} err="failed to get container status \"32ee692b1c18f2953e24371a5c20fd835dc7dce63f05a4c7c9136ba4edf02f31\": rpc error: code = NotFound desc = could not find container \"32ee692b1c18f2953e24371a5c20fd835dc7dce63f05a4c7c9136ba4edf02f31\": container with ID starting with 32ee692b1c18f2953e24371a5c20fd835dc7dce63f05a4c7c9136ba4edf02f31 not found: ID does not exist" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.584798 4700 scope.go:117] "RemoveContainer" containerID="b49c8d5c91529dc1ecfa9452fcd09735ab3f7589725e3dc78a86b3ac9064a73a" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.598995 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.612602 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.625105 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 27 17:25:01 crc kubenswrapper[4700]: E0227 17:25:01.625785 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="884d838f-ded4-4ef7-92d5-7a6a1fcdecf2" containerName="init" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.625930 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="884d838f-ded4-4ef7-92d5-7a6a1fcdecf2" containerName="init" Feb 27 17:25:01 crc kubenswrapper[4700]: E0227 17:25:01.626010 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb9499b9-4b5c-4f5a-9579-e3f7329f36a7" containerName="nova-metadata-metadata" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.626076 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb9499b9-4b5c-4f5a-9579-e3f7329f36a7" containerName="nova-metadata-metadata" Feb 27 17:25:01 crc kubenswrapper[4700]: E0227 17:25:01.626152 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="884d838f-ded4-4ef7-92d5-7a6a1fcdecf2" containerName="dnsmasq-dns" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.626250 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="884d838f-ded4-4ef7-92d5-7a6a1fcdecf2" containerName="dnsmasq-dns" Feb 27 17:25:01 crc kubenswrapper[4700]: E0227 17:25:01.626349 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa7db367-fd4d-4915-803f-bdd113ddc82b" containerName="nova-api-log" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.626422 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa7db367-fd4d-4915-803f-bdd113ddc82b" containerName="nova-api-log" Feb 27 17:25:01 crc kubenswrapper[4700]: E0227 17:25:01.626511 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb9499b9-4b5c-4f5a-9579-e3f7329f36a7" containerName="nova-metadata-log" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.626602 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb9499b9-4b5c-4f5a-9579-e3f7329f36a7" containerName="nova-metadata-log" Feb 27 17:25:01 crc kubenswrapper[4700]: E0227 17:25:01.626684 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5e54324-9f69-4bc9-a6af-02b5ca6445e8" containerName="nova-manage" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.626748 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5e54324-9f69-4bc9-a6af-02b5ca6445e8" containerName="nova-manage" Feb 27 17:25:01 crc kubenswrapper[4700]: E0227 17:25:01.626835 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa7db367-fd4d-4915-803f-bdd113ddc82b" containerName="nova-api-api" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.626910 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa7db367-fd4d-4915-803f-bdd113ddc82b" containerName="nova-api-api" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.627210 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb9499b9-4b5c-4f5a-9579-e3f7329f36a7" containerName="nova-metadata-metadata" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.627291 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="884d838f-ded4-4ef7-92d5-7a6a1fcdecf2" containerName="dnsmasq-dns" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.627364 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5e54324-9f69-4bc9-a6af-02b5ca6445e8" containerName="nova-manage" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.627436 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa7db367-fd4d-4915-803f-bdd113ddc82b" containerName="nova-api-log" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.627539 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb9499b9-4b5c-4f5a-9579-e3f7329f36a7" containerName="nova-metadata-log" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.627639 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa7db367-fd4d-4915-803f-bdd113ddc82b" containerName="nova-api-api" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.629268 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.632919 4700 scope.go:117] "RemoveContainer" containerID="6fb577ceb4ebed75979c2a5787d8e9babf6acb979630029fa2e18c525a970734" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.633254 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.635148 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.635663 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.659621 4700 scope.go:117] "RemoveContainer" containerID="b49c8d5c91529dc1ecfa9452fcd09735ab3f7589725e3dc78a86b3ac9064a73a" Feb 27 17:25:01 crc kubenswrapper[4700]: E0227 17:25:01.660804 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b49c8d5c91529dc1ecfa9452fcd09735ab3f7589725e3dc78a86b3ac9064a73a\": container with ID starting with b49c8d5c91529dc1ecfa9452fcd09735ab3f7589725e3dc78a86b3ac9064a73a not found: ID does not exist" containerID="b49c8d5c91529dc1ecfa9452fcd09735ab3f7589725e3dc78a86b3ac9064a73a" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.660838 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b49c8d5c91529dc1ecfa9452fcd09735ab3f7589725e3dc78a86b3ac9064a73a"} err="failed to get container status \"b49c8d5c91529dc1ecfa9452fcd09735ab3f7589725e3dc78a86b3ac9064a73a\": rpc error: code = NotFound desc = could not find container \"b49c8d5c91529dc1ecfa9452fcd09735ab3f7589725e3dc78a86b3ac9064a73a\": container with ID starting with b49c8d5c91529dc1ecfa9452fcd09735ab3f7589725e3dc78a86b3ac9064a73a not found: ID does not exist" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.660857 4700 scope.go:117] "RemoveContainer" containerID="6fb577ceb4ebed75979c2a5787d8e9babf6acb979630029fa2e18c525a970734" Feb 27 17:25:01 crc kubenswrapper[4700]: E0227 17:25:01.661114 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6fb577ceb4ebed75979c2a5787d8e9babf6acb979630029fa2e18c525a970734\": container with ID starting with 6fb577ceb4ebed75979c2a5787d8e9babf6acb979630029fa2e18c525a970734 not found: ID does not exist" containerID="6fb577ceb4ebed75979c2a5787d8e9babf6acb979630029fa2e18c525a970734" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.661167 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fb577ceb4ebed75979c2a5787d8e9babf6acb979630029fa2e18c525a970734"} err="failed to get container status \"6fb577ceb4ebed75979c2a5787d8e9babf6acb979630029fa2e18c525a970734\": rpc error: code = NotFound desc = could not find container \"6fb577ceb4ebed75979c2a5787d8e9babf6acb979630029fa2e18c525a970734\": container with ID starting with 6fb577ceb4ebed75979c2a5787d8e9babf6acb979630029fa2e18c525a970734 not found: ID does not exist" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.736697 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/08f8b14c-db2a-4ac2-818a-4978ca9abf8b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"08f8b14c-db2a-4ac2-818a-4978ca9abf8b\") " pod="openstack/nova-metadata-0" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.737045 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08f8b14c-db2a-4ac2-818a-4978ca9abf8b-logs\") pod \"nova-metadata-0\" (UID: \"08f8b14c-db2a-4ac2-818a-4978ca9abf8b\") " pod="openstack/nova-metadata-0" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.737374 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f8b14c-db2a-4ac2-818a-4978ca9abf8b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"08f8b14c-db2a-4ac2-818a-4978ca9abf8b\") " pod="openstack/nova-metadata-0" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.737474 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08f8b14c-db2a-4ac2-818a-4978ca9abf8b-config-data\") pod \"nova-metadata-0\" (UID: \"08f8b14c-db2a-4ac2-818a-4978ca9abf8b\") " pod="openstack/nova-metadata-0" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.737680 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-296md\" (UniqueName: \"kubernetes.io/projected/08f8b14c-db2a-4ac2-818a-4978ca9abf8b-kube-api-access-296md\") pod \"nova-metadata-0\" (UID: \"08f8b14c-db2a-4ac2-818a-4978ca9abf8b\") " pod="openstack/nova-metadata-0" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.838846 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-296md\" (UniqueName: \"kubernetes.io/projected/08f8b14c-db2a-4ac2-818a-4978ca9abf8b-kube-api-access-296md\") pod \"nova-metadata-0\" (UID: \"08f8b14c-db2a-4ac2-818a-4978ca9abf8b\") " pod="openstack/nova-metadata-0" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.838927 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/08f8b14c-db2a-4ac2-818a-4978ca9abf8b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"08f8b14c-db2a-4ac2-818a-4978ca9abf8b\") " pod="openstack/nova-metadata-0" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.838973 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08f8b14c-db2a-4ac2-818a-4978ca9abf8b-logs\") pod \"nova-metadata-0\" (UID: \"08f8b14c-db2a-4ac2-818a-4978ca9abf8b\") " pod="openstack/nova-metadata-0" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.839025 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f8b14c-db2a-4ac2-818a-4978ca9abf8b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"08f8b14c-db2a-4ac2-818a-4978ca9abf8b\") " pod="openstack/nova-metadata-0" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.839052 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08f8b14c-db2a-4ac2-818a-4978ca9abf8b-config-data\") pod \"nova-metadata-0\" (UID: \"08f8b14c-db2a-4ac2-818a-4978ca9abf8b\") " pod="openstack/nova-metadata-0" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.840011 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08f8b14c-db2a-4ac2-818a-4978ca9abf8b-logs\") pod \"nova-metadata-0\" (UID: \"08f8b14c-db2a-4ac2-818a-4978ca9abf8b\") " pod="openstack/nova-metadata-0" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.842527 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08f8b14c-db2a-4ac2-818a-4978ca9abf8b-config-data\") pod \"nova-metadata-0\" (UID: \"08f8b14c-db2a-4ac2-818a-4978ca9abf8b\") " pod="openstack/nova-metadata-0" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.843765 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08f8b14c-db2a-4ac2-818a-4978ca9abf8b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"08f8b14c-db2a-4ac2-818a-4978ca9abf8b\") " pod="openstack/nova-metadata-0" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.844046 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/08f8b14c-db2a-4ac2-818a-4978ca9abf8b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"08f8b14c-db2a-4ac2-818a-4978ca9abf8b\") " pod="openstack/nova-metadata-0" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.854613 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-296md\" (UniqueName: \"kubernetes.io/projected/08f8b14c-db2a-4ac2-818a-4978ca9abf8b-kube-api-access-296md\") pod \"nova-metadata-0\" (UID: \"08f8b14c-db2a-4ac2-818a-4978ca9abf8b\") " pod="openstack/nova-metadata-0" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.868744 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.878210 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.898710 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.903046 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.905056 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.905371 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.905398 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.927249 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 27 17:25:01 crc kubenswrapper[4700]: I0227 17:25:01.952746 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 17:25:02 crc kubenswrapper[4700]: I0227 17:25:02.042596 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87dac950-8564-49c9-a8bc-d51b0dd9773b-public-tls-certs\") pod \"nova-api-0\" (UID: \"87dac950-8564-49c9-a8bc-d51b0dd9773b\") " pod="openstack/nova-api-0" Feb 27 17:25:02 crc kubenswrapper[4700]: I0227 17:25:02.043839 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87dac950-8564-49c9-a8bc-d51b0dd9773b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"87dac950-8564-49c9-a8bc-d51b0dd9773b\") " pod="openstack/nova-api-0" Feb 27 17:25:02 crc kubenswrapper[4700]: I0227 17:25:02.044358 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87dac950-8564-49c9-a8bc-d51b0dd9773b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"87dac950-8564-49c9-a8bc-d51b0dd9773b\") " pod="openstack/nova-api-0" Feb 27 17:25:02 crc kubenswrapper[4700]: I0227 17:25:02.044831 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87dac950-8564-49c9-a8bc-d51b0dd9773b-config-data\") pod \"nova-api-0\" (UID: \"87dac950-8564-49c9-a8bc-d51b0dd9773b\") " pod="openstack/nova-api-0" Feb 27 17:25:02 crc kubenswrapper[4700]: I0227 17:25:02.045003 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbhd8\" (UniqueName: \"kubernetes.io/projected/87dac950-8564-49c9-a8bc-d51b0dd9773b-kube-api-access-mbhd8\") pod \"nova-api-0\" (UID: \"87dac950-8564-49c9-a8bc-d51b0dd9773b\") " pod="openstack/nova-api-0" Feb 27 17:25:02 crc kubenswrapper[4700]: I0227 17:25:02.045159 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87dac950-8564-49c9-a8bc-d51b0dd9773b-logs\") pod \"nova-api-0\" (UID: \"87dac950-8564-49c9-a8bc-d51b0dd9773b\") " pod="openstack/nova-api-0" Feb 27 17:25:02 crc kubenswrapper[4700]: I0227 17:25:02.151701 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87dac950-8564-49c9-a8bc-d51b0dd9773b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"87dac950-8564-49c9-a8bc-d51b0dd9773b\") " pod="openstack/nova-api-0" Feb 27 17:25:02 crc kubenswrapper[4700]: I0227 17:25:02.151805 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87dac950-8564-49c9-a8bc-d51b0dd9773b-config-data\") pod \"nova-api-0\" (UID: \"87dac950-8564-49c9-a8bc-d51b0dd9773b\") " pod="openstack/nova-api-0" Feb 27 17:25:02 crc kubenswrapper[4700]: I0227 17:25:02.151828 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbhd8\" (UniqueName: \"kubernetes.io/projected/87dac950-8564-49c9-a8bc-d51b0dd9773b-kube-api-access-mbhd8\") pod \"nova-api-0\" (UID: \"87dac950-8564-49c9-a8bc-d51b0dd9773b\") " pod="openstack/nova-api-0" Feb 27 17:25:02 crc kubenswrapper[4700]: I0227 17:25:02.151853 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87dac950-8564-49c9-a8bc-d51b0dd9773b-logs\") pod \"nova-api-0\" (UID: \"87dac950-8564-49c9-a8bc-d51b0dd9773b\") " pod="openstack/nova-api-0" Feb 27 17:25:02 crc kubenswrapper[4700]: I0227 17:25:02.151909 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87dac950-8564-49c9-a8bc-d51b0dd9773b-public-tls-certs\") pod \"nova-api-0\" (UID: \"87dac950-8564-49c9-a8bc-d51b0dd9773b\") " pod="openstack/nova-api-0" Feb 27 17:25:02 crc kubenswrapper[4700]: I0227 17:25:02.151950 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87dac950-8564-49c9-a8bc-d51b0dd9773b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"87dac950-8564-49c9-a8bc-d51b0dd9773b\") " pod="openstack/nova-api-0" Feb 27 17:25:02 crc kubenswrapper[4700]: I0227 17:25:02.152968 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87dac950-8564-49c9-a8bc-d51b0dd9773b-logs\") pod \"nova-api-0\" (UID: \"87dac950-8564-49c9-a8bc-d51b0dd9773b\") " pod="openstack/nova-api-0" Feb 27 17:25:02 crc kubenswrapper[4700]: I0227 17:25:02.156916 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87dac950-8564-49c9-a8bc-d51b0dd9773b-public-tls-certs\") pod \"nova-api-0\" (UID: \"87dac950-8564-49c9-a8bc-d51b0dd9773b\") " pod="openstack/nova-api-0" Feb 27 17:25:02 crc kubenswrapper[4700]: I0227 17:25:02.157140 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87dac950-8564-49c9-a8bc-d51b0dd9773b-config-data\") pod \"nova-api-0\" (UID: \"87dac950-8564-49c9-a8bc-d51b0dd9773b\") " pod="openstack/nova-api-0" Feb 27 17:25:02 crc kubenswrapper[4700]: I0227 17:25:02.159010 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87dac950-8564-49c9-a8bc-d51b0dd9773b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"87dac950-8564-49c9-a8bc-d51b0dd9773b\") " pod="openstack/nova-api-0" Feb 27 17:25:02 crc kubenswrapper[4700]: I0227 17:25:02.161127 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87dac950-8564-49c9-a8bc-d51b0dd9773b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"87dac950-8564-49c9-a8bc-d51b0dd9773b\") " pod="openstack/nova-api-0" Feb 27 17:25:02 crc kubenswrapper[4700]: I0227 17:25:02.169863 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbhd8\" (UniqueName: \"kubernetes.io/projected/87dac950-8564-49c9-a8bc-d51b0dd9773b-kube-api-access-mbhd8\") pod \"nova-api-0\" (UID: \"87dac950-8564-49c9-a8bc-d51b0dd9773b\") " pod="openstack/nova-api-0" Feb 27 17:25:02 crc kubenswrapper[4700]: I0227 17:25:02.395149 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 17:25:02 crc kubenswrapper[4700]: I0227 17:25:02.449324 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 17:25:02 crc kubenswrapper[4700]: W0227 17:25:02.453625 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08f8b14c_db2a_4ac2_818a_4978ca9abf8b.slice/crio-ef810393ec55087926acd224d6b009254fb2a84c62792041d16ae428121ef59b WatchSource:0}: Error finding container ef810393ec55087926acd224d6b009254fb2a84c62792041d16ae428121ef59b: Status 404 returned error can't find the container with id ef810393ec55087926acd224d6b009254fb2a84c62792041d16ae428121ef59b Feb 27 17:25:02 crc kubenswrapper[4700]: I0227 17:25:02.541185 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"08f8b14c-db2a-4ac2-818a-4978ca9abf8b","Type":"ContainerStarted","Data":"ef810393ec55087926acd224d6b009254fb2a84c62792041d16ae428121ef59b"} Feb 27 17:25:02 crc kubenswrapper[4700]: I0227 17:25:02.946895 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 27 17:25:02 crc kubenswrapper[4700]: W0227 17:25:02.948691 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87dac950_8564_49c9_a8bc_d51b0dd9773b.slice/crio-ab88a0ce94a8ed934128501b44a53fbcd17385b5d7caee094ae5e5f95c6166d0 WatchSource:0}: Error finding container ab88a0ce94a8ed934128501b44a53fbcd17385b5d7caee094ae5e5f95c6166d0: Status 404 returned error can't find the container with id ab88a0ce94a8ed934128501b44a53fbcd17385b5d7caee094ae5e5f95c6166d0 Feb 27 17:25:02 crc kubenswrapper[4700]: I0227 17:25:02.992489 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa7db367-fd4d-4915-803f-bdd113ddc82b" path="/var/lib/kubelet/pods/aa7db367-fd4d-4915-803f-bdd113ddc82b/volumes" Feb 27 17:25:02 crc kubenswrapper[4700]: I0227 17:25:02.993151 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb9499b9-4b5c-4f5a-9579-e3f7329f36a7" path="/var/lib/kubelet/pods/fb9499b9-4b5c-4f5a-9579-e3f7329f36a7/volumes" Feb 27 17:25:03 crc kubenswrapper[4700]: I0227 17:25:03.576167 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"87dac950-8564-49c9-a8bc-d51b0dd9773b","Type":"ContainerStarted","Data":"0b39a6a21b3f3e6b3ddd8bf300e4b162e187658ea94609729254c3d7bf859d56"} Feb 27 17:25:03 crc kubenswrapper[4700]: I0227 17:25:03.576737 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"87dac950-8564-49c9-a8bc-d51b0dd9773b","Type":"ContainerStarted","Data":"4122ddfd80724c573aa5458122369500c32358c5409c69db11d787561785f5e2"} Feb 27 17:25:03 crc kubenswrapper[4700]: I0227 17:25:03.576776 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"87dac950-8564-49c9-a8bc-d51b0dd9773b","Type":"ContainerStarted","Data":"ab88a0ce94a8ed934128501b44a53fbcd17385b5d7caee094ae5e5f95c6166d0"} Feb 27 17:25:03 crc kubenswrapper[4700]: I0227 17:25:03.581862 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"08f8b14c-db2a-4ac2-818a-4978ca9abf8b","Type":"ContainerStarted","Data":"7c8474fc8aeea650e4cf1ba9b2e0bcccf57df92ad619943d03b618bd854739f6"} Feb 27 17:25:03 crc kubenswrapper[4700]: I0227 17:25:03.582061 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"08f8b14c-db2a-4ac2-818a-4978ca9abf8b","Type":"ContainerStarted","Data":"c9b4b49ac398b28f202d94adff8b0ca327e8a188c3e91258bc52e973b806131f"} Feb 27 17:25:03 crc kubenswrapper[4700]: I0227 17:25:03.605304 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.605284488 podStartE2EDuration="2.605284488s" podCreationTimestamp="2026-02-27 17:25:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:25:03.60082371 +0000 UTC m=+1463.586136497" watchObservedRunningTime="2026-02-27 17:25:03.605284488 +0000 UTC m=+1463.590597245" Feb 27 17:25:03 crc kubenswrapper[4700]: I0227 17:25:03.626113 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.6260836850000002 podStartE2EDuration="2.626083685s" podCreationTimestamp="2026-02-27 17:25:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:25:03.625180012 +0000 UTC m=+1463.610492769" watchObservedRunningTime="2026-02-27 17:25:03.626083685 +0000 UTC m=+1463.611396462" Feb 27 17:25:05 crc kubenswrapper[4700]: I0227 17:25:05.236893 4700 scope.go:117] "RemoveContainer" containerID="a92bd1eb5131a729ffcb950c63978cef03db723ace5353e6cb1430d02d09d7e9" Feb 27 17:25:05 crc kubenswrapper[4700]: I0227 17:25:05.609244 4700 generic.go:334] "Generic (PLEG): container finished" podID="fd4be9bd-89e3-4e6c-b5a9-6ced858cb974" containerID="edb119d383758b51fad5555188e081aba816283333b85f5639fad1cfbaf98c58" exitCode=0 Feb 27 17:25:05 crc kubenswrapper[4700]: I0227 17:25:05.609325 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"fd4be9bd-89e3-4e6c-b5a9-6ced858cb974","Type":"ContainerDied","Data":"edb119d383758b51fad5555188e081aba816283333b85f5639fad1cfbaf98c58"} Feb 27 17:25:05 crc kubenswrapper[4700]: I0227 17:25:05.609732 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"fd4be9bd-89e3-4e6c-b5a9-6ced858cb974","Type":"ContainerDied","Data":"6c8c2b139534e5bb14ca0c8bd2dee47022a7332b82a8b4775b8ec4a421ad39ee"} Feb 27 17:25:05 crc kubenswrapper[4700]: I0227 17:25:05.609757 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c8c2b139534e5bb14ca0c8bd2dee47022a7332b82a8b4775b8ec4a421ad39ee" Feb 27 17:25:05 crc kubenswrapper[4700]: I0227 17:25:05.617749 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 27 17:25:05 crc kubenswrapper[4700]: I0227 17:25:05.741684 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd4be9bd-89e3-4e6c-b5a9-6ced858cb974-combined-ca-bundle\") pod \"fd4be9bd-89e3-4e6c-b5a9-6ced858cb974\" (UID: \"fd4be9bd-89e3-4e6c-b5a9-6ced858cb974\") " Feb 27 17:25:05 crc kubenswrapper[4700]: I0227 17:25:05.741910 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd4be9bd-89e3-4e6c-b5a9-6ced858cb974-config-data\") pod \"fd4be9bd-89e3-4e6c-b5a9-6ced858cb974\" (UID: \"fd4be9bd-89e3-4e6c-b5a9-6ced858cb974\") " Feb 27 17:25:05 crc kubenswrapper[4700]: I0227 17:25:05.742023 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kpwtf\" (UniqueName: \"kubernetes.io/projected/fd4be9bd-89e3-4e6c-b5a9-6ced858cb974-kube-api-access-kpwtf\") pod \"fd4be9bd-89e3-4e6c-b5a9-6ced858cb974\" (UID: \"fd4be9bd-89e3-4e6c-b5a9-6ced858cb974\") " Feb 27 17:25:05 crc kubenswrapper[4700]: I0227 17:25:05.755785 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd4be9bd-89e3-4e6c-b5a9-6ced858cb974-kube-api-access-kpwtf" (OuterVolumeSpecName: "kube-api-access-kpwtf") pod "fd4be9bd-89e3-4e6c-b5a9-6ced858cb974" (UID: "fd4be9bd-89e3-4e6c-b5a9-6ced858cb974"). InnerVolumeSpecName "kube-api-access-kpwtf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:25:05 crc kubenswrapper[4700]: I0227 17:25:05.796057 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd4be9bd-89e3-4e6c-b5a9-6ced858cb974-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fd4be9bd-89e3-4e6c-b5a9-6ced858cb974" (UID: "fd4be9bd-89e3-4e6c-b5a9-6ced858cb974"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:25:05 crc kubenswrapper[4700]: I0227 17:25:05.796544 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd4be9bd-89e3-4e6c-b5a9-6ced858cb974-config-data" (OuterVolumeSpecName: "config-data") pod "fd4be9bd-89e3-4e6c-b5a9-6ced858cb974" (UID: "fd4be9bd-89e3-4e6c-b5a9-6ced858cb974"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:25:05 crc kubenswrapper[4700]: I0227 17:25:05.846786 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd4be9bd-89e3-4e6c-b5a9-6ced858cb974-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:05 crc kubenswrapper[4700]: I0227 17:25:05.846851 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kpwtf\" (UniqueName: \"kubernetes.io/projected/fd4be9bd-89e3-4e6c-b5a9-6ced858cb974-kube-api-access-kpwtf\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:05 crc kubenswrapper[4700]: I0227 17:25:05.846886 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd4be9bd-89e3-4e6c-b5a9-6ced858cb974-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:06 crc kubenswrapper[4700]: I0227 17:25:06.410636 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:25:06 crc kubenswrapper[4700]: I0227 17:25:06.411159 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:25:06 crc kubenswrapper[4700]: I0227 17:25:06.411226 4700 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" Feb 27 17:25:06 crc kubenswrapper[4700]: I0227 17:25:06.412440 4700 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7f3dc0b799607b839ef8d740e7569600055eb53b8ad9a51f0d8e9d8e3a22c10d"} pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 17:25:06 crc kubenswrapper[4700]: I0227 17:25:06.412569 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" containerID="cri-o://7f3dc0b799607b839ef8d740e7569600055eb53b8ad9a51f0d8e9d8e3a22c10d" gracePeriod=600 Feb 27 17:25:06 crc kubenswrapper[4700]: I0227 17:25:06.640709 4700 generic.go:334] "Generic (PLEG): container finished" podID="84b45511-b94c-479f-98db-fd2c4eceec46" containerID="7f3dc0b799607b839ef8d740e7569600055eb53b8ad9a51f0d8e9d8e3a22c10d" exitCode=0 Feb 27 17:25:06 crc kubenswrapper[4700]: I0227 17:25:06.640800 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 27 17:25:06 crc kubenswrapper[4700]: I0227 17:25:06.640826 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerDied","Data":"7f3dc0b799607b839ef8d740e7569600055eb53b8ad9a51f0d8e9d8e3a22c10d"} Feb 27 17:25:06 crc kubenswrapper[4700]: I0227 17:25:06.640912 4700 scope.go:117] "RemoveContainer" containerID="c99a71cb46c45356ba1bf62a0adf7979247710d0e51e5eac769c9c248f05463d" Feb 27 17:25:06 crc kubenswrapper[4700]: I0227 17:25:06.698259 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 17:25:06 crc kubenswrapper[4700]: I0227 17:25:06.713516 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 17:25:06 crc kubenswrapper[4700]: I0227 17:25:06.726929 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 17:25:06 crc kubenswrapper[4700]: E0227 17:25:06.727615 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd4be9bd-89e3-4e6c-b5a9-6ced858cb974" containerName="nova-scheduler-scheduler" Feb 27 17:25:06 crc kubenswrapper[4700]: I0227 17:25:06.727649 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd4be9bd-89e3-4e6c-b5a9-6ced858cb974" containerName="nova-scheduler-scheduler" Feb 27 17:25:06 crc kubenswrapper[4700]: I0227 17:25:06.728051 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd4be9bd-89e3-4e6c-b5a9-6ced858cb974" containerName="nova-scheduler-scheduler" Feb 27 17:25:06 crc kubenswrapper[4700]: I0227 17:25:06.729160 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 27 17:25:06 crc kubenswrapper[4700]: I0227 17:25:06.732308 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 27 17:25:06 crc kubenswrapper[4700]: I0227 17:25:06.739574 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 17:25:06 crc kubenswrapper[4700]: I0227 17:25:06.872799 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9b4g\" (UniqueName: \"kubernetes.io/projected/c0c63170-09b4-45f8-9b8b-d52ec9af2181-kube-api-access-v9b4g\") pod \"nova-scheduler-0\" (UID: \"c0c63170-09b4-45f8-9b8b-d52ec9af2181\") " pod="openstack/nova-scheduler-0" Feb 27 17:25:06 crc kubenswrapper[4700]: I0227 17:25:06.873266 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0c63170-09b4-45f8-9b8b-d52ec9af2181-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c0c63170-09b4-45f8-9b8b-d52ec9af2181\") " pod="openstack/nova-scheduler-0" Feb 27 17:25:06 crc kubenswrapper[4700]: I0227 17:25:06.873329 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0c63170-09b4-45f8-9b8b-d52ec9af2181-config-data\") pod \"nova-scheduler-0\" (UID: \"c0c63170-09b4-45f8-9b8b-d52ec9af2181\") " pod="openstack/nova-scheduler-0" Feb 27 17:25:06 crc kubenswrapper[4700]: I0227 17:25:06.953633 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 27 17:25:06 crc kubenswrapper[4700]: I0227 17:25:06.953679 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 27 17:25:06 crc kubenswrapper[4700]: I0227 17:25:06.975242 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0c63170-09b4-45f8-9b8b-d52ec9af2181-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c0c63170-09b4-45f8-9b8b-d52ec9af2181\") " pod="openstack/nova-scheduler-0" Feb 27 17:25:06 crc kubenswrapper[4700]: I0227 17:25:06.975308 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0c63170-09b4-45f8-9b8b-d52ec9af2181-config-data\") pod \"nova-scheduler-0\" (UID: \"c0c63170-09b4-45f8-9b8b-d52ec9af2181\") " pod="openstack/nova-scheduler-0" Feb 27 17:25:06 crc kubenswrapper[4700]: I0227 17:25:06.975394 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9b4g\" (UniqueName: \"kubernetes.io/projected/c0c63170-09b4-45f8-9b8b-d52ec9af2181-kube-api-access-v9b4g\") pod \"nova-scheduler-0\" (UID: \"c0c63170-09b4-45f8-9b8b-d52ec9af2181\") " pod="openstack/nova-scheduler-0" Feb 27 17:25:06 crc kubenswrapper[4700]: I0227 17:25:06.982107 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0c63170-09b4-45f8-9b8b-d52ec9af2181-config-data\") pod \"nova-scheduler-0\" (UID: \"c0c63170-09b4-45f8-9b8b-d52ec9af2181\") " pod="openstack/nova-scheduler-0" Feb 27 17:25:06 crc kubenswrapper[4700]: I0227 17:25:06.987370 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0c63170-09b4-45f8-9b8b-d52ec9af2181-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c0c63170-09b4-45f8-9b8b-d52ec9af2181\") " pod="openstack/nova-scheduler-0" Feb 27 17:25:06 crc kubenswrapper[4700]: I0227 17:25:06.996078 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd4be9bd-89e3-4e6c-b5a9-6ced858cb974" path="/var/lib/kubelet/pods/fd4be9bd-89e3-4e6c-b5a9-6ced858cb974/volumes" Feb 27 17:25:07 crc kubenswrapper[4700]: I0227 17:25:07.003334 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9b4g\" (UniqueName: \"kubernetes.io/projected/c0c63170-09b4-45f8-9b8b-d52ec9af2181-kube-api-access-v9b4g\") pod \"nova-scheduler-0\" (UID: \"c0c63170-09b4-45f8-9b8b-d52ec9af2181\") " pod="openstack/nova-scheduler-0" Feb 27 17:25:07 crc kubenswrapper[4700]: I0227 17:25:07.056374 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 27 17:25:07 crc kubenswrapper[4700]: I0227 17:25:07.578090 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 17:25:07 crc kubenswrapper[4700]: W0227 17:25:07.592190 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0c63170_09b4_45f8_9b8b_d52ec9af2181.slice/crio-d85c189c53ffe43ad511086bd69345221e871f91ba19dbbf1a6cec930efdad54 WatchSource:0}: Error finding container d85c189c53ffe43ad511086bd69345221e871f91ba19dbbf1a6cec930efdad54: Status 404 returned error can't find the container with id d85c189c53ffe43ad511086bd69345221e871f91ba19dbbf1a6cec930efdad54 Feb 27 17:25:07 crc kubenswrapper[4700]: I0227 17:25:07.663272 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c0c63170-09b4-45f8-9b8b-d52ec9af2181","Type":"ContainerStarted","Data":"d85c189c53ffe43ad511086bd69345221e871f91ba19dbbf1a6cec930efdad54"} Feb 27 17:25:07 crc kubenswrapper[4700]: I0227 17:25:07.667454 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerStarted","Data":"88adee85741f363c1411884023846cd0aab91e2b1c4dca4a79262d1b15c4192f"} Feb 27 17:25:08 crc kubenswrapper[4700]: I0227 17:25:08.691646 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c0c63170-09b4-45f8-9b8b-d52ec9af2181","Type":"ContainerStarted","Data":"5ab2ab3d95739373494126a41070dc25c98505a8a36bc66ebf2f23f6305b3dee"} Feb 27 17:25:08 crc kubenswrapper[4700]: I0227 17:25:08.717243 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.717219688 podStartE2EDuration="2.717219688s" podCreationTimestamp="2026-02-27 17:25:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:25:08.715936244 +0000 UTC m=+1468.701249001" watchObservedRunningTime="2026-02-27 17:25:08.717219688 +0000 UTC m=+1468.702532445" Feb 27 17:25:11 crc kubenswrapper[4700]: I0227 17:25:11.953416 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 27 17:25:11 crc kubenswrapper[4700]: I0227 17:25:11.954067 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 27 17:25:12 crc kubenswrapper[4700]: I0227 17:25:12.057136 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 27 17:25:12 crc kubenswrapper[4700]: I0227 17:25:12.395815 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 27 17:25:12 crc kubenswrapper[4700]: I0227 17:25:12.395887 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 27 17:25:12 crc kubenswrapper[4700]: I0227 17:25:12.995657 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="08f8b14c-db2a-4ac2-818a-4978ca9abf8b" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.233:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 27 17:25:12 crc kubenswrapper[4700]: I0227 17:25:12.995690 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="08f8b14c-db2a-4ac2-818a-4978ca9abf8b" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.233:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 27 17:25:13 crc kubenswrapper[4700]: I0227 17:25:13.408730 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="87dac950-8564-49c9-a8bc-d51b0dd9773b" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.234:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 27 17:25:13 crc kubenswrapper[4700]: I0227 17:25:13.409219 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="87dac950-8564-49c9-a8bc-d51b0dd9773b" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.234:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 27 17:25:17 crc kubenswrapper[4700]: I0227 17:25:17.057229 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 27 17:25:17 crc kubenswrapper[4700]: I0227 17:25:17.105350 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 27 17:25:17 crc kubenswrapper[4700]: I0227 17:25:17.869843 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 27 17:25:18 crc kubenswrapper[4700]: I0227 17:25:18.723189 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 27 17:25:18 crc kubenswrapper[4700]: I0227 17:25:18.938442 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-b75gs"] Feb 27 17:25:18 crc kubenswrapper[4700]: I0227 17:25:18.940757 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b75gs" Feb 27 17:25:18 crc kubenswrapper[4700]: I0227 17:25:18.959347 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-b75gs"] Feb 27 17:25:19 crc kubenswrapper[4700]: I0227 17:25:19.097770 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdb80efc-e0bf-48fe-b8b8-60bc95f0972b-utilities\") pod \"community-operators-b75gs\" (UID: \"fdb80efc-e0bf-48fe-b8b8-60bc95f0972b\") " pod="openshift-marketplace/community-operators-b75gs" Feb 27 17:25:19 crc kubenswrapper[4700]: I0227 17:25:19.097964 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdb80efc-e0bf-48fe-b8b8-60bc95f0972b-catalog-content\") pod \"community-operators-b75gs\" (UID: \"fdb80efc-e0bf-48fe-b8b8-60bc95f0972b\") " pod="openshift-marketplace/community-operators-b75gs" Feb 27 17:25:19 crc kubenswrapper[4700]: I0227 17:25:19.098116 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wvj9\" (UniqueName: \"kubernetes.io/projected/fdb80efc-e0bf-48fe-b8b8-60bc95f0972b-kube-api-access-6wvj9\") pod \"community-operators-b75gs\" (UID: \"fdb80efc-e0bf-48fe-b8b8-60bc95f0972b\") " pod="openshift-marketplace/community-operators-b75gs" Feb 27 17:25:19 crc kubenswrapper[4700]: I0227 17:25:19.199904 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wvj9\" (UniqueName: \"kubernetes.io/projected/fdb80efc-e0bf-48fe-b8b8-60bc95f0972b-kube-api-access-6wvj9\") pod \"community-operators-b75gs\" (UID: \"fdb80efc-e0bf-48fe-b8b8-60bc95f0972b\") " pod="openshift-marketplace/community-operators-b75gs" Feb 27 17:25:19 crc kubenswrapper[4700]: I0227 17:25:19.200291 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdb80efc-e0bf-48fe-b8b8-60bc95f0972b-utilities\") pod \"community-operators-b75gs\" (UID: \"fdb80efc-e0bf-48fe-b8b8-60bc95f0972b\") " pod="openshift-marketplace/community-operators-b75gs" Feb 27 17:25:19 crc kubenswrapper[4700]: I0227 17:25:19.200437 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdb80efc-e0bf-48fe-b8b8-60bc95f0972b-catalog-content\") pod \"community-operators-b75gs\" (UID: \"fdb80efc-e0bf-48fe-b8b8-60bc95f0972b\") " pod="openshift-marketplace/community-operators-b75gs" Feb 27 17:25:19 crc kubenswrapper[4700]: I0227 17:25:19.201045 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdb80efc-e0bf-48fe-b8b8-60bc95f0972b-utilities\") pod \"community-operators-b75gs\" (UID: \"fdb80efc-e0bf-48fe-b8b8-60bc95f0972b\") " pod="openshift-marketplace/community-operators-b75gs" Feb 27 17:25:19 crc kubenswrapper[4700]: I0227 17:25:19.201114 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdb80efc-e0bf-48fe-b8b8-60bc95f0972b-catalog-content\") pod \"community-operators-b75gs\" (UID: \"fdb80efc-e0bf-48fe-b8b8-60bc95f0972b\") " pod="openshift-marketplace/community-operators-b75gs" Feb 27 17:25:19 crc kubenswrapper[4700]: I0227 17:25:19.233330 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wvj9\" (UniqueName: \"kubernetes.io/projected/fdb80efc-e0bf-48fe-b8b8-60bc95f0972b-kube-api-access-6wvj9\") pod \"community-operators-b75gs\" (UID: \"fdb80efc-e0bf-48fe-b8b8-60bc95f0972b\") " pod="openshift-marketplace/community-operators-b75gs" Feb 27 17:25:19 crc kubenswrapper[4700]: I0227 17:25:19.264198 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b75gs" Feb 27 17:25:19 crc kubenswrapper[4700]: I0227 17:25:19.741396 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-b75gs"] Feb 27 17:25:19 crc kubenswrapper[4700]: I0227 17:25:19.853800 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b75gs" event={"ID":"fdb80efc-e0bf-48fe-b8b8-60bc95f0972b","Type":"ContainerStarted","Data":"d66760630fabdf4ff1181b729de49a9bcf19f9a453df189df7145dd2d6cf7c71"} Feb 27 17:25:20 crc kubenswrapper[4700]: I0227 17:25:20.866224 4700 generic.go:334] "Generic (PLEG): container finished" podID="fdb80efc-e0bf-48fe-b8b8-60bc95f0972b" containerID="851fac12adbb43ef7c4eba118c1658e395a5450b5679f0c2c388dfafd48d482a" exitCode=0 Feb 27 17:25:20 crc kubenswrapper[4700]: I0227 17:25:20.866324 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b75gs" event={"ID":"fdb80efc-e0bf-48fe-b8b8-60bc95f0972b","Type":"ContainerDied","Data":"851fac12adbb43ef7c4eba118c1658e395a5450b5679f0c2c388dfafd48d482a"} Feb 27 17:25:20 crc kubenswrapper[4700]: I0227 17:25:20.869335 4700 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 17:25:21 crc kubenswrapper[4700]: I0227 17:25:21.967880 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 27 17:25:21 crc kubenswrapper[4700]: I0227 17:25:21.972731 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 27 17:25:21 crc kubenswrapper[4700]: I0227 17:25:21.976506 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 27 17:25:22 crc kubenswrapper[4700]: I0227 17:25:22.406971 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 27 17:25:22 crc kubenswrapper[4700]: I0227 17:25:22.409665 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 27 17:25:22 crc kubenswrapper[4700]: I0227 17:25:22.410130 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 27 17:25:22 crc kubenswrapper[4700]: I0227 17:25:22.420675 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 27 17:25:22 crc kubenswrapper[4700]: I0227 17:25:22.895308 4700 generic.go:334] "Generic (PLEG): container finished" podID="fdb80efc-e0bf-48fe-b8b8-60bc95f0972b" containerID="7a8af121a08fce0c1a93e21c783a57b3d9692a80de64822f366f45bf758be84f" exitCode=0 Feb 27 17:25:22 crc kubenswrapper[4700]: I0227 17:25:22.895437 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b75gs" event={"ID":"fdb80efc-e0bf-48fe-b8b8-60bc95f0972b","Type":"ContainerDied","Data":"7a8af121a08fce0c1a93e21c783a57b3d9692a80de64822f366f45bf758be84f"} Feb 27 17:25:22 crc kubenswrapper[4700]: I0227 17:25:22.895573 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 27 17:25:22 crc kubenswrapper[4700]: I0227 17:25:22.906444 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 27 17:25:22 crc kubenswrapper[4700]: I0227 17:25:22.910978 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 27 17:25:23 crc kubenswrapper[4700]: I0227 17:25:23.910649 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b75gs" event={"ID":"fdb80efc-e0bf-48fe-b8b8-60bc95f0972b","Type":"ContainerStarted","Data":"22f0632f71ffea19fab9b12e25f03e6082da32ef313f491f8772d004b0458ce3"} Feb 27 17:25:23 crc kubenswrapper[4700]: I0227 17:25:23.948065 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-b75gs" podStartSLOduration=3.315237937 podStartE2EDuration="5.948027616s" podCreationTimestamp="2026-02-27 17:25:18 +0000 UTC" firstStartedPulling="2026-02-27 17:25:20.868768939 +0000 UTC m=+1480.854081726" lastFinishedPulling="2026-02-27 17:25:23.501558648 +0000 UTC m=+1483.486871405" observedRunningTime="2026-02-27 17:25:23.930132715 +0000 UTC m=+1483.915445502" watchObservedRunningTime="2026-02-27 17:25:23.948027616 +0000 UTC m=+1483.933340413" Feb 27 17:25:29 crc kubenswrapper[4700]: I0227 17:25:29.265192 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-b75gs" Feb 27 17:25:29 crc kubenswrapper[4700]: I0227 17:25:29.265896 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-b75gs" Feb 27 17:25:29 crc kubenswrapper[4700]: I0227 17:25:29.355719 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-b75gs" Feb 27 17:25:30 crc kubenswrapper[4700]: I0227 17:25:30.071027 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-b75gs" Feb 27 17:25:30 crc kubenswrapper[4700]: I0227 17:25:30.660348 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-b75gs"] Feb 27 17:25:31 crc kubenswrapper[4700]: I0227 17:25:31.061290 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 27 17:25:31 crc kubenswrapper[4700]: I0227 17:25:31.249134 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6bw84"] Feb 27 17:25:31 crc kubenswrapper[4700]: I0227 17:25:31.251828 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6bw84" Feb 27 17:25:31 crc kubenswrapper[4700]: I0227 17:25:31.262199 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6bw84"] Feb 27 17:25:31 crc kubenswrapper[4700]: I0227 17:25:31.333268 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77aad150-188b-472a-bae0-2f5f7011617a-utilities\") pod \"redhat-operators-6bw84\" (UID: \"77aad150-188b-472a-bae0-2f5f7011617a\") " pod="openshift-marketplace/redhat-operators-6bw84" Feb 27 17:25:31 crc kubenswrapper[4700]: I0227 17:25:31.333371 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqq9f\" (UniqueName: \"kubernetes.io/projected/77aad150-188b-472a-bae0-2f5f7011617a-kube-api-access-qqq9f\") pod \"redhat-operators-6bw84\" (UID: \"77aad150-188b-472a-bae0-2f5f7011617a\") " pod="openshift-marketplace/redhat-operators-6bw84" Feb 27 17:25:31 crc kubenswrapper[4700]: I0227 17:25:31.333442 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77aad150-188b-472a-bae0-2f5f7011617a-catalog-content\") pod \"redhat-operators-6bw84\" (UID: \"77aad150-188b-472a-bae0-2f5f7011617a\") " pod="openshift-marketplace/redhat-operators-6bw84" Feb 27 17:25:31 crc kubenswrapper[4700]: I0227 17:25:31.434802 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77aad150-188b-472a-bae0-2f5f7011617a-utilities\") pod \"redhat-operators-6bw84\" (UID: \"77aad150-188b-472a-bae0-2f5f7011617a\") " pod="openshift-marketplace/redhat-operators-6bw84" Feb 27 17:25:31 crc kubenswrapper[4700]: I0227 17:25:31.434894 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqq9f\" (UniqueName: \"kubernetes.io/projected/77aad150-188b-472a-bae0-2f5f7011617a-kube-api-access-qqq9f\") pod \"redhat-operators-6bw84\" (UID: \"77aad150-188b-472a-bae0-2f5f7011617a\") " pod="openshift-marketplace/redhat-operators-6bw84" Feb 27 17:25:31 crc kubenswrapper[4700]: I0227 17:25:31.434993 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77aad150-188b-472a-bae0-2f5f7011617a-catalog-content\") pod \"redhat-operators-6bw84\" (UID: \"77aad150-188b-472a-bae0-2f5f7011617a\") " pod="openshift-marketplace/redhat-operators-6bw84" Feb 27 17:25:31 crc kubenswrapper[4700]: I0227 17:25:31.435393 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77aad150-188b-472a-bae0-2f5f7011617a-utilities\") pod \"redhat-operators-6bw84\" (UID: \"77aad150-188b-472a-bae0-2f5f7011617a\") " pod="openshift-marketplace/redhat-operators-6bw84" Feb 27 17:25:31 crc kubenswrapper[4700]: I0227 17:25:31.435520 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77aad150-188b-472a-bae0-2f5f7011617a-catalog-content\") pod \"redhat-operators-6bw84\" (UID: \"77aad150-188b-472a-bae0-2f5f7011617a\") " pod="openshift-marketplace/redhat-operators-6bw84" Feb 27 17:25:31 crc kubenswrapper[4700]: I0227 17:25:31.457334 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqq9f\" (UniqueName: \"kubernetes.io/projected/77aad150-188b-472a-bae0-2f5f7011617a-kube-api-access-qqq9f\") pod \"redhat-operators-6bw84\" (UID: \"77aad150-188b-472a-bae0-2f5f7011617a\") " pod="openshift-marketplace/redhat-operators-6bw84" Feb 27 17:25:31 crc kubenswrapper[4700]: I0227 17:25:31.572085 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6bw84" Feb 27 17:25:32 crc kubenswrapper[4700]: I0227 17:25:32.023994 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-b75gs" podUID="fdb80efc-e0bf-48fe-b8b8-60bc95f0972b" containerName="registry-server" containerID="cri-o://22f0632f71ffea19fab9b12e25f03e6082da32ef313f491f8772d004b0458ce3" gracePeriod=2 Feb 27 17:25:32 crc kubenswrapper[4700]: I0227 17:25:32.095030 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6bw84"] Feb 27 17:25:32 crc kubenswrapper[4700]: W0227 17:25:32.131657 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod77aad150_188b_472a_bae0_2f5f7011617a.slice/crio-d7ac175a43c8cade8c56d01540be666f6fd175cac23298d596d7132ad813d28c WatchSource:0}: Error finding container d7ac175a43c8cade8c56d01540be666f6fd175cac23298d596d7132ad813d28c: Status 404 returned error can't find the container with id d7ac175a43c8cade8c56d01540be666f6fd175cac23298d596d7132ad813d28c Feb 27 17:25:32 crc kubenswrapper[4700]: I0227 17:25:32.198941 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 27 17:25:32 crc kubenswrapper[4700]: I0227 17:25:32.462036 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b75gs" Feb 27 17:25:32 crc kubenswrapper[4700]: I0227 17:25:32.554581 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdb80efc-e0bf-48fe-b8b8-60bc95f0972b-catalog-content\") pod \"fdb80efc-e0bf-48fe-b8b8-60bc95f0972b\" (UID: \"fdb80efc-e0bf-48fe-b8b8-60bc95f0972b\") " Feb 27 17:25:32 crc kubenswrapper[4700]: I0227 17:25:32.554758 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6wvj9\" (UniqueName: \"kubernetes.io/projected/fdb80efc-e0bf-48fe-b8b8-60bc95f0972b-kube-api-access-6wvj9\") pod \"fdb80efc-e0bf-48fe-b8b8-60bc95f0972b\" (UID: \"fdb80efc-e0bf-48fe-b8b8-60bc95f0972b\") " Feb 27 17:25:32 crc kubenswrapper[4700]: I0227 17:25:32.554868 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdb80efc-e0bf-48fe-b8b8-60bc95f0972b-utilities\") pod \"fdb80efc-e0bf-48fe-b8b8-60bc95f0972b\" (UID: \"fdb80efc-e0bf-48fe-b8b8-60bc95f0972b\") " Feb 27 17:25:32 crc kubenswrapper[4700]: I0227 17:25:32.555721 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fdb80efc-e0bf-48fe-b8b8-60bc95f0972b-utilities" (OuterVolumeSpecName: "utilities") pod "fdb80efc-e0bf-48fe-b8b8-60bc95f0972b" (UID: "fdb80efc-e0bf-48fe-b8b8-60bc95f0972b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:25:32 crc kubenswrapper[4700]: I0227 17:25:32.560708 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdb80efc-e0bf-48fe-b8b8-60bc95f0972b-kube-api-access-6wvj9" (OuterVolumeSpecName: "kube-api-access-6wvj9") pod "fdb80efc-e0bf-48fe-b8b8-60bc95f0972b" (UID: "fdb80efc-e0bf-48fe-b8b8-60bc95f0972b"). InnerVolumeSpecName "kube-api-access-6wvj9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:25:32 crc kubenswrapper[4700]: I0227 17:25:32.603196 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fdb80efc-e0bf-48fe-b8b8-60bc95f0972b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fdb80efc-e0bf-48fe-b8b8-60bc95f0972b" (UID: "fdb80efc-e0bf-48fe-b8b8-60bc95f0972b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:25:32 crc kubenswrapper[4700]: I0227 17:25:32.656357 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6wvj9\" (UniqueName: \"kubernetes.io/projected/fdb80efc-e0bf-48fe-b8b8-60bc95f0972b-kube-api-access-6wvj9\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:32 crc kubenswrapper[4700]: I0227 17:25:32.656390 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdb80efc-e0bf-48fe-b8b8-60bc95f0972b-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:32 crc kubenswrapper[4700]: I0227 17:25:32.656399 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdb80efc-e0bf-48fe-b8b8-60bc95f0972b-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:33 crc kubenswrapper[4700]: I0227 17:25:33.035278 4700 generic.go:334] "Generic (PLEG): container finished" podID="77aad150-188b-472a-bae0-2f5f7011617a" containerID="bc133afdf262414257afca6233a3ba754f8358fc335a7bca6d17f753bc9aada6" exitCode=0 Feb 27 17:25:33 crc kubenswrapper[4700]: I0227 17:25:33.035346 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6bw84" event={"ID":"77aad150-188b-472a-bae0-2f5f7011617a","Type":"ContainerDied","Data":"bc133afdf262414257afca6233a3ba754f8358fc335a7bca6d17f753bc9aada6"} Feb 27 17:25:33 crc kubenswrapper[4700]: I0227 17:25:33.035370 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6bw84" event={"ID":"77aad150-188b-472a-bae0-2f5f7011617a","Type":"ContainerStarted","Data":"d7ac175a43c8cade8c56d01540be666f6fd175cac23298d596d7132ad813d28c"} Feb 27 17:25:33 crc kubenswrapper[4700]: I0227 17:25:33.039432 4700 generic.go:334] "Generic (PLEG): container finished" podID="fdb80efc-e0bf-48fe-b8b8-60bc95f0972b" containerID="22f0632f71ffea19fab9b12e25f03e6082da32ef313f491f8772d004b0458ce3" exitCode=0 Feb 27 17:25:33 crc kubenswrapper[4700]: I0227 17:25:33.039497 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b75gs" event={"ID":"fdb80efc-e0bf-48fe-b8b8-60bc95f0972b","Type":"ContainerDied","Data":"22f0632f71ffea19fab9b12e25f03e6082da32ef313f491f8772d004b0458ce3"} Feb 27 17:25:33 crc kubenswrapper[4700]: I0227 17:25:33.039528 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b75gs" event={"ID":"fdb80efc-e0bf-48fe-b8b8-60bc95f0972b","Type":"ContainerDied","Data":"d66760630fabdf4ff1181b729de49a9bcf19f9a453df189df7145dd2d6cf7c71"} Feb 27 17:25:33 crc kubenswrapper[4700]: I0227 17:25:33.039550 4700 scope.go:117] "RemoveContainer" containerID="22f0632f71ffea19fab9b12e25f03e6082da32ef313f491f8772d004b0458ce3" Feb 27 17:25:33 crc kubenswrapper[4700]: I0227 17:25:33.039764 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b75gs" Feb 27 17:25:33 crc kubenswrapper[4700]: I0227 17:25:33.097051 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-b75gs"] Feb 27 17:25:33 crc kubenswrapper[4700]: I0227 17:25:33.110554 4700 scope.go:117] "RemoveContainer" containerID="7a8af121a08fce0c1a93e21c783a57b3d9692a80de64822f366f45bf758be84f" Feb 27 17:25:33 crc kubenswrapper[4700]: I0227 17:25:33.113489 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-b75gs"] Feb 27 17:25:33 crc kubenswrapper[4700]: I0227 17:25:33.152205 4700 scope.go:117] "RemoveContainer" containerID="851fac12adbb43ef7c4eba118c1658e395a5450b5679f0c2c388dfafd48d482a" Feb 27 17:25:33 crc kubenswrapper[4700]: I0227 17:25:33.211680 4700 scope.go:117] "RemoveContainer" containerID="22f0632f71ffea19fab9b12e25f03e6082da32ef313f491f8772d004b0458ce3" Feb 27 17:25:33 crc kubenswrapper[4700]: E0227 17:25:33.212252 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22f0632f71ffea19fab9b12e25f03e6082da32ef313f491f8772d004b0458ce3\": container with ID starting with 22f0632f71ffea19fab9b12e25f03e6082da32ef313f491f8772d004b0458ce3 not found: ID does not exist" containerID="22f0632f71ffea19fab9b12e25f03e6082da32ef313f491f8772d004b0458ce3" Feb 27 17:25:33 crc kubenswrapper[4700]: I0227 17:25:33.212312 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22f0632f71ffea19fab9b12e25f03e6082da32ef313f491f8772d004b0458ce3"} err="failed to get container status \"22f0632f71ffea19fab9b12e25f03e6082da32ef313f491f8772d004b0458ce3\": rpc error: code = NotFound desc = could not find container \"22f0632f71ffea19fab9b12e25f03e6082da32ef313f491f8772d004b0458ce3\": container with ID starting with 22f0632f71ffea19fab9b12e25f03e6082da32ef313f491f8772d004b0458ce3 not found: ID does not exist" Feb 27 17:25:33 crc kubenswrapper[4700]: I0227 17:25:33.212363 4700 scope.go:117] "RemoveContainer" containerID="7a8af121a08fce0c1a93e21c783a57b3d9692a80de64822f366f45bf758be84f" Feb 27 17:25:33 crc kubenswrapper[4700]: E0227 17:25:33.213182 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a8af121a08fce0c1a93e21c783a57b3d9692a80de64822f366f45bf758be84f\": container with ID starting with 7a8af121a08fce0c1a93e21c783a57b3d9692a80de64822f366f45bf758be84f not found: ID does not exist" containerID="7a8af121a08fce0c1a93e21c783a57b3d9692a80de64822f366f45bf758be84f" Feb 27 17:25:33 crc kubenswrapper[4700]: I0227 17:25:33.213224 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a8af121a08fce0c1a93e21c783a57b3d9692a80de64822f366f45bf758be84f"} err="failed to get container status \"7a8af121a08fce0c1a93e21c783a57b3d9692a80de64822f366f45bf758be84f\": rpc error: code = NotFound desc = could not find container \"7a8af121a08fce0c1a93e21c783a57b3d9692a80de64822f366f45bf758be84f\": container with ID starting with 7a8af121a08fce0c1a93e21c783a57b3d9692a80de64822f366f45bf758be84f not found: ID does not exist" Feb 27 17:25:33 crc kubenswrapper[4700]: I0227 17:25:33.213245 4700 scope.go:117] "RemoveContainer" containerID="851fac12adbb43ef7c4eba118c1658e395a5450b5679f0c2c388dfafd48d482a" Feb 27 17:25:33 crc kubenswrapper[4700]: E0227 17:25:33.213934 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"851fac12adbb43ef7c4eba118c1658e395a5450b5679f0c2c388dfafd48d482a\": container with ID starting with 851fac12adbb43ef7c4eba118c1658e395a5450b5679f0c2c388dfafd48d482a not found: ID does not exist" containerID="851fac12adbb43ef7c4eba118c1658e395a5450b5679f0c2c388dfafd48d482a" Feb 27 17:25:33 crc kubenswrapper[4700]: I0227 17:25:33.213963 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"851fac12adbb43ef7c4eba118c1658e395a5450b5679f0c2c388dfafd48d482a"} err="failed to get container status \"851fac12adbb43ef7c4eba118c1658e395a5450b5679f0c2c388dfafd48d482a\": rpc error: code = NotFound desc = could not find container \"851fac12adbb43ef7c4eba118c1658e395a5450b5679f0c2c388dfafd48d482a\": container with ID starting with 851fac12adbb43ef7c4eba118c1658e395a5450b5679f0c2c388dfafd48d482a not found: ID does not exist" Feb 27 17:25:34 crc kubenswrapper[4700]: I0227 17:25:34.607318 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="efe7d637-da3a-4995-a40f-fae00257ac1d" containerName="rabbitmq" containerID="cri-o://3a5a3f9cb7f22f3a2b43a0279b69f515963804662da1ec627fe5bcf97b746739" gracePeriod=604797 Feb 27 17:25:34 crc kubenswrapper[4700]: I0227 17:25:34.994928 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdb80efc-e0bf-48fe-b8b8-60bc95f0972b" path="/var/lib/kubelet/pods/fdb80efc-e0bf-48fe-b8b8-60bc95f0972b/volumes" Feb 27 17:25:35 crc kubenswrapper[4700]: I0227 17:25:35.060744 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6bw84" event={"ID":"77aad150-188b-472a-bae0-2f5f7011617a","Type":"ContainerStarted","Data":"ce1d734fc2fecdbcabe47e8046e0a4948755e41754dab4f61c5bb5f3fe38110f"} Feb 27 17:25:35 crc kubenswrapper[4700]: I0227 17:25:35.780802 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="458e6422-b856-4bcf-8308-4b6cb9ec8fef" containerName="rabbitmq" containerID="cri-o://520ea869bacd9c0536d61aea44f09946e35bd5b34aac34d9e9100c9c2ed55787" gracePeriod=604797 Feb 27 17:25:37 crc kubenswrapper[4700]: I0227 17:25:37.089974 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="efe7d637-da3a-4995-a40f-fae00257ac1d" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.109:5671: connect: connection refused" Feb 27 17:25:37 crc kubenswrapper[4700]: I0227 17:25:37.092954 4700 generic.go:334] "Generic (PLEG): container finished" podID="77aad150-188b-472a-bae0-2f5f7011617a" containerID="ce1d734fc2fecdbcabe47e8046e0a4948755e41754dab4f61c5bb5f3fe38110f" exitCode=0 Feb 27 17:25:37 crc kubenswrapper[4700]: I0227 17:25:37.093091 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6bw84" event={"ID":"77aad150-188b-472a-bae0-2f5f7011617a","Type":"ContainerDied","Data":"ce1d734fc2fecdbcabe47e8046e0a4948755e41754dab4f61c5bb5f3fe38110f"} Feb 27 17:25:37 crc kubenswrapper[4700]: I0227 17:25:37.703333 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="458e6422-b856-4bcf-8308-4b6cb9ec8fef" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.111:5671: connect: connection refused" Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.112617 4700 generic.go:334] "Generic (PLEG): container finished" podID="efe7d637-da3a-4995-a40f-fae00257ac1d" containerID="3a5a3f9cb7f22f3a2b43a0279b69f515963804662da1ec627fe5bcf97b746739" exitCode=0 Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.112690 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"efe7d637-da3a-4995-a40f-fae00257ac1d","Type":"ContainerDied","Data":"3a5a3f9cb7f22f3a2b43a0279b69f515963804662da1ec627fe5bcf97b746739"} Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.559008 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.687230 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/efe7d637-da3a-4995-a40f-fae00257ac1d-rabbitmq-confd\") pod \"efe7d637-da3a-4995-a40f-fae00257ac1d\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.687316 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/efe7d637-da3a-4995-a40f-fae00257ac1d-server-conf\") pod \"efe7d637-da3a-4995-a40f-fae00257ac1d\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.687411 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/efe7d637-da3a-4995-a40f-fae00257ac1d-pod-info\") pod \"efe7d637-da3a-4995-a40f-fae00257ac1d\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.687473 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/efe7d637-da3a-4995-a40f-fae00257ac1d-rabbitmq-tls\") pod \"efe7d637-da3a-4995-a40f-fae00257ac1d\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.687545 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rw6kl\" (UniqueName: \"kubernetes.io/projected/efe7d637-da3a-4995-a40f-fae00257ac1d-kube-api-access-rw6kl\") pod \"efe7d637-da3a-4995-a40f-fae00257ac1d\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.687595 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/efe7d637-da3a-4995-a40f-fae00257ac1d-config-data\") pod \"efe7d637-da3a-4995-a40f-fae00257ac1d\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.687667 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/efe7d637-da3a-4995-a40f-fae00257ac1d-plugins-conf\") pod \"efe7d637-da3a-4995-a40f-fae00257ac1d\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.687703 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"efe7d637-da3a-4995-a40f-fae00257ac1d\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.687752 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/efe7d637-da3a-4995-a40f-fae00257ac1d-rabbitmq-plugins\") pod \"efe7d637-da3a-4995-a40f-fae00257ac1d\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.687795 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/efe7d637-da3a-4995-a40f-fae00257ac1d-erlang-cookie-secret\") pod \"efe7d637-da3a-4995-a40f-fae00257ac1d\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.687832 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/efe7d637-da3a-4995-a40f-fae00257ac1d-rabbitmq-erlang-cookie\") pod \"efe7d637-da3a-4995-a40f-fae00257ac1d\" (UID: \"efe7d637-da3a-4995-a40f-fae00257ac1d\") " Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.694248 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efe7d637-da3a-4995-a40f-fae00257ac1d-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "efe7d637-da3a-4995-a40f-fae00257ac1d" (UID: "efe7d637-da3a-4995-a40f-fae00257ac1d"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.694868 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/efe7d637-da3a-4995-a40f-fae00257ac1d-pod-info" (OuterVolumeSpecName: "pod-info") pod "efe7d637-da3a-4995-a40f-fae00257ac1d" (UID: "efe7d637-da3a-4995-a40f-fae00257ac1d"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.697579 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efe7d637-da3a-4995-a40f-fae00257ac1d-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "efe7d637-da3a-4995-a40f-fae00257ac1d" (UID: "efe7d637-da3a-4995-a40f-fae00257ac1d"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.698569 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efe7d637-da3a-4995-a40f-fae00257ac1d-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "efe7d637-da3a-4995-a40f-fae00257ac1d" (UID: "efe7d637-da3a-4995-a40f-fae00257ac1d"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.699622 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "persistence") pod "efe7d637-da3a-4995-a40f-fae00257ac1d" (UID: "efe7d637-da3a-4995-a40f-fae00257ac1d"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.716222 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efe7d637-da3a-4995-a40f-fae00257ac1d-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "efe7d637-da3a-4995-a40f-fae00257ac1d" (UID: "efe7d637-da3a-4995-a40f-fae00257ac1d"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.717920 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efe7d637-da3a-4995-a40f-fae00257ac1d-kube-api-access-rw6kl" (OuterVolumeSpecName: "kube-api-access-rw6kl") pod "efe7d637-da3a-4995-a40f-fae00257ac1d" (UID: "efe7d637-da3a-4995-a40f-fae00257ac1d"). InnerVolumeSpecName "kube-api-access-rw6kl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.726268 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efe7d637-da3a-4995-a40f-fae00257ac1d-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "efe7d637-da3a-4995-a40f-fae00257ac1d" (UID: "efe7d637-da3a-4995-a40f-fae00257ac1d"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.788141 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efe7d637-da3a-4995-a40f-fae00257ac1d-server-conf" (OuterVolumeSpecName: "server-conf") pod "efe7d637-da3a-4995-a40f-fae00257ac1d" (UID: "efe7d637-da3a-4995-a40f-fae00257ac1d"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.790455 4700 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/efe7d637-da3a-4995-a40f-fae00257ac1d-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.790499 4700 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.790516 4700 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/efe7d637-da3a-4995-a40f-fae00257ac1d-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.790525 4700 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/efe7d637-da3a-4995-a40f-fae00257ac1d-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.790534 4700 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/efe7d637-da3a-4995-a40f-fae00257ac1d-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.790543 4700 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/efe7d637-da3a-4995-a40f-fae00257ac1d-server-conf\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.790551 4700 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/efe7d637-da3a-4995-a40f-fae00257ac1d-pod-info\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.790558 4700 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/efe7d637-da3a-4995-a40f-fae00257ac1d-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.790568 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rw6kl\" (UniqueName: \"kubernetes.io/projected/efe7d637-da3a-4995-a40f-fae00257ac1d-kube-api-access-rw6kl\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.797720 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efe7d637-da3a-4995-a40f-fae00257ac1d-config-data" (OuterVolumeSpecName: "config-data") pod "efe7d637-da3a-4995-a40f-fae00257ac1d" (UID: "efe7d637-da3a-4995-a40f-fae00257ac1d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.823784 4700 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.861198 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efe7d637-da3a-4995-a40f-fae00257ac1d-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "efe7d637-da3a-4995-a40f-fae00257ac1d" (UID: "efe7d637-da3a-4995-a40f-fae00257ac1d"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.892748 4700 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.892782 4700 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/efe7d637-da3a-4995-a40f-fae00257ac1d-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:38 crc kubenswrapper[4700]: I0227 17:25:38.892794 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/efe7d637-da3a-4995-a40f-fae00257ac1d-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.121491 4700 generic.go:334] "Generic (PLEG): container finished" podID="458e6422-b856-4bcf-8308-4b6cb9ec8fef" containerID="520ea869bacd9c0536d61aea44f09946e35bd5b34aac34d9e9100c9c2ed55787" exitCode=0 Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.121546 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"458e6422-b856-4bcf-8308-4b6cb9ec8fef","Type":"ContainerDied","Data":"520ea869bacd9c0536d61aea44f09946e35bd5b34aac34d9e9100c9c2ed55787"} Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.122751 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"efe7d637-da3a-4995-a40f-fae00257ac1d","Type":"ContainerDied","Data":"d359c4ba60ca395c18f917a41cc24f1757315df110bc33704754d767fc0749b0"} Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.122774 4700 scope.go:117] "RemoveContainer" containerID="3a5a3f9cb7f22f3a2b43a0279b69f515963804662da1ec627fe5bcf97b746739" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.122881 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.128202 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6bw84" event={"ID":"77aad150-188b-472a-bae0-2f5f7011617a","Type":"ContainerStarted","Data":"eb39d5946ec84e1def84398ed6d684d12fd97f698b941c8a89a0cd0d98bbd0cb"} Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.152125 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6bw84" podStartSLOduration=3.017294576 podStartE2EDuration="8.15210969s" podCreationTimestamp="2026-02-27 17:25:31 +0000 UTC" firstStartedPulling="2026-02-27 17:25:33.037156433 +0000 UTC m=+1493.022469180" lastFinishedPulling="2026-02-27 17:25:38.171971507 +0000 UTC m=+1498.157284294" observedRunningTime="2026-02-27 17:25:39.143855513 +0000 UTC m=+1499.129168270" watchObservedRunningTime="2026-02-27 17:25:39.15210969 +0000 UTC m=+1499.137422437" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.175738 4700 scope.go:117] "RemoveContainer" containerID="eb50098b16c98521dac5370ccb4478720cceb5dc15fabc0e5682e5ccc113eb78" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.183514 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.195168 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.209884 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Feb 27 17:25:39 crc kubenswrapper[4700]: E0227 17:25:39.210375 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdb80efc-e0bf-48fe-b8b8-60bc95f0972b" containerName="extract-utilities" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.210393 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdb80efc-e0bf-48fe-b8b8-60bc95f0972b" containerName="extract-utilities" Feb 27 17:25:39 crc kubenswrapper[4700]: E0227 17:25:39.210416 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdb80efc-e0bf-48fe-b8b8-60bc95f0972b" containerName="extract-content" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.210423 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdb80efc-e0bf-48fe-b8b8-60bc95f0972b" containerName="extract-content" Feb 27 17:25:39 crc kubenswrapper[4700]: E0227 17:25:39.210432 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efe7d637-da3a-4995-a40f-fae00257ac1d" containerName="setup-container" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.210438 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="efe7d637-da3a-4995-a40f-fae00257ac1d" containerName="setup-container" Feb 27 17:25:39 crc kubenswrapper[4700]: E0227 17:25:39.210453 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efe7d637-da3a-4995-a40f-fae00257ac1d" containerName="rabbitmq" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.210472 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="efe7d637-da3a-4995-a40f-fae00257ac1d" containerName="rabbitmq" Feb 27 17:25:39 crc kubenswrapper[4700]: E0227 17:25:39.210495 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdb80efc-e0bf-48fe-b8b8-60bc95f0972b" containerName="registry-server" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.210501 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdb80efc-e0bf-48fe-b8b8-60bc95f0972b" containerName="registry-server" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.210666 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdb80efc-e0bf-48fe-b8b8-60bc95f0972b" containerName="registry-server" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.210696 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="efe7d637-da3a-4995-a40f-fae00257ac1d" containerName="rabbitmq" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.211759 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.215392 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.215842 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.215918 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.216070 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.216107 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-7fq6k" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.219806 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.219989 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.242958 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.247266 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.301708 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jl7p\" (UniqueName: \"kubernetes.io/projected/15c786bb-7353-4d06-b410-a44458a5f954-kube-api-access-7jl7p\") pod \"rabbitmq-server-0\" (UID: \"15c786bb-7353-4d06-b410-a44458a5f954\") " pod="openstack/rabbitmq-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.301978 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/15c786bb-7353-4d06-b410-a44458a5f954-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"15c786bb-7353-4d06-b410-a44458a5f954\") " pod="openstack/rabbitmq-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.302072 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/15c786bb-7353-4d06-b410-a44458a5f954-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"15c786bb-7353-4d06-b410-a44458a5f954\") " pod="openstack/rabbitmq-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.302148 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/15c786bb-7353-4d06-b410-a44458a5f954-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"15c786bb-7353-4d06-b410-a44458a5f954\") " pod="openstack/rabbitmq-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.302242 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/15c786bb-7353-4d06-b410-a44458a5f954-server-conf\") pod \"rabbitmq-server-0\" (UID: \"15c786bb-7353-4d06-b410-a44458a5f954\") " pod="openstack/rabbitmq-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.302514 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/15c786bb-7353-4d06-b410-a44458a5f954-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"15c786bb-7353-4d06-b410-a44458a5f954\") " pod="openstack/rabbitmq-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.302600 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"15c786bb-7353-4d06-b410-a44458a5f954\") " pod="openstack/rabbitmq-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.302690 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/15c786bb-7353-4d06-b410-a44458a5f954-config-data\") pod \"rabbitmq-server-0\" (UID: \"15c786bb-7353-4d06-b410-a44458a5f954\") " pod="openstack/rabbitmq-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.302783 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/15c786bb-7353-4d06-b410-a44458a5f954-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"15c786bb-7353-4d06-b410-a44458a5f954\") " pod="openstack/rabbitmq-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.302897 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/15c786bb-7353-4d06-b410-a44458a5f954-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"15c786bb-7353-4d06-b410-a44458a5f954\") " pod="openstack/rabbitmq-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.302978 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/15c786bb-7353-4d06-b410-a44458a5f954-pod-info\") pod \"rabbitmq-server-0\" (UID: \"15c786bb-7353-4d06-b410-a44458a5f954\") " pod="openstack/rabbitmq-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.406401 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/458e6422-b856-4bcf-8308-4b6cb9ec8fef-server-conf\") pod \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.406492 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/458e6422-b856-4bcf-8308-4b6cb9ec8fef-pod-info\") pod \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.406535 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/458e6422-b856-4bcf-8308-4b6cb9ec8fef-rabbitmq-confd\") pod \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.406580 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbjsk\" (UniqueName: \"kubernetes.io/projected/458e6422-b856-4bcf-8308-4b6cb9ec8fef-kube-api-access-dbjsk\") pod \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.406611 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/458e6422-b856-4bcf-8308-4b6cb9ec8fef-config-data\") pod \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.406669 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/458e6422-b856-4bcf-8308-4b6cb9ec8fef-plugins-conf\") pod \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.406706 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/458e6422-b856-4bcf-8308-4b6cb9ec8fef-rabbitmq-erlang-cookie\") pod \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.406722 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/458e6422-b856-4bcf-8308-4b6cb9ec8fef-erlang-cookie-secret\") pod \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.406764 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/458e6422-b856-4bcf-8308-4b6cb9ec8fef-rabbitmq-plugins\") pod \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.406813 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/458e6422-b856-4bcf-8308-4b6cb9ec8fef-rabbitmq-tls\") pod \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.406849 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\" (UID: \"458e6422-b856-4bcf-8308-4b6cb9ec8fef\") " Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.407058 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/15c786bb-7353-4d06-b410-a44458a5f954-config-data\") pod \"rabbitmq-server-0\" (UID: \"15c786bb-7353-4d06-b410-a44458a5f954\") " pod="openstack/rabbitmq-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.407101 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/15c786bb-7353-4d06-b410-a44458a5f954-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"15c786bb-7353-4d06-b410-a44458a5f954\") " pod="openstack/rabbitmq-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.407172 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/15c786bb-7353-4d06-b410-a44458a5f954-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"15c786bb-7353-4d06-b410-a44458a5f954\") " pod="openstack/rabbitmq-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.407199 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/15c786bb-7353-4d06-b410-a44458a5f954-pod-info\") pod \"rabbitmq-server-0\" (UID: \"15c786bb-7353-4d06-b410-a44458a5f954\") " pod="openstack/rabbitmq-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.407228 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jl7p\" (UniqueName: \"kubernetes.io/projected/15c786bb-7353-4d06-b410-a44458a5f954-kube-api-access-7jl7p\") pod \"rabbitmq-server-0\" (UID: \"15c786bb-7353-4d06-b410-a44458a5f954\") " pod="openstack/rabbitmq-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.407262 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/15c786bb-7353-4d06-b410-a44458a5f954-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"15c786bb-7353-4d06-b410-a44458a5f954\") " pod="openstack/rabbitmq-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.407290 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/15c786bb-7353-4d06-b410-a44458a5f954-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"15c786bb-7353-4d06-b410-a44458a5f954\") " pod="openstack/rabbitmq-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.407309 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/15c786bb-7353-4d06-b410-a44458a5f954-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"15c786bb-7353-4d06-b410-a44458a5f954\") " pod="openstack/rabbitmq-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.407343 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/15c786bb-7353-4d06-b410-a44458a5f954-server-conf\") pod \"rabbitmq-server-0\" (UID: \"15c786bb-7353-4d06-b410-a44458a5f954\") " pod="openstack/rabbitmq-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.407359 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/15c786bb-7353-4d06-b410-a44458a5f954-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"15c786bb-7353-4d06-b410-a44458a5f954\") " pod="openstack/rabbitmq-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.407383 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"15c786bb-7353-4d06-b410-a44458a5f954\") " pod="openstack/rabbitmq-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.407642 4700 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"15c786bb-7353-4d06-b410-a44458a5f954\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.418518 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/15c786bb-7353-4d06-b410-a44458a5f954-config-data\") pod \"rabbitmq-server-0\" (UID: \"15c786bb-7353-4d06-b410-a44458a5f954\") " pod="openstack/rabbitmq-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.419084 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/15c786bb-7353-4d06-b410-a44458a5f954-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"15c786bb-7353-4d06-b410-a44458a5f954\") " pod="openstack/rabbitmq-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.419202 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/15c786bb-7353-4d06-b410-a44458a5f954-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"15c786bb-7353-4d06-b410-a44458a5f954\") " pod="openstack/rabbitmq-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.419414 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/458e6422-b856-4bcf-8308-4b6cb9ec8fef-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "458e6422-b856-4bcf-8308-4b6cb9ec8fef" (UID: "458e6422-b856-4bcf-8308-4b6cb9ec8fef"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.419468 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/15c786bb-7353-4d06-b410-a44458a5f954-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"15c786bb-7353-4d06-b410-a44458a5f954\") " pod="openstack/rabbitmq-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.420067 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/15c786bb-7353-4d06-b410-a44458a5f954-server-conf\") pod \"rabbitmq-server-0\" (UID: \"15c786bb-7353-4d06-b410-a44458a5f954\") " pod="openstack/rabbitmq-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.420205 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/458e6422-b856-4bcf-8308-4b6cb9ec8fef-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "458e6422-b856-4bcf-8308-4b6cb9ec8fef" (UID: "458e6422-b856-4bcf-8308-4b6cb9ec8fef"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.422960 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/458e6422-b856-4bcf-8308-4b6cb9ec8fef-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "458e6422-b856-4bcf-8308-4b6cb9ec8fef" (UID: "458e6422-b856-4bcf-8308-4b6cb9ec8fef"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.424803 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/458e6422-b856-4bcf-8308-4b6cb9ec8fef-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "458e6422-b856-4bcf-8308-4b6cb9ec8fef" (UID: "458e6422-b856-4bcf-8308-4b6cb9ec8fef"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.425286 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/15c786bb-7353-4d06-b410-a44458a5f954-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"15c786bb-7353-4d06-b410-a44458a5f954\") " pod="openstack/rabbitmq-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.426203 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/15c786bb-7353-4d06-b410-a44458a5f954-pod-info\") pod \"rabbitmq-server-0\" (UID: \"15c786bb-7353-4d06-b410-a44458a5f954\") " pod="openstack/rabbitmq-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.428035 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/15c786bb-7353-4d06-b410-a44458a5f954-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"15c786bb-7353-4d06-b410-a44458a5f954\") " pod="openstack/rabbitmq-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.430658 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "458e6422-b856-4bcf-8308-4b6cb9ec8fef" (UID: "458e6422-b856-4bcf-8308-4b6cb9ec8fef"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.431315 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/15c786bb-7353-4d06-b410-a44458a5f954-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"15c786bb-7353-4d06-b410-a44458a5f954\") " pod="openstack/rabbitmq-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.434261 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/458e6422-b856-4bcf-8308-4b6cb9ec8fef-pod-info" (OuterVolumeSpecName: "pod-info") pod "458e6422-b856-4bcf-8308-4b6cb9ec8fef" (UID: "458e6422-b856-4bcf-8308-4b6cb9ec8fef"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.436255 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/458e6422-b856-4bcf-8308-4b6cb9ec8fef-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "458e6422-b856-4bcf-8308-4b6cb9ec8fef" (UID: "458e6422-b856-4bcf-8308-4b6cb9ec8fef"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.449681 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/458e6422-b856-4bcf-8308-4b6cb9ec8fef-kube-api-access-dbjsk" (OuterVolumeSpecName: "kube-api-access-dbjsk") pod "458e6422-b856-4bcf-8308-4b6cb9ec8fef" (UID: "458e6422-b856-4bcf-8308-4b6cb9ec8fef"). InnerVolumeSpecName "kube-api-access-dbjsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.453221 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jl7p\" (UniqueName: \"kubernetes.io/projected/15c786bb-7353-4d06-b410-a44458a5f954-kube-api-access-7jl7p\") pod \"rabbitmq-server-0\" (UID: \"15c786bb-7353-4d06-b410-a44458a5f954\") " pod="openstack/rabbitmq-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.460958 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"15c786bb-7353-4d06-b410-a44458a5f954\") " pod="openstack/rabbitmq-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.468619 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/458e6422-b856-4bcf-8308-4b6cb9ec8fef-config-data" (OuterVolumeSpecName: "config-data") pod "458e6422-b856-4bcf-8308-4b6cb9ec8fef" (UID: "458e6422-b856-4bcf-8308-4b6cb9ec8fef"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.511094 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbjsk\" (UniqueName: \"kubernetes.io/projected/458e6422-b856-4bcf-8308-4b6cb9ec8fef-kube-api-access-dbjsk\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.511128 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/458e6422-b856-4bcf-8308-4b6cb9ec8fef-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.511138 4700 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/458e6422-b856-4bcf-8308-4b6cb9ec8fef-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.511147 4700 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/458e6422-b856-4bcf-8308-4b6cb9ec8fef-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.511155 4700 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/458e6422-b856-4bcf-8308-4b6cb9ec8fef-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.511167 4700 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/458e6422-b856-4bcf-8308-4b6cb9ec8fef-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.511174 4700 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/458e6422-b856-4bcf-8308-4b6cb9ec8fef-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.511218 4700 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.511226 4700 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/458e6422-b856-4bcf-8308-4b6cb9ec8fef-pod-info\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.516748 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/458e6422-b856-4bcf-8308-4b6cb9ec8fef-server-conf" (OuterVolumeSpecName: "server-conf") pod "458e6422-b856-4bcf-8308-4b6cb9ec8fef" (UID: "458e6422-b856-4bcf-8308-4b6cb9ec8fef"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.534904 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.549066 4700 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.571657 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/458e6422-b856-4bcf-8308-4b6cb9ec8fef-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "458e6422-b856-4bcf-8308-4b6cb9ec8fef" (UID: "458e6422-b856-4bcf-8308-4b6cb9ec8fef"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.612919 4700 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/458e6422-b856-4bcf-8308-4b6cb9ec8fef-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.612949 4700 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:39 crc kubenswrapper[4700]: I0227 17:25:39.612959 4700 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/458e6422-b856-4bcf-8308-4b6cb9ec8fef-server-conf\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.049581 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.140488 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"15c786bb-7353-4d06-b410-a44458a5f954","Type":"ContainerStarted","Data":"13d39877d946a9c7a2a0622ef79b171bb5885623f71c71793332197892b04a62"} Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.143836 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"458e6422-b856-4bcf-8308-4b6cb9ec8fef","Type":"ContainerDied","Data":"5b6a1b53951776cb57deade96773bdd54ce581139e41fe6311edce8336aacd0d"} Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.144174 4700 scope.go:117] "RemoveContainer" containerID="520ea869bacd9c0536d61aea44f09946e35bd5b34aac34d9e9100c9c2ed55787" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.143872 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.174010 4700 scope.go:117] "RemoveContainer" containerID="3c345536df56b1f58f851e4bdb0af75eeb80bcf3825e3bbbaf62f032bc472793" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.216275 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.234791 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.242916 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 27 17:25:40 crc kubenswrapper[4700]: E0227 17:25:40.243378 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="458e6422-b856-4bcf-8308-4b6cb9ec8fef" containerName="rabbitmq" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.243394 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="458e6422-b856-4bcf-8308-4b6cb9ec8fef" containerName="rabbitmq" Feb 27 17:25:40 crc kubenswrapper[4700]: E0227 17:25:40.243415 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="458e6422-b856-4bcf-8308-4b6cb9ec8fef" containerName="setup-container" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.243422 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="458e6422-b856-4bcf-8308-4b6cb9ec8fef" containerName="setup-container" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.243636 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="458e6422-b856-4bcf-8308-4b6cb9ec8fef" containerName="rabbitmq" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.244718 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.250081 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.250597 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.250706 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.250779 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.250994 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.251623 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.251757 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-xxmpg" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.251908 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.325878 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/43a2852f-4d8f-468c-988a-60dcac328039-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"43a2852f-4d8f-468c-988a-60dcac328039\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.325956 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98vv5\" (UniqueName: \"kubernetes.io/projected/43a2852f-4d8f-468c-988a-60dcac328039-kube-api-access-98vv5\") pod \"rabbitmq-cell1-server-0\" (UID: \"43a2852f-4d8f-468c-988a-60dcac328039\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.326013 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/43a2852f-4d8f-468c-988a-60dcac328039-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"43a2852f-4d8f-468c-988a-60dcac328039\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.326074 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/43a2852f-4d8f-468c-988a-60dcac328039-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"43a2852f-4d8f-468c-988a-60dcac328039\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.326091 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/43a2852f-4d8f-468c-988a-60dcac328039-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"43a2852f-4d8f-468c-988a-60dcac328039\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.326126 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/43a2852f-4d8f-468c-988a-60dcac328039-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"43a2852f-4d8f-468c-988a-60dcac328039\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.326141 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/43a2852f-4d8f-468c-988a-60dcac328039-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"43a2852f-4d8f-468c-988a-60dcac328039\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.326162 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/43a2852f-4d8f-468c-988a-60dcac328039-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"43a2852f-4d8f-468c-988a-60dcac328039\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.326341 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"43a2852f-4d8f-468c-988a-60dcac328039\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.326397 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/43a2852f-4d8f-468c-988a-60dcac328039-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"43a2852f-4d8f-468c-988a-60dcac328039\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.326451 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/43a2852f-4d8f-468c-988a-60dcac328039-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"43a2852f-4d8f-468c-988a-60dcac328039\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.427872 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/43a2852f-4d8f-468c-988a-60dcac328039-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"43a2852f-4d8f-468c-988a-60dcac328039\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.428010 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/43a2852f-4d8f-468c-988a-60dcac328039-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"43a2852f-4d8f-468c-988a-60dcac328039\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.428126 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/43a2852f-4d8f-468c-988a-60dcac328039-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"43a2852f-4d8f-468c-988a-60dcac328039\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.428175 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/43a2852f-4d8f-468c-988a-60dcac328039-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"43a2852f-4d8f-468c-988a-60dcac328039\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.428197 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/43a2852f-4d8f-468c-988a-60dcac328039-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"43a2852f-4d8f-468c-988a-60dcac328039\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.429429 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/43a2852f-4d8f-468c-988a-60dcac328039-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"43a2852f-4d8f-468c-988a-60dcac328039\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.429946 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"43a2852f-4d8f-468c-988a-60dcac328039\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.429990 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/43a2852f-4d8f-468c-988a-60dcac328039-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"43a2852f-4d8f-468c-988a-60dcac328039\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.430027 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/43a2852f-4d8f-468c-988a-60dcac328039-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"43a2852f-4d8f-468c-988a-60dcac328039\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.430119 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/43a2852f-4d8f-468c-988a-60dcac328039-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"43a2852f-4d8f-468c-988a-60dcac328039\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.430189 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98vv5\" (UniqueName: \"kubernetes.io/projected/43a2852f-4d8f-468c-988a-60dcac328039-kube-api-access-98vv5\") pod \"rabbitmq-cell1-server-0\" (UID: \"43a2852f-4d8f-468c-988a-60dcac328039\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.429355 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/43a2852f-4d8f-468c-988a-60dcac328039-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"43a2852f-4d8f-468c-988a-60dcac328039\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.430951 4700 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"43a2852f-4d8f-468c-988a-60dcac328039\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.432256 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/43a2852f-4d8f-468c-988a-60dcac328039-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"43a2852f-4d8f-468c-988a-60dcac328039\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.429857 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/43a2852f-4d8f-468c-988a-60dcac328039-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"43a2852f-4d8f-468c-988a-60dcac328039\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.432629 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/43a2852f-4d8f-468c-988a-60dcac328039-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"43a2852f-4d8f-468c-988a-60dcac328039\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.432957 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/43a2852f-4d8f-468c-988a-60dcac328039-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"43a2852f-4d8f-468c-988a-60dcac328039\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.432962 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/43a2852f-4d8f-468c-988a-60dcac328039-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"43a2852f-4d8f-468c-988a-60dcac328039\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.434080 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/43a2852f-4d8f-468c-988a-60dcac328039-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"43a2852f-4d8f-468c-988a-60dcac328039\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.435165 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/43a2852f-4d8f-468c-988a-60dcac328039-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"43a2852f-4d8f-468c-988a-60dcac328039\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.436076 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/43a2852f-4d8f-468c-988a-60dcac328039-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"43a2852f-4d8f-468c-988a-60dcac328039\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.449176 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98vv5\" (UniqueName: \"kubernetes.io/projected/43a2852f-4d8f-468c-988a-60dcac328039-kube-api-access-98vv5\") pod \"rabbitmq-cell1-server-0\" (UID: \"43a2852f-4d8f-468c-988a-60dcac328039\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.475137 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"43a2852f-4d8f-468c-988a-60dcac328039\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.602569 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:25:40 crc kubenswrapper[4700]: I0227 17:25:40.999169 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="458e6422-b856-4bcf-8308-4b6cb9ec8fef" path="/var/lib/kubelet/pods/458e6422-b856-4bcf-8308-4b6cb9ec8fef/volumes" Feb 27 17:25:41 crc kubenswrapper[4700]: I0227 17:25:41.001256 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efe7d637-da3a-4995-a40f-fae00257ac1d" path="/var/lib/kubelet/pods/efe7d637-da3a-4995-a40f-fae00257ac1d/volumes" Feb 27 17:25:41 crc kubenswrapper[4700]: I0227 17:25:41.128876 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 27 17:25:41 crc kubenswrapper[4700]: I0227 17:25:41.167893 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"43a2852f-4d8f-468c-988a-60dcac328039","Type":"ContainerStarted","Data":"74291d2b0e2f8d22c91f35c1b41657a1dd305009a2aee804e7d967145ec2edf8"} Feb 27 17:25:42 crc kubenswrapper[4700]: I0227 17:25:42.071320 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6bw84" Feb 27 17:25:42 crc kubenswrapper[4700]: I0227 17:25:42.074312 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6bw84" Feb 27 17:25:43 crc kubenswrapper[4700]: I0227 17:25:43.180924 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6bw84" podUID="77aad150-188b-472a-bae0-2f5f7011617a" containerName="registry-server" probeResult="failure" output=< Feb 27 17:25:43 crc kubenswrapper[4700]: timeout: failed to connect service ":50051" within 1s Feb 27 17:25:43 crc kubenswrapper[4700]: > Feb 27 17:25:43 crc kubenswrapper[4700]: I0227 17:25:43.211298 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"15c786bb-7353-4d06-b410-a44458a5f954","Type":"ContainerStarted","Data":"a2f26fc9d1979c04bd0885f2d6f1958a1bbbf31130e63c328f47e02c6adc1fe7"} Feb 27 17:25:44 crc kubenswrapper[4700]: I0227 17:25:44.224209 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"43a2852f-4d8f-468c-988a-60dcac328039","Type":"ContainerStarted","Data":"d5d66a235a89017e5a008ac40902ca259716009a8acc38ac2a27a24dee26973d"} Feb 27 17:25:47 crc kubenswrapper[4700]: I0227 17:25:47.575796 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77b594dc5c-ndvwx"] Feb 27 17:25:47 crc kubenswrapper[4700]: I0227 17:25:47.581402 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77b594dc5c-ndvwx" Feb 27 17:25:47 crc kubenswrapper[4700]: I0227 17:25:47.584053 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Feb 27 17:25:47 crc kubenswrapper[4700]: I0227 17:25:47.597445 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77b594dc5c-ndvwx"] Feb 27 17:25:47 crc kubenswrapper[4700]: I0227 17:25:47.727440 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c2ee20a-f966-4207-b52c-7b387989ddf5-dns-svc\") pod \"dnsmasq-dns-77b594dc5c-ndvwx\" (UID: \"0c2ee20a-f966-4207-b52c-7b387989ddf5\") " pod="openstack/dnsmasq-dns-77b594dc5c-ndvwx" Feb 27 17:25:47 crc kubenswrapper[4700]: I0227 17:25:47.727544 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c2ee20a-f966-4207-b52c-7b387989ddf5-config\") pod \"dnsmasq-dns-77b594dc5c-ndvwx\" (UID: \"0c2ee20a-f966-4207-b52c-7b387989ddf5\") " pod="openstack/dnsmasq-dns-77b594dc5c-ndvwx" Feb 27 17:25:47 crc kubenswrapper[4700]: I0227 17:25:47.727596 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8jdl\" (UniqueName: \"kubernetes.io/projected/0c2ee20a-f966-4207-b52c-7b387989ddf5-kube-api-access-v8jdl\") pod \"dnsmasq-dns-77b594dc5c-ndvwx\" (UID: \"0c2ee20a-f966-4207-b52c-7b387989ddf5\") " pod="openstack/dnsmasq-dns-77b594dc5c-ndvwx" Feb 27 17:25:47 crc kubenswrapper[4700]: I0227 17:25:47.727617 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0c2ee20a-f966-4207-b52c-7b387989ddf5-openstack-edpm-ipam\") pod \"dnsmasq-dns-77b594dc5c-ndvwx\" (UID: \"0c2ee20a-f966-4207-b52c-7b387989ddf5\") " pod="openstack/dnsmasq-dns-77b594dc5c-ndvwx" Feb 27 17:25:47 crc kubenswrapper[4700]: I0227 17:25:47.727873 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c2ee20a-f966-4207-b52c-7b387989ddf5-dns-swift-storage-0\") pod \"dnsmasq-dns-77b594dc5c-ndvwx\" (UID: \"0c2ee20a-f966-4207-b52c-7b387989ddf5\") " pod="openstack/dnsmasq-dns-77b594dc5c-ndvwx" Feb 27 17:25:47 crc kubenswrapper[4700]: I0227 17:25:47.727973 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c2ee20a-f966-4207-b52c-7b387989ddf5-ovsdbserver-sb\") pod \"dnsmasq-dns-77b594dc5c-ndvwx\" (UID: \"0c2ee20a-f966-4207-b52c-7b387989ddf5\") " pod="openstack/dnsmasq-dns-77b594dc5c-ndvwx" Feb 27 17:25:47 crc kubenswrapper[4700]: I0227 17:25:47.728023 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c2ee20a-f966-4207-b52c-7b387989ddf5-ovsdbserver-nb\") pod \"dnsmasq-dns-77b594dc5c-ndvwx\" (UID: \"0c2ee20a-f966-4207-b52c-7b387989ddf5\") " pod="openstack/dnsmasq-dns-77b594dc5c-ndvwx" Feb 27 17:25:47 crc kubenswrapper[4700]: I0227 17:25:47.830176 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8jdl\" (UniqueName: \"kubernetes.io/projected/0c2ee20a-f966-4207-b52c-7b387989ddf5-kube-api-access-v8jdl\") pod \"dnsmasq-dns-77b594dc5c-ndvwx\" (UID: \"0c2ee20a-f966-4207-b52c-7b387989ddf5\") " pod="openstack/dnsmasq-dns-77b594dc5c-ndvwx" Feb 27 17:25:47 crc kubenswrapper[4700]: I0227 17:25:47.830274 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0c2ee20a-f966-4207-b52c-7b387989ddf5-openstack-edpm-ipam\") pod \"dnsmasq-dns-77b594dc5c-ndvwx\" (UID: \"0c2ee20a-f966-4207-b52c-7b387989ddf5\") " pod="openstack/dnsmasq-dns-77b594dc5c-ndvwx" Feb 27 17:25:47 crc kubenswrapper[4700]: I0227 17:25:47.830524 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c2ee20a-f966-4207-b52c-7b387989ddf5-dns-swift-storage-0\") pod \"dnsmasq-dns-77b594dc5c-ndvwx\" (UID: \"0c2ee20a-f966-4207-b52c-7b387989ddf5\") " pod="openstack/dnsmasq-dns-77b594dc5c-ndvwx" Feb 27 17:25:47 crc kubenswrapper[4700]: I0227 17:25:47.830659 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c2ee20a-f966-4207-b52c-7b387989ddf5-ovsdbserver-sb\") pod \"dnsmasq-dns-77b594dc5c-ndvwx\" (UID: \"0c2ee20a-f966-4207-b52c-7b387989ddf5\") " pod="openstack/dnsmasq-dns-77b594dc5c-ndvwx" Feb 27 17:25:47 crc kubenswrapper[4700]: I0227 17:25:47.830727 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c2ee20a-f966-4207-b52c-7b387989ddf5-ovsdbserver-nb\") pod \"dnsmasq-dns-77b594dc5c-ndvwx\" (UID: \"0c2ee20a-f966-4207-b52c-7b387989ddf5\") " pod="openstack/dnsmasq-dns-77b594dc5c-ndvwx" Feb 27 17:25:47 crc kubenswrapper[4700]: I0227 17:25:47.830865 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c2ee20a-f966-4207-b52c-7b387989ddf5-dns-svc\") pod \"dnsmasq-dns-77b594dc5c-ndvwx\" (UID: \"0c2ee20a-f966-4207-b52c-7b387989ddf5\") " pod="openstack/dnsmasq-dns-77b594dc5c-ndvwx" Feb 27 17:25:47 crc kubenswrapper[4700]: I0227 17:25:47.831011 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c2ee20a-f966-4207-b52c-7b387989ddf5-config\") pod \"dnsmasq-dns-77b594dc5c-ndvwx\" (UID: \"0c2ee20a-f966-4207-b52c-7b387989ddf5\") " pod="openstack/dnsmasq-dns-77b594dc5c-ndvwx" Feb 27 17:25:47 crc kubenswrapper[4700]: I0227 17:25:47.832132 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c2ee20a-f966-4207-b52c-7b387989ddf5-ovsdbserver-nb\") pod \"dnsmasq-dns-77b594dc5c-ndvwx\" (UID: \"0c2ee20a-f966-4207-b52c-7b387989ddf5\") " pod="openstack/dnsmasq-dns-77b594dc5c-ndvwx" Feb 27 17:25:47 crc kubenswrapper[4700]: I0227 17:25:47.832259 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c2ee20a-f966-4207-b52c-7b387989ddf5-ovsdbserver-sb\") pod \"dnsmasq-dns-77b594dc5c-ndvwx\" (UID: \"0c2ee20a-f966-4207-b52c-7b387989ddf5\") " pod="openstack/dnsmasq-dns-77b594dc5c-ndvwx" Feb 27 17:25:47 crc kubenswrapper[4700]: I0227 17:25:47.832265 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0c2ee20a-f966-4207-b52c-7b387989ddf5-openstack-edpm-ipam\") pod \"dnsmasq-dns-77b594dc5c-ndvwx\" (UID: \"0c2ee20a-f966-4207-b52c-7b387989ddf5\") " pod="openstack/dnsmasq-dns-77b594dc5c-ndvwx" Feb 27 17:25:47 crc kubenswrapper[4700]: I0227 17:25:47.832511 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c2ee20a-f966-4207-b52c-7b387989ddf5-dns-swift-storage-0\") pod \"dnsmasq-dns-77b594dc5c-ndvwx\" (UID: \"0c2ee20a-f966-4207-b52c-7b387989ddf5\") " pod="openstack/dnsmasq-dns-77b594dc5c-ndvwx" Feb 27 17:25:47 crc kubenswrapper[4700]: I0227 17:25:47.832844 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c2ee20a-f966-4207-b52c-7b387989ddf5-dns-svc\") pod \"dnsmasq-dns-77b594dc5c-ndvwx\" (UID: \"0c2ee20a-f966-4207-b52c-7b387989ddf5\") " pod="openstack/dnsmasq-dns-77b594dc5c-ndvwx" Feb 27 17:25:47 crc kubenswrapper[4700]: I0227 17:25:47.833828 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c2ee20a-f966-4207-b52c-7b387989ddf5-config\") pod \"dnsmasq-dns-77b594dc5c-ndvwx\" (UID: \"0c2ee20a-f966-4207-b52c-7b387989ddf5\") " pod="openstack/dnsmasq-dns-77b594dc5c-ndvwx" Feb 27 17:25:47 crc kubenswrapper[4700]: I0227 17:25:47.852933 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8jdl\" (UniqueName: \"kubernetes.io/projected/0c2ee20a-f966-4207-b52c-7b387989ddf5-kube-api-access-v8jdl\") pod \"dnsmasq-dns-77b594dc5c-ndvwx\" (UID: \"0c2ee20a-f966-4207-b52c-7b387989ddf5\") " pod="openstack/dnsmasq-dns-77b594dc5c-ndvwx" Feb 27 17:25:47 crc kubenswrapper[4700]: I0227 17:25:47.908279 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77b594dc5c-ndvwx" Feb 27 17:25:48 crc kubenswrapper[4700]: I0227 17:25:48.452384 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77b594dc5c-ndvwx"] Feb 27 17:25:49 crc kubenswrapper[4700]: I0227 17:25:49.293663 4700 generic.go:334] "Generic (PLEG): container finished" podID="0c2ee20a-f966-4207-b52c-7b387989ddf5" containerID="a0caff94988c0f1b5123903aeafeeecd44188499ef02bacd206c5512b253325e" exitCode=0 Feb 27 17:25:49 crc kubenswrapper[4700]: I0227 17:25:49.294296 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77b594dc5c-ndvwx" event={"ID":"0c2ee20a-f966-4207-b52c-7b387989ddf5","Type":"ContainerDied","Data":"a0caff94988c0f1b5123903aeafeeecd44188499ef02bacd206c5512b253325e"} Feb 27 17:25:49 crc kubenswrapper[4700]: I0227 17:25:49.294330 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77b594dc5c-ndvwx" event={"ID":"0c2ee20a-f966-4207-b52c-7b387989ddf5","Type":"ContainerStarted","Data":"9f23e9e947e32a56850bd1330754915faec8fd997dbe2201782e2b5ad516e579"} Feb 27 17:25:50 crc kubenswrapper[4700]: I0227 17:25:50.309105 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77b594dc5c-ndvwx" event={"ID":"0c2ee20a-f966-4207-b52c-7b387989ddf5","Type":"ContainerStarted","Data":"aa083400b53a68fd0d45d930cbb2bf2f683ca648fe2d8f1536a51b0f4f774c56"} Feb 27 17:25:50 crc kubenswrapper[4700]: I0227 17:25:50.309702 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77b594dc5c-ndvwx" Feb 27 17:25:50 crc kubenswrapper[4700]: I0227 17:25:50.344523 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77b594dc5c-ndvwx" podStartSLOduration=3.34448468 podStartE2EDuration="3.34448468s" podCreationTimestamp="2026-02-27 17:25:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:25:50.333674935 +0000 UTC m=+1510.318987692" watchObservedRunningTime="2026-02-27 17:25:50.34448468 +0000 UTC m=+1510.329797477" Feb 27 17:25:51 crc kubenswrapper[4700]: I0227 17:25:51.655676 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6bw84" Feb 27 17:25:51 crc kubenswrapper[4700]: I0227 17:25:51.720611 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6bw84" Feb 27 17:25:51 crc kubenswrapper[4700]: I0227 17:25:51.933706 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6bw84"] Feb 27 17:25:53 crc kubenswrapper[4700]: I0227 17:25:53.342533 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6bw84" podUID="77aad150-188b-472a-bae0-2f5f7011617a" containerName="registry-server" containerID="cri-o://eb39d5946ec84e1def84398ed6d684d12fd97f698b941c8a89a0cd0d98bbd0cb" gracePeriod=2 Feb 27 17:25:53 crc kubenswrapper[4700]: I0227 17:25:53.912719 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6bw84" Feb 27 17:25:54 crc kubenswrapper[4700]: I0227 17:25:54.081369 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77aad150-188b-472a-bae0-2f5f7011617a-catalog-content\") pod \"77aad150-188b-472a-bae0-2f5f7011617a\" (UID: \"77aad150-188b-472a-bae0-2f5f7011617a\") " Feb 27 17:25:54 crc kubenswrapper[4700]: I0227 17:25:54.081511 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqq9f\" (UniqueName: \"kubernetes.io/projected/77aad150-188b-472a-bae0-2f5f7011617a-kube-api-access-qqq9f\") pod \"77aad150-188b-472a-bae0-2f5f7011617a\" (UID: \"77aad150-188b-472a-bae0-2f5f7011617a\") " Feb 27 17:25:54 crc kubenswrapper[4700]: I0227 17:25:54.081598 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77aad150-188b-472a-bae0-2f5f7011617a-utilities\") pod \"77aad150-188b-472a-bae0-2f5f7011617a\" (UID: \"77aad150-188b-472a-bae0-2f5f7011617a\") " Feb 27 17:25:54 crc kubenswrapper[4700]: I0227 17:25:54.083300 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77aad150-188b-472a-bae0-2f5f7011617a-utilities" (OuterVolumeSpecName: "utilities") pod "77aad150-188b-472a-bae0-2f5f7011617a" (UID: "77aad150-188b-472a-bae0-2f5f7011617a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:25:54 crc kubenswrapper[4700]: I0227 17:25:54.088408 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77aad150-188b-472a-bae0-2f5f7011617a-kube-api-access-qqq9f" (OuterVolumeSpecName: "kube-api-access-qqq9f") pod "77aad150-188b-472a-bae0-2f5f7011617a" (UID: "77aad150-188b-472a-bae0-2f5f7011617a"). InnerVolumeSpecName "kube-api-access-qqq9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:25:54 crc kubenswrapper[4700]: I0227 17:25:54.185469 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqq9f\" (UniqueName: \"kubernetes.io/projected/77aad150-188b-472a-bae0-2f5f7011617a-kube-api-access-qqq9f\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:54 crc kubenswrapper[4700]: I0227 17:25:54.185505 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77aad150-188b-472a-bae0-2f5f7011617a-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:54 crc kubenswrapper[4700]: I0227 17:25:54.260775 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77aad150-188b-472a-bae0-2f5f7011617a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "77aad150-188b-472a-bae0-2f5f7011617a" (UID: "77aad150-188b-472a-bae0-2f5f7011617a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:25:54 crc kubenswrapper[4700]: I0227 17:25:54.287189 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77aad150-188b-472a-bae0-2f5f7011617a-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:54 crc kubenswrapper[4700]: I0227 17:25:54.356946 4700 generic.go:334] "Generic (PLEG): container finished" podID="77aad150-188b-472a-bae0-2f5f7011617a" containerID="eb39d5946ec84e1def84398ed6d684d12fd97f698b941c8a89a0cd0d98bbd0cb" exitCode=0 Feb 27 17:25:54 crc kubenswrapper[4700]: I0227 17:25:54.357020 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6bw84" event={"ID":"77aad150-188b-472a-bae0-2f5f7011617a","Type":"ContainerDied","Data":"eb39d5946ec84e1def84398ed6d684d12fd97f698b941c8a89a0cd0d98bbd0cb"} Feb 27 17:25:54 crc kubenswrapper[4700]: I0227 17:25:54.357111 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6bw84" event={"ID":"77aad150-188b-472a-bae0-2f5f7011617a","Type":"ContainerDied","Data":"d7ac175a43c8cade8c56d01540be666f6fd175cac23298d596d7132ad813d28c"} Feb 27 17:25:54 crc kubenswrapper[4700]: I0227 17:25:54.357144 4700 scope.go:117] "RemoveContainer" containerID="eb39d5946ec84e1def84398ed6d684d12fd97f698b941c8a89a0cd0d98bbd0cb" Feb 27 17:25:54 crc kubenswrapper[4700]: I0227 17:25:54.357051 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6bw84" Feb 27 17:25:54 crc kubenswrapper[4700]: I0227 17:25:54.395537 4700 scope.go:117] "RemoveContainer" containerID="ce1d734fc2fecdbcabe47e8046e0a4948755e41754dab4f61c5bb5f3fe38110f" Feb 27 17:25:54 crc kubenswrapper[4700]: I0227 17:25:54.428854 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6bw84"] Feb 27 17:25:54 crc kubenswrapper[4700]: I0227 17:25:54.445069 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6bw84"] Feb 27 17:25:54 crc kubenswrapper[4700]: I0227 17:25:54.456276 4700 scope.go:117] "RemoveContainer" containerID="bc133afdf262414257afca6233a3ba754f8358fc335a7bca6d17f753bc9aada6" Feb 27 17:25:54 crc kubenswrapper[4700]: I0227 17:25:54.517216 4700 scope.go:117] "RemoveContainer" containerID="eb39d5946ec84e1def84398ed6d684d12fd97f698b941c8a89a0cd0d98bbd0cb" Feb 27 17:25:54 crc kubenswrapper[4700]: E0227 17:25:54.517808 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb39d5946ec84e1def84398ed6d684d12fd97f698b941c8a89a0cd0d98bbd0cb\": container with ID starting with eb39d5946ec84e1def84398ed6d684d12fd97f698b941c8a89a0cd0d98bbd0cb not found: ID does not exist" containerID="eb39d5946ec84e1def84398ed6d684d12fd97f698b941c8a89a0cd0d98bbd0cb" Feb 27 17:25:54 crc kubenswrapper[4700]: I0227 17:25:54.517873 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb39d5946ec84e1def84398ed6d684d12fd97f698b941c8a89a0cd0d98bbd0cb"} err="failed to get container status \"eb39d5946ec84e1def84398ed6d684d12fd97f698b941c8a89a0cd0d98bbd0cb\": rpc error: code = NotFound desc = could not find container \"eb39d5946ec84e1def84398ed6d684d12fd97f698b941c8a89a0cd0d98bbd0cb\": container with ID starting with eb39d5946ec84e1def84398ed6d684d12fd97f698b941c8a89a0cd0d98bbd0cb not found: ID does not exist" Feb 27 17:25:54 crc kubenswrapper[4700]: I0227 17:25:54.517915 4700 scope.go:117] "RemoveContainer" containerID="ce1d734fc2fecdbcabe47e8046e0a4948755e41754dab4f61c5bb5f3fe38110f" Feb 27 17:25:54 crc kubenswrapper[4700]: E0227 17:25:54.518557 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce1d734fc2fecdbcabe47e8046e0a4948755e41754dab4f61c5bb5f3fe38110f\": container with ID starting with ce1d734fc2fecdbcabe47e8046e0a4948755e41754dab4f61c5bb5f3fe38110f not found: ID does not exist" containerID="ce1d734fc2fecdbcabe47e8046e0a4948755e41754dab4f61c5bb5f3fe38110f" Feb 27 17:25:54 crc kubenswrapper[4700]: I0227 17:25:54.518607 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce1d734fc2fecdbcabe47e8046e0a4948755e41754dab4f61c5bb5f3fe38110f"} err="failed to get container status \"ce1d734fc2fecdbcabe47e8046e0a4948755e41754dab4f61c5bb5f3fe38110f\": rpc error: code = NotFound desc = could not find container \"ce1d734fc2fecdbcabe47e8046e0a4948755e41754dab4f61c5bb5f3fe38110f\": container with ID starting with ce1d734fc2fecdbcabe47e8046e0a4948755e41754dab4f61c5bb5f3fe38110f not found: ID does not exist" Feb 27 17:25:54 crc kubenswrapper[4700]: I0227 17:25:54.518644 4700 scope.go:117] "RemoveContainer" containerID="bc133afdf262414257afca6233a3ba754f8358fc335a7bca6d17f753bc9aada6" Feb 27 17:25:54 crc kubenswrapper[4700]: E0227 17:25:54.519763 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc133afdf262414257afca6233a3ba754f8358fc335a7bca6d17f753bc9aada6\": container with ID starting with bc133afdf262414257afca6233a3ba754f8358fc335a7bca6d17f753bc9aada6 not found: ID does not exist" containerID="bc133afdf262414257afca6233a3ba754f8358fc335a7bca6d17f753bc9aada6" Feb 27 17:25:54 crc kubenswrapper[4700]: I0227 17:25:54.519809 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc133afdf262414257afca6233a3ba754f8358fc335a7bca6d17f753bc9aada6"} err="failed to get container status \"bc133afdf262414257afca6233a3ba754f8358fc335a7bca6d17f753bc9aada6\": rpc error: code = NotFound desc = could not find container \"bc133afdf262414257afca6233a3ba754f8358fc335a7bca6d17f753bc9aada6\": container with ID starting with bc133afdf262414257afca6233a3ba754f8358fc335a7bca6d17f753bc9aada6 not found: ID does not exist" Feb 27 17:25:55 crc kubenswrapper[4700]: I0227 17:25:55.001061 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77aad150-188b-472a-bae0-2f5f7011617a" path="/var/lib/kubelet/pods/77aad150-188b-472a-bae0-2f5f7011617a/volumes" Feb 27 17:25:57 crc kubenswrapper[4700]: I0227 17:25:57.909746 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77b594dc5c-ndvwx" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.026519 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d8cd4fdf7-74zlx"] Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.026906 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5d8cd4fdf7-74zlx" podUID="b896d6a2-a26c-4b15-adcf-bb73998922f2" containerName="dnsmasq-dns" containerID="cri-o://2d1317865a2ce03538491b00c99808acf65eedb1a3994f3245a6fc4013fc05b4" gracePeriod=10 Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.179561 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cfbbd847f-bd2kg"] Feb 27 17:25:58 crc kubenswrapper[4700]: E0227 17:25:58.179994 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77aad150-188b-472a-bae0-2f5f7011617a" containerName="extract-content" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.180010 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="77aad150-188b-472a-bae0-2f5f7011617a" containerName="extract-content" Feb 27 17:25:58 crc kubenswrapper[4700]: E0227 17:25:58.180022 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77aad150-188b-472a-bae0-2f5f7011617a" containerName="registry-server" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.180029 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="77aad150-188b-472a-bae0-2f5f7011617a" containerName="registry-server" Feb 27 17:25:58 crc kubenswrapper[4700]: E0227 17:25:58.180058 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77aad150-188b-472a-bae0-2f5f7011617a" containerName="extract-utilities" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.180065 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="77aad150-188b-472a-bae0-2f5f7011617a" containerName="extract-utilities" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.180239 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="77aad150-188b-472a-bae0-2f5f7011617a" containerName="registry-server" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.181267 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cfbbd847f-bd2kg" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.210496 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cfbbd847f-bd2kg"] Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.287359 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eaac16a0-f258-43bb-9c2f-ca23dec08397-dns-swift-storage-0\") pod \"dnsmasq-dns-cfbbd847f-bd2kg\" (UID: \"eaac16a0-f258-43bb-9c2f-ca23dec08397\") " pod="openstack/dnsmasq-dns-cfbbd847f-bd2kg" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.287409 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eaac16a0-f258-43bb-9c2f-ca23dec08397-ovsdbserver-nb\") pod \"dnsmasq-dns-cfbbd847f-bd2kg\" (UID: \"eaac16a0-f258-43bb-9c2f-ca23dec08397\") " pod="openstack/dnsmasq-dns-cfbbd847f-bd2kg" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.287457 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/eaac16a0-f258-43bb-9c2f-ca23dec08397-openstack-edpm-ipam\") pod \"dnsmasq-dns-cfbbd847f-bd2kg\" (UID: \"eaac16a0-f258-43bb-9c2f-ca23dec08397\") " pod="openstack/dnsmasq-dns-cfbbd847f-bd2kg" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.287689 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eaac16a0-f258-43bb-9c2f-ca23dec08397-dns-svc\") pod \"dnsmasq-dns-cfbbd847f-bd2kg\" (UID: \"eaac16a0-f258-43bb-9c2f-ca23dec08397\") " pod="openstack/dnsmasq-dns-cfbbd847f-bd2kg" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.287715 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4mjt\" (UniqueName: \"kubernetes.io/projected/eaac16a0-f258-43bb-9c2f-ca23dec08397-kube-api-access-z4mjt\") pod \"dnsmasq-dns-cfbbd847f-bd2kg\" (UID: \"eaac16a0-f258-43bb-9c2f-ca23dec08397\") " pod="openstack/dnsmasq-dns-cfbbd847f-bd2kg" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.287738 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eaac16a0-f258-43bb-9c2f-ca23dec08397-config\") pod \"dnsmasq-dns-cfbbd847f-bd2kg\" (UID: \"eaac16a0-f258-43bb-9c2f-ca23dec08397\") " pod="openstack/dnsmasq-dns-cfbbd847f-bd2kg" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.287825 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eaac16a0-f258-43bb-9c2f-ca23dec08397-ovsdbserver-sb\") pod \"dnsmasq-dns-cfbbd847f-bd2kg\" (UID: \"eaac16a0-f258-43bb-9c2f-ca23dec08397\") " pod="openstack/dnsmasq-dns-cfbbd847f-bd2kg" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.389260 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eaac16a0-f258-43bb-9c2f-ca23dec08397-ovsdbserver-nb\") pod \"dnsmasq-dns-cfbbd847f-bd2kg\" (UID: \"eaac16a0-f258-43bb-9c2f-ca23dec08397\") " pod="openstack/dnsmasq-dns-cfbbd847f-bd2kg" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.389316 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/eaac16a0-f258-43bb-9c2f-ca23dec08397-openstack-edpm-ipam\") pod \"dnsmasq-dns-cfbbd847f-bd2kg\" (UID: \"eaac16a0-f258-43bb-9c2f-ca23dec08397\") " pod="openstack/dnsmasq-dns-cfbbd847f-bd2kg" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.389448 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eaac16a0-f258-43bb-9c2f-ca23dec08397-dns-svc\") pod \"dnsmasq-dns-cfbbd847f-bd2kg\" (UID: \"eaac16a0-f258-43bb-9c2f-ca23dec08397\") " pod="openstack/dnsmasq-dns-cfbbd847f-bd2kg" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.389492 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4mjt\" (UniqueName: \"kubernetes.io/projected/eaac16a0-f258-43bb-9c2f-ca23dec08397-kube-api-access-z4mjt\") pod \"dnsmasq-dns-cfbbd847f-bd2kg\" (UID: \"eaac16a0-f258-43bb-9c2f-ca23dec08397\") " pod="openstack/dnsmasq-dns-cfbbd847f-bd2kg" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.389519 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eaac16a0-f258-43bb-9c2f-ca23dec08397-config\") pod \"dnsmasq-dns-cfbbd847f-bd2kg\" (UID: \"eaac16a0-f258-43bb-9c2f-ca23dec08397\") " pod="openstack/dnsmasq-dns-cfbbd847f-bd2kg" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.389588 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eaac16a0-f258-43bb-9c2f-ca23dec08397-ovsdbserver-sb\") pod \"dnsmasq-dns-cfbbd847f-bd2kg\" (UID: \"eaac16a0-f258-43bb-9c2f-ca23dec08397\") " pod="openstack/dnsmasq-dns-cfbbd847f-bd2kg" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.389659 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eaac16a0-f258-43bb-9c2f-ca23dec08397-dns-swift-storage-0\") pod \"dnsmasq-dns-cfbbd847f-bd2kg\" (UID: \"eaac16a0-f258-43bb-9c2f-ca23dec08397\") " pod="openstack/dnsmasq-dns-cfbbd847f-bd2kg" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.390703 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eaac16a0-f258-43bb-9c2f-ca23dec08397-dns-swift-storage-0\") pod \"dnsmasq-dns-cfbbd847f-bd2kg\" (UID: \"eaac16a0-f258-43bb-9c2f-ca23dec08397\") " pod="openstack/dnsmasq-dns-cfbbd847f-bd2kg" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.390839 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eaac16a0-f258-43bb-9c2f-ca23dec08397-ovsdbserver-nb\") pod \"dnsmasq-dns-cfbbd847f-bd2kg\" (UID: \"eaac16a0-f258-43bb-9c2f-ca23dec08397\") " pod="openstack/dnsmasq-dns-cfbbd847f-bd2kg" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.391378 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eaac16a0-f258-43bb-9c2f-ca23dec08397-ovsdbserver-sb\") pod \"dnsmasq-dns-cfbbd847f-bd2kg\" (UID: \"eaac16a0-f258-43bb-9c2f-ca23dec08397\") " pod="openstack/dnsmasq-dns-cfbbd847f-bd2kg" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.392228 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/eaac16a0-f258-43bb-9c2f-ca23dec08397-openstack-edpm-ipam\") pod \"dnsmasq-dns-cfbbd847f-bd2kg\" (UID: \"eaac16a0-f258-43bb-9c2f-ca23dec08397\") " pod="openstack/dnsmasq-dns-cfbbd847f-bd2kg" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.392396 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eaac16a0-f258-43bb-9c2f-ca23dec08397-dns-svc\") pod \"dnsmasq-dns-cfbbd847f-bd2kg\" (UID: \"eaac16a0-f258-43bb-9c2f-ca23dec08397\") " pod="openstack/dnsmasq-dns-cfbbd847f-bd2kg" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.393775 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eaac16a0-f258-43bb-9c2f-ca23dec08397-config\") pod \"dnsmasq-dns-cfbbd847f-bd2kg\" (UID: \"eaac16a0-f258-43bb-9c2f-ca23dec08397\") " pod="openstack/dnsmasq-dns-cfbbd847f-bd2kg" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.416336 4700 generic.go:334] "Generic (PLEG): container finished" podID="b896d6a2-a26c-4b15-adcf-bb73998922f2" containerID="2d1317865a2ce03538491b00c99808acf65eedb1a3994f3245a6fc4013fc05b4" exitCode=0 Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.416381 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d8cd4fdf7-74zlx" event={"ID":"b896d6a2-a26c-4b15-adcf-bb73998922f2","Type":"ContainerDied","Data":"2d1317865a2ce03538491b00c99808acf65eedb1a3994f3245a6fc4013fc05b4"} Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.419736 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4mjt\" (UniqueName: \"kubernetes.io/projected/eaac16a0-f258-43bb-9c2f-ca23dec08397-kube-api-access-z4mjt\") pod \"dnsmasq-dns-cfbbd847f-bd2kg\" (UID: \"eaac16a0-f258-43bb-9c2f-ca23dec08397\") " pod="openstack/dnsmasq-dns-cfbbd847f-bd2kg" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.532000 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cfbbd847f-bd2kg" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.616854 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d8cd4fdf7-74zlx" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.699408 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b896d6a2-a26c-4b15-adcf-bb73998922f2-config\") pod \"b896d6a2-a26c-4b15-adcf-bb73998922f2\" (UID: \"b896d6a2-a26c-4b15-adcf-bb73998922f2\") " Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.699803 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b896d6a2-a26c-4b15-adcf-bb73998922f2-dns-swift-storage-0\") pod \"b896d6a2-a26c-4b15-adcf-bb73998922f2\" (UID: \"b896d6a2-a26c-4b15-adcf-bb73998922f2\") " Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.699949 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b896d6a2-a26c-4b15-adcf-bb73998922f2-ovsdbserver-nb\") pod \"b896d6a2-a26c-4b15-adcf-bb73998922f2\" (UID: \"b896d6a2-a26c-4b15-adcf-bb73998922f2\") " Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.699975 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvx5p\" (UniqueName: \"kubernetes.io/projected/b896d6a2-a26c-4b15-adcf-bb73998922f2-kube-api-access-nvx5p\") pod \"b896d6a2-a26c-4b15-adcf-bb73998922f2\" (UID: \"b896d6a2-a26c-4b15-adcf-bb73998922f2\") " Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.700107 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b896d6a2-a26c-4b15-adcf-bb73998922f2-ovsdbserver-sb\") pod \"b896d6a2-a26c-4b15-adcf-bb73998922f2\" (UID: \"b896d6a2-a26c-4b15-adcf-bb73998922f2\") " Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.700139 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b896d6a2-a26c-4b15-adcf-bb73998922f2-dns-svc\") pod \"b896d6a2-a26c-4b15-adcf-bb73998922f2\" (UID: \"b896d6a2-a26c-4b15-adcf-bb73998922f2\") " Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.704055 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b896d6a2-a26c-4b15-adcf-bb73998922f2-kube-api-access-nvx5p" (OuterVolumeSpecName: "kube-api-access-nvx5p") pod "b896d6a2-a26c-4b15-adcf-bb73998922f2" (UID: "b896d6a2-a26c-4b15-adcf-bb73998922f2"). InnerVolumeSpecName "kube-api-access-nvx5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.777092 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b896d6a2-a26c-4b15-adcf-bb73998922f2-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b896d6a2-a26c-4b15-adcf-bb73998922f2" (UID: "b896d6a2-a26c-4b15-adcf-bb73998922f2"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.780478 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b896d6a2-a26c-4b15-adcf-bb73998922f2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b896d6a2-a26c-4b15-adcf-bb73998922f2" (UID: "b896d6a2-a26c-4b15-adcf-bb73998922f2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.789645 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b896d6a2-a26c-4b15-adcf-bb73998922f2-config" (OuterVolumeSpecName: "config") pod "b896d6a2-a26c-4b15-adcf-bb73998922f2" (UID: "b896d6a2-a26c-4b15-adcf-bb73998922f2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.794602 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b896d6a2-a26c-4b15-adcf-bb73998922f2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b896d6a2-a26c-4b15-adcf-bb73998922f2" (UID: "b896d6a2-a26c-4b15-adcf-bb73998922f2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.802535 4700 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b896d6a2-a26c-4b15-adcf-bb73998922f2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.802561 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b896d6a2-a26c-4b15-adcf-bb73998922f2-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.802571 4700 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b896d6a2-a26c-4b15-adcf-bb73998922f2-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.802579 4700 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b896d6a2-a26c-4b15-adcf-bb73998922f2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.802590 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvx5p\" (UniqueName: \"kubernetes.io/projected/b896d6a2-a26c-4b15-adcf-bb73998922f2-kube-api-access-nvx5p\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.804648 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b896d6a2-a26c-4b15-adcf-bb73998922f2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b896d6a2-a26c-4b15-adcf-bb73998922f2" (UID: "b896d6a2-a26c-4b15-adcf-bb73998922f2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:25:58 crc kubenswrapper[4700]: I0227 17:25:58.905245 4700 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b896d6a2-a26c-4b15-adcf-bb73998922f2-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 17:25:59 crc kubenswrapper[4700]: I0227 17:25:59.034222 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cfbbd847f-bd2kg"] Feb 27 17:25:59 crc kubenswrapper[4700]: W0227 17:25:59.034341 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeaac16a0_f258_43bb_9c2f_ca23dec08397.slice/crio-3ed756b3e70ad005a2136217e2a4ff8634b8293f9b7f7ff77434eb5e6114669a WatchSource:0}: Error finding container 3ed756b3e70ad005a2136217e2a4ff8634b8293f9b7f7ff77434eb5e6114669a: Status 404 returned error can't find the container with id 3ed756b3e70ad005a2136217e2a4ff8634b8293f9b7f7ff77434eb5e6114669a Feb 27 17:25:59 crc kubenswrapper[4700]: I0227 17:25:59.428377 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d8cd4fdf7-74zlx" event={"ID":"b896d6a2-a26c-4b15-adcf-bb73998922f2","Type":"ContainerDied","Data":"468bfd23b39aec2505dfb99e2b91f26af3a75f29bd3b2c5fdd537b97f41d6c82"} Feb 27 17:25:59 crc kubenswrapper[4700]: I0227 17:25:59.428409 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d8cd4fdf7-74zlx" Feb 27 17:25:59 crc kubenswrapper[4700]: I0227 17:25:59.428445 4700 scope.go:117] "RemoveContainer" containerID="2d1317865a2ce03538491b00c99808acf65eedb1a3994f3245a6fc4013fc05b4" Feb 27 17:25:59 crc kubenswrapper[4700]: I0227 17:25:59.430415 4700 generic.go:334] "Generic (PLEG): container finished" podID="eaac16a0-f258-43bb-9c2f-ca23dec08397" containerID="6c165f52f905fba16ef4b8b7e21ad0efe79eddc8c8767ebfbcc42e1304a3769f" exitCode=0 Feb 27 17:25:59 crc kubenswrapper[4700]: I0227 17:25:59.430451 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cfbbd847f-bd2kg" event={"ID":"eaac16a0-f258-43bb-9c2f-ca23dec08397","Type":"ContainerDied","Data":"6c165f52f905fba16ef4b8b7e21ad0efe79eddc8c8767ebfbcc42e1304a3769f"} Feb 27 17:25:59 crc kubenswrapper[4700]: I0227 17:25:59.430488 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cfbbd847f-bd2kg" event={"ID":"eaac16a0-f258-43bb-9c2f-ca23dec08397","Type":"ContainerStarted","Data":"3ed756b3e70ad005a2136217e2a4ff8634b8293f9b7f7ff77434eb5e6114669a"} Feb 27 17:25:59 crc kubenswrapper[4700]: I0227 17:25:59.467517 4700 scope.go:117] "RemoveContainer" containerID="9d5028a387e275812362f6e533e459777ff21a701a1c370699589c2a8fa193ac" Feb 27 17:25:59 crc kubenswrapper[4700]: I0227 17:25:59.470149 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d8cd4fdf7-74zlx"] Feb 27 17:25:59 crc kubenswrapper[4700]: I0227 17:25:59.496432 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d8cd4fdf7-74zlx"] Feb 27 17:26:00 crc kubenswrapper[4700]: I0227 17:26:00.146262 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536886-87g88"] Feb 27 17:26:00 crc kubenswrapper[4700]: E0227 17:26:00.146846 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b896d6a2-a26c-4b15-adcf-bb73998922f2" containerName="dnsmasq-dns" Feb 27 17:26:00 crc kubenswrapper[4700]: I0227 17:26:00.146861 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="b896d6a2-a26c-4b15-adcf-bb73998922f2" containerName="dnsmasq-dns" Feb 27 17:26:00 crc kubenswrapper[4700]: E0227 17:26:00.146902 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b896d6a2-a26c-4b15-adcf-bb73998922f2" containerName="init" Feb 27 17:26:00 crc kubenswrapper[4700]: I0227 17:26:00.146911 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="b896d6a2-a26c-4b15-adcf-bb73998922f2" containerName="init" Feb 27 17:26:00 crc kubenswrapper[4700]: I0227 17:26:00.147177 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="b896d6a2-a26c-4b15-adcf-bb73998922f2" containerName="dnsmasq-dns" Feb 27 17:26:00 crc kubenswrapper[4700]: I0227 17:26:00.148286 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536886-87g88" Feb 27 17:26:00 crc kubenswrapper[4700]: I0227 17:26:00.150912 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:26:00 crc kubenswrapper[4700]: I0227 17:26:00.154941 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d62kc" Feb 27 17:26:00 crc kubenswrapper[4700]: I0227 17:26:00.157342 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:26:00 crc kubenswrapper[4700]: I0227 17:26:00.158371 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536886-87g88"] Feb 27 17:26:00 crc kubenswrapper[4700]: I0227 17:26:00.230816 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g28jq\" (UniqueName: \"kubernetes.io/projected/a4b01082-c066-4fbd-aef5-169e9ec31fff-kube-api-access-g28jq\") pod \"auto-csr-approver-29536886-87g88\" (UID: \"a4b01082-c066-4fbd-aef5-169e9ec31fff\") " pod="openshift-infra/auto-csr-approver-29536886-87g88" Feb 27 17:26:00 crc kubenswrapper[4700]: I0227 17:26:00.333497 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g28jq\" (UniqueName: \"kubernetes.io/projected/a4b01082-c066-4fbd-aef5-169e9ec31fff-kube-api-access-g28jq\") pod \"auto-csr-approver-29536886-87g88\" (UID: \"a4b01082-c066-4fbd-aef5-169e9ec31fff\") " pod="openshift-infra/auto-csr-approver-29536886-87g88" Feb 27 17:26:00 crc kubenswrapper[4700]: I0227 17:26:00.364260 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g28jq\" (UniqueName: \"kubernetes.io/projected/a4b01082-c066-4fbd-aef5-169e9ec31fff-kube-api-access-g28jq\") pod \"auto-csr-approver-29536886-87g88\" (UID: \"a4b01082-c066-4fbd-aef5-169e9ec31fff\") " pod="openshift-infra/auto-csr-approver-29536886-87g88" Feb 27 17:26:00 crc kubenswrapper[4700]: I0227 17:26:00.442776 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cfbbd847f-bd2kg" event={"ID":"eaac16a0-f258-43bb-9c2f-ca23dec08397","Type":"ContainerStarted","Data":"f630109401a234c0645abbdb3c0126896651a4d07cd6581e5f49c51d36c9e8e1"} Feb 27 17:26:00 crc kubenswrapper[4700]: I0227 17:26:00.443122 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cfbbd847f-bd2kg" Feb 27 17:26:00 crc kubenswrapper[4700]: I0227 17:26:00.468151 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cfbbd847f-bd2kg" podStartSLOduration=2.468128822 podStartE2EDuration="2.468128822s" podCreationTimestamp="2026-02-27 17:25:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:26:00.467314421 +0000 UTC m=+1520.452627168" watchObservedRunningTime="2026-02-27 17:26:00.468128822 +0000 UTC m=+1520.453441589" Feb 27 17:26:00 crc kubenswrapper[4700]: I0227 17:26:00.510393 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536886-87g88" Feb 27 17:26:00 crc kubenswrapper[4700]: I0227 17:26:00.971027 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536886-87g88"] Feb 27 17:26:01 crc kubenswrapper[4700]: I0227 17:26:01.007379 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b896d6a2-a26c-4b15-adcf-bb73998922f2" path="/var/lib/kubelet/pods/b896d6a2-a26c-4b15-adcf-bb73998922f2/volumes" Feb 27 17:26:01 crc kubenswrapper[4700]: I0227 17:26:01.457782 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536886-87g88" event={"ID":"a4b01082-c066-4fbd-aef5-169e9ec31fff","Type":"ContainerStarted","Data":"93c0db3011f0ca04624bff5b9a7f8e7500afe9acb6d218cf58288748c210141b"} Feb 27 17:26:02 crc kubenswrapper[4700]: I0227 17:26:02.483158 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536886-87g88" event={"ID":"a4b01082-c066-4fbd-aef5-169e9ec31fff","Type":"ContainerStarted","Data":"3cc4589589d2611cf5d42961ae0bae28311f2045634dd9d7392d4f4d2a990485"} Feb 27 17:26:02 crc kubenswrapper[4700]: I0227 17:26:02.500687 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536886-87g88" podStartSLOduration=1.616525997 podStartE2EDuration="2.500669892s" podCreationTimestamp="2026-02-27 17:26:00 +0000 UTC" firstStartedPulling="2026-02-27 17:26:00.992913794 +0000 UTC m=+1520.978226541" lastFinishedPulling="2026-02-27 17:26:01.877057649 +0000 UTC m=+1521.862370436" observedRunningTime="2026-02-27 17:26:02.494553091 +0000 UTC m=+1522.479865848" watchObservedRunningTime="2026-02-27 17:26:02.500669892 +0000 UTC m=+1522.485982639" Feb 27 17:26:03 crc kubenswrapper[4700]: I0227 17:26:03.496832 4700 generic.go:334] "Generic (PLEG): container finished" podID="a4b01082-c066-4fbd-aef5-169e9ec31fff" containerID="3cc4589589d2611cf5d42961ae0bae28311f2045634dd9d7392d4f4d2a990485" exitCode=0 Feb 27 17:26:03 crc kubenswrapper[4700]: I0227 17:26:03.497004 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536886-87g88" event={"ID":"a4b01082-c066-4fbd-aef5-169e9ec31fff","Type":"ContainerDied","Data":"3cc4589589d2611cf5d42961ae0bae28311f2045634dd9d7392d4f4d2a990485"} Feb 27 17:26:04 crc kubenswrapper[4700]: I0227 17:26:04.991874 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536886-87g88" Feb 27 17:26:05 crc kubenswrapper[4700]: I0227 17:26:05.143426 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g28jq\" (UniqueName: \"kubernetes.io/projected/a4b01082-c066-4fbd-aef5-169e9ec31fff-kube-api-access-g28jq\") pod \"a4b01082-c066-4fbd-aef5-169e9ec31fff\" (UID: \"a4b01082-c066-4fbd-aef5-169e9ec31fff\") " Feb 27 17:26:05 crc kubenswrapper[4700]: I0227 17:26:05.156807 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4b01082-c066-4fbd-aef5-169e9ec31fff-kube-api-access-g28jq" (OuterVolumeSpecName: "kube-api-access-g28jq") pod "a4b01082-c066-4fbd-aef5-169e9ec31fff" (UID: "a4b01082-c066-4fbd-aef5-169e9ec31fff"). InnerVolumeSpecName "kube-api-access-g28jq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:26:05 crc kubenswrapper[4700]: I0227 17:26:05.246780 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g28jq\" (UniqueName: \"kubernetes.io/projected/a4b01082-c066-4fbd-aef5-169e9ec31fff-kube-api-access-g28jq\") on node \"crc\" DevicePath \"\"" Feb 27 17:26:05 crc kubenswrapper[4700]: I0227 17:26:05.527976 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536886-87g88" event={"ID":"a4b01082-c066-4fbd-aef5-169e9ec31fff","Type":"ContainerDied","Data":"93c0db3011f0ca04624bff5b9a7f8e7500afe9acb6d218cf58288748c210141b"} Feb 27 17:26:05 crc kubenswrapper[4700]: I0227 17:26:05.528273 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="93c0db3011f0ca04624bff5b9a7f8e7500afe9acb6d218cf58288748c210141b" Feb 27 17:26:05 crc kubenswrapper[4700]: I0227 17:26:05.528324 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536886-87g88" Feb 27 17:26:05 crc kubenswrapper[4700]: I0227 17:26:05.601808 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536880-7gbpd"] Feb 27 17:26:05 crc kubenswrapper[4700]: I0227 17:26:05.624352 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536880-7gbpd"] Feb 27 17:26:07 crc kubenswrapper[4700]: I0227 17:26:07.002060 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4392a4d6-4e34-4821-8350-a82f8c06fc8d" path="/var/lib/kubelet/pods/4392a4d6-4e34-4821-8350-a82f8c06fc8d/volumes" Feb 27 17:26:08 crc kubenswrapper[4700]: I0227 17:26:08.536009 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cfbbd847f-bd2kg" Feb 27 17:26:08 crc kubenswrapper[4700]: I0227 17:26:08.655847 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77b594dc5c-ndvwx"] Feb 27 17:26:08 crc kubenswrapper[4700]: I0227 17:26:08.656126 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77b594dc5c-ndvwx" podUID="0c2ee20a-f966-4207-b52c-7b387989ddf5" containerName="dnsmasq-dns" containerID="cri-o://aa083400b53a68fd0d45d930cbb2bf2f683ca648fe2d8f1536a51b0f4f774c56" gracePeriod=10 Feb 27 17:26:09 crc kubenswrapper[4700]: I0227 17:26:09.190523 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77b594dc5c-ndvwx" Feb 27 17:26:09 crc kubenswrapper[4700]: I0227 17:26:09.232238 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c2ee20a-f966-4207-b52c-7b387989ddf5-ovsdbserver-nb\") pod \"0c2ee20a-f966-4207-b52c-7b387989ddf5\" (UID: \"0c2ee20a-f966-4207-b52c-7b387989ddf5\") " Feb 27 17:26:09 crc kubenswrapper[4700]: I0227 17:26:09.232285 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c2ee20a-f966-4207-b52c-7b387989ddf5-ovsdbserver-sb\") pod \"0c2ee20a-f966-4207-b52c-7b387989ddf5\" (UID: \"0c2ee20a-f966-4207-b52c-7b387989ddf5\") " Feb 27 17:26:09 crc kubenswrapper[4700]: I0227 17:26:09.232356 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c2ee20a-f966-4207-b52c-7b387989ddf5-dns-swift-storage-0\") pod \"0c2ee20a-f966-4207-b52c-7b387989ddf5\" (UID: \"0c2ee20a-f966-4207-b52c-7b387989ddf5\") " Feb 27 17:26:09 crc kubenswrapper[4700]: I0227 17:26:09.232488 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c2ee20a-f966-4207-b52c-7b387989ddf5-dns-svc\") pod \"0c2ee20a-f966-4207-b52c-7b387989ddf5\" (UID: \"0c2ee20a-f966-4207-b52c-7b387989ddf5\") " Feb 27 17:26:09 crc kubenswrapper[4700]: I0227 17:26:09.232549 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0c2ee20a-f966-4207-b52c-7b387989ddf5-openstack-edpm-ipam\") pod \"0c2ee20a-f966-4207-b52c-7b387989ddf5\" (UID: \"0c2ee20a-f966-4207-b52c-7b387989ddf5\") " Feb 27 17:26:09 crc kubenswrapper[4700]: I0227 17:26:09.232607 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8jdl\" (UniqueName: \"kubernetes.io/projected/0c2ee20a-f966-4207-b52c-7b387989ddf5-kube-api-access-v8jdl\") pod \"0c2ee20a-f966-4207-b52c-7b387989ddf5\" (UID: \"0c2ee20a-f966-4207-b52c-7b387989ddf5\") " Feb 27 17:26:09 crc kubenswrapper[4700]: I0227 17:26:09.232694 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c2ee20a-f966-4207-b52c-7b387989ddf5-config\") pod \"0c2ee20a-f966-4207-b52c-7b387989ddf5\" (UID: \"0c2ee20a-f966-4207-b52c-7b387989ddf5\") " Feb 27 17:26:09 crc kubenswrapper[4700]: I0227 17:26:09.240975 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c2ee20a-f966-4207-b52c-7b387989ddf5-kube-api-access-v8jdl" (OuterVolumeSpecName: "kube-api-access-v8jdl") pod "0c2ee20a-f966-4207-b52c-7b387989ddf5" (UID: "0c2ee20a-f966-4207-b52c-7b387989ddf5"). InnerVolumeSpecName "kube-api-access-v8jdl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:26:09 crc kubenswrapper[4700]: I0227 17:26:09.289411 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c2ee20a-f966-4207-b52c-7b387989ddf5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0c2ee20a-f966-4207-b52c-7b387989ddf5" (UID: "0c2ee20a-f966-4207-b52c-7b387989ddf5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:26:09 crc kubenswrapper[4700]: I0227 17:26:09.297754 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c2ee20a-f966-4207-b52c-7b387989ddf5-config" (OuterVolumeSpecName: "config") pod "0c2ee20a-f966-4207-b52c-7b387989ddf5" (UID: "0c2ee20a-f966-4207-b52c-7b387989ddf5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:26:09 crc kubenswrapper[4700]: I0227 17:26:09.299204 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c2ee20a-f966-4207-b52c-7b387989ddf5-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "0c2ee20a-f966-4207-b52c-7b387989ddf5" (UID: "0c2ee20a-f966-4207-b52c-7b387989ddf5"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:26:09 crc kubenswrapper[4700]: I0227 17:26:09.303727 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c2ee20a-f966-4207-b52c-7b387989ddf5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0c2ee20a-f966-4207-b52c-7b387989ddf5" (UID: "0c2ee20a-f966-4207-b52c-7b387989ddf5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:26:09 crc kubenswrapper[4700]: I0227 17:26:09.308040 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c2ee20a-f966-4207-b52c-7b387989ddf5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0c2ee20a-f966-4207-b52c-7b387989ddf5" (UID: "0c2ee20a-f966-4207-b52c-7b387989ddf5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:26:09 crc kubenswrapper[4700]: I0227 17:26:09.315869 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c2ee20a-f966-4207-b52c-7b387989ddf5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0c2ee20a-f966-4207-b52c-7b387989ddf5" (UID: "0c2ee20a-f966-4207-b52c-7b387989ddf5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:26:09 crc kubenswrapper[4700]: I0227 17:26:09.334840 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8jdl\" (UniqueName: \"kubernetes.io/projected/0c2ee20a-f966-4207-b52c-7b387989ddf5-kube-api-access-v8jdl\") on node \"crc\" DevicePath \"\"" Feb 27 17:26:09 crc kubenswrapper[4700]: I0227 17:26:09.334870 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c2ee20a-f966-4207-b52c-7b387989ddf5-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:26:09 crc kubenswrapper[4700]: I0227 17:26:09.334884 4700 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c2ee20a-f966-4207-b52c-7b387989ddf5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 27 17:26:09 crc kubenswrapper[4700]: I0227 17:26:09.334892 4700 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c2ee20a-f966-4207-b52c-7b387989ddf5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 17:26:09 crc kubenswrapper[4700]: I0227 17:26:09.334902 4700 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c2ee20a-f966-4207-b52c-7b387989ddf5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 27 17:26:09 crc kubenswrapper[4700]: I0227 17:26:09.334910 4700 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c2ee20a-f966-4207-b52c-7b387989ddf5-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 17:26:09 crc kubenswrapper[4700]: I0227 17:26:09.334917 4700 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0c2ee20a-f966-4207-b52c-7b387989ddf5-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 17:26:09 crc kubenswrapper[4700]: I0227 17:26:09.592435 4700 generic.go:334] "Generic (PLEG): container finished" podID="0c2ee20a-f966-4207-b52c-7b387989ddf5" containerID="aa083400b53a68fd0d45d930cbb2bf2f683ca648fe2d8f1536a51b0f4f774c56" exitCode=0 Feb 27 17:26:09 crc kubenswrapper[4700]: I0227 17:26:09.592491 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77b594dc5c-ndvwx" event={"ID":"0c2ee20a-f966-4207-b52c-7b387989ddf5","Type":"ContainerDied","Data":"aa083400b53a68fd0d45d930cbb2bf2f683ca648fe2d8f1536a51b0f4f774c56"} Feb 27 17:26:09 crc kubenswrapper[4700]: I0227 17:26:09.592515 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77b594dc5c-ndvwx" event={"ID":"0c2ee20a-f966-4207-b52c-7b387989ddf5","Type":"ContainerDied","Data":"9f23e9e947e32a56850bd1330754915faec8fd997dbe2201782e2b5ad516e579"} Feb 27 17:26:09 crc kubenswrapper[4700]: I0227 17:26:09.592533 4700 scope.go:117] "RemoveContainer" containerID="aa083400b53a68fd0d45d930cbb2bf2f683ca648fe2d8f1536a51b0f4f774c56" Feb 27 17:26:09 crc kubenswrapper[4700]: I0227 17:26:09.592520 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77b594dc5c-ndvwx" Feb 27 17:26:09 crc kubenswrapper[4700]: I0227 17:26:09.619914 4700 scope.go:117] "RemoveContainer" containerID="a0caff94988c0f1b5123903aeafeeecd44188499ef02bacd206c5512b253325e" Feb 27 17:26:09 crc kubenswrapper[4700]: I0227 17:26:09.651595 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77b594dc5c-ndvwx"] Feb 27 17:26:09 crc kubenswrapper[4700]: I0227 17:26:09.660007 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77b594dc5c-ndvwx"] Feb 27 17:26:09 crc kubenswrapper[4700]: I0227 17:26:09.669617 4700 scope.go:117] "RemoveContainer" containerID="aa083400b53a68fd0d45d930cbb2bf2f683ca648fe2d8f1536a51b0f4f774c56" Feb 27 17:26:09 crc kubenswrapper[4700]: E0227 17:26:09.671433 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa083400b53a68fd0d45d930cbb2bf2f683ca648fe2d8f1536a51b0f4f774c56\": container with ID starting with aa083400b53a68fd0d45d930cbb2bf2f683ca648fe2d8f1536a51b0f4f774c56 not found: ID does not exist" containerID="aa083400b53a68fd0d45d930cbb2bf2f683ca648fe2d8f1536a51b0f4f774c56" Feb 27 17:26:09 crc kubenswrapper[4700]: I0227 17:26:09.671570 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa083400b53a68fd0d45d930cbb2bf2f683ca648fe2d8f1536a51b0f4f774c56"} err="failed to get container status \"aa083400b53a68fd0d45d930cbb2bf2f683ca648fe2d8f1536a51b0f4f774c56\": rpc error: code = NotFound desc = could not find container \"aa083400b53a68fd0d45d930cbb2bf2f683ca648fe2d8f1536a51b0f4f774c56\": container with ID starting with aa083400b53a68fd0d45d930cbb2bf2f683ca648fe2d8f1536a51b0f4f774c56 not found: ID does not exist" Feb 27 17:26:09 crc kubenswrapper[4700]: I0227 17:26:09.671619 4700 scope.go:117] "RemoveContainer" containerID="a0caff94988c0f1b5123903aeafeeecd44188499ef02bacd206c5512b253325e" Feb 27 17:26:09 crc kubenswrapper[4700]: E0227 17:26:09.672091 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0caff94988c0f1b5123903aeafeeecd44188499ef02bacd206c5512b253325e\": container with ID starting with a0caff94988c0f1b5123903aeafeeecd44188499ef02bacd206c5512b253325e not found: ID does not exist" containerID="a0caff94988c0f1b5123903aeafeeecd44188499ef02bacd206c5512b253325e" Feb 27 17:26:09 crc kubenswrapper[4700]: I0227 17:26:09.672129 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0caff94988c0f1b5123903aeafeeecd44188499ef02bacd206c5512b253325e"} err="failed to get container status \"a0caff94988c0f1b5123903aeafeeecd44188499ef02bacd206c5512b253325e\": rpc error: code = NotFound desc = could not find container \"a0caff94988c0f1b5123903aeafeeecd44188499ef02bacd206c5512b253325e\": container with ID starting with a0caff94988c0f1b5123903aeafeeecd44188499ef02bacd206c5512b253325e not found: ID does not exist" Feb 27 17:26:10 crc kubenswrapper[4700]: I0227 17:26:10.998918 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c2ee20a-f966-4207-b52c-7b387989ddf5" path="/var/lib/kubelet/pods/0c2ee20a-f966-4207-b52c-7b387989ddf5/volumes" Feb 27 17:26:15 crc kubenswrapper[4700]: I0227 17:26:15.660419 4700 generic.go:334] "Generic (PLEG): container finished" podID="15c786bb-7353-4d06-b410-a44458a5f954" containerID="a2f26fc9d1979c04bd0885f2d6f1958a1bbbf31130e63c328f47e02c6adc1fe7" exitCode=0 Feb 27 17:26:15 crc kubenswrapper[4700]: I0227 17:26:15.660553 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"15c786bb-7353-4d06-b410-a44458a5f954","Type":"ContainerDied","Data":"a2f26fc9d1979c04bd0885f2d6f1958a1bbbf31130e63c328f47e02c6adc1fe7"} Feb 27 17:26:16 crc kubenswrapper[4700]: I0227 17:26:16.672176 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"15c786bb-7353-4d06-b410-a44458a5f954","Type":"ContainerStarted","Data":"bd076193a341ef3618796d8cab4a80f51ab08077cc5da4425411fb806e17a892"} Feb 27 17:26:16 crc kubenswrapper[4700]: I0227 17:26:16.672852 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Feb 27 17:26:16 crc kubenswrapper[4700]: I0227 17:26:16.706699 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.706678031 podStartE2EDuration="37.706678031s" podCreationTimestamp="2026-02-27 17:25:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:26:16.692134037 +0000 UTC m=+1536.677446784" watchObservedRunningTime="2026-02-27 17:26:16.706678031 +0000 UTC m=+1536.691990768" Feb 27 17:26:17 crc kubenswrapper[4700]: I0227 17:26:17.695633 4700 generic.go:334] "Generic (PLEG): container finished" podID="43a2852f-4d8f-468c-988a-60dcac328039" containerID="d5d66a235a89017e5a008ac40902ca259716009a8acc38ac2a27a24dee26973d" exitCode=0 Feb 27 17:26:17 crc kubenswrapper[4700]: I0227 17:26:17.695705 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"43a2852f-4d8f-468c-988a-60dcac328039","Type":"ContainerDied","Data":"d5d66a235a89017e5a008ac40902ca259716009a8acc38ac2a27a24dee26973d"} Feb 27 17:26:18 crc kubenswrapper[4700]: I0227 17:26:18.707173 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"43a2852f-4d8f-468c-988a-60dcac328039","Type":"ContainerStarted","Data":"091911e1496640c735e25563a957bddfc8d279b5b13923e641bfe3db16678c56"} Feb 27 17:26:18 crc kubenswrapper[4700]: I0227 17:26:18.707606 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:26:18 crc kubenswrapper[4700]: I0227 17:26:18.731430 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.731415746 podStartE2EDuration="38.731415746s" podCreationTimestamp="2026-02-27 17:25:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:26:18.730713857 +0000 UTC m=+1538.716026644" watchObservedRunningTime="2026-02-27 17:26:18.731415746 +0000 UTC m=+1538.716728493" Feb 27 17:26:21 crc kubenswrapper[4700]: I0227 17:26:21.061106 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rqjvd"] Feb 27 17:26:21 crc kubenswrapper[4700]: E0227 17:26:21.061574 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c2ee20a-f966-4207-b52c-7b387989ddf5" containerName="init" Feb 27 17:26:21 crc kubenswrapper[4700]: I0227 17:26:21.061586 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c2ee20a-f966-4207-b52c-7b387989ddf5" containerName="init" Feb 27 17:26:21 crc kubenswrapper[4700]: E0227 17:26:21.061602 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c2ee20a-f966-4207-b52c-7b387989ddf5" containerName="dnsmasq-dns" Feb 27 17:26:21 crc kubenswrapper[4700]: I0227 17:26:21.061608 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c2ee20a-f966-4207-b52c-7b387989ddf5" containerName="dnsmasq-dns" Feb 27 17:26:21 crc kubenswrapper[4700]: E0227 17:26:21.061618 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4b01082-c066-4fbd-aef5-169e9ec31fff" containerName="oc" Feb 27 17:26:21 crc kubenswrapper[4700]: I0227 17:26:21.061624 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4b01082-c066-4fbd-aef5-169e9ec31fff" containerName="oc" Feb 27 17:26:21 crc kubenswrapper[4700]: I0227 17:26:21.061842 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4b01082-c066-4fbd-aef5-169e9ec31fff" containerName="oc" Feb 27 17:26:21 crc kubenswrapper[4700]: I0227 17:26:21.061863 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c2ee20a-f966-4207-b52c-7b387989ddf5" containerName="dnsmasq-dns" Feb 27 17:26:21 crc kubenswrapper[4700]: I0227 17:26:21.062633 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rqjvd" Feb 27 17:26:21 crc kubenswrapper[4700]: I0227 17:26:21.064913 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dtdm4" Feb 27 17:26:21 crc kubenswrapper[4700]: I0227 17:26:21.065563 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 17:26:21 crc kubenswrapper[4700]: I0227 17:26:21.065929 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 17:26:21 crc kubenswrapper[4700]: I0227 17:26:21.067315 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 17:26:21 crc kubenswrapper[4700]: I0227 17:26:21.075229 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rqjvd"] Feb 27 17:26:21 crc kubenswrapper[4700]: I0227 17:26:21.104353 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85c1092c-f937-492c-9791-9df6150f439d-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-rqjvd\" (UID: \"85c1092c-f937-492c-9791-9df6150f439d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rqjvd" Feb 27 17:26:21 crc kubenswrapper[4700]: I0227 17:26:21.104783 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/85c1092c-f937-492c-9791-9df6150f439d-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-rqjvd\" (UID: \"85c1092c-f937-492c-9791-9df6150f439d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rqjvd" Feb 27 17:26:21 crc kubenswrapper[4700]: I0227 17:26:21.104823 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/85c1092c-f937-492c-9791-9df6150f439d-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-rqjvd\" (UID: \"85c1092c-f937-492c-9791-9df6150f439d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rqjvd" Feb 27 17:26:21 crc kubenswrapper[4700]: I0227 17:26:21.104922 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pprn\" (UniqueName: \"kubernetes.io/projected/85c1092c-f937-492c-9791-9df6150f439d-kube-api-access-8pprn\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-rqjvd\" (UID: \"85c1092c-f937-492c-9791-9df6150f439d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rqjvd" Feb 27 17:26:21 crc kubenswrapper[4700]: I0227 17:26:21.206472 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/85c1092c-f937-492c-9791-9df6150f439d-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-rqjvd\" (UID: \"85c1092c-f937-492c-9791-9df6150f439d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rqjvd" Feb 27 17:26:21 crc kubenswrapper[4700]: I0227 17:26:21.206755 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/85c1092c-f937-492c-9791-9df6150f439d-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-rqjvd\" (UID: \"85c1092c-f937-492c-9791-9df6150f439d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rqjvd" Feb 27 17:26:21 crc kubenswrapper[4700]: I0227 17:26:21.206893 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pprn\" (UniqueName: \"kubernetes.io/projected/85c1092c-f937-492c-9791-9df6150f439d-kube-api-access-8pprn\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-rqjvd\" (UID: \"85c1092c-f937-492c-9791-9df6150f439d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rqjvd" Feb 27 17:26:21 crc kubenswrapper[4700]: I0227 17:26:21.207001 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85c1092c-f937-492c-9791-9df6150f439d-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-rqjvd\" (UID: \"85c1092c-f937-492c-9791-9df6150f439d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rqjvd" Feb 27 17:26:21 crc kubenswrapper[4700]: I0227 17:26:21.213160 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/85c1092c-f937-492c-9791-9df6150f439d-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-rqjvd\" (UID: \"85c1092c-f937-492c-9791-9df6150f439d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rqjvd" Feb 27 17:26:21 crc kubenswrapper[4700]: I0227 17:26:21.214005 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/85c1092c-f937-492c-9791-9df6150f439d-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-rqjvd\" (UID: \"85c1092c-f937-492c-9791-9df6150f439d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rqjvd" Feb 27 17:26:21 crc kubenswrapper[4700]: I0227 17:26:21.217240 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85c1092c-f937-492c-9791-9df6150f439d-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-rqjvd\" (UID: \"85c1092c-f937-492c-9791-9df6150f439d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rqjvd" Feb 27 17:26:21 crc kubenswrapper[4700]: I0227 17:26:21.261163 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pprn\" (UniqueName: \"kubernetes.io/projected/85c1092c-f937-492c-9791-9df6150f439d-kube-api-access-8pprn\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-rqjvd\" (UID: \"85c1092c-f937-492c-9791-9df6150f439d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rqjvd" Feb 27 17:26:21 crc kubenswrapper[4700]: I0227 17:26:21.380831 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rqjvd" Feb 27 17:26:22 crc kubenswrapper[4700]: I0227 17:26:22.063704 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rqjvd"] Feb 27 17:26:22 crc kubenswrapper[4700]: W0227 17:26:22.073608 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod85c1092c_f937_492c_9791_9df6150f439d.slice/crio-d1baea438ae164c7f25e310e6e841f42c0ebc9381854bd1ccda09d2db8f4e00d WatchSource:0}: Error finding container d1baea438ae164c7f25e310e6e841f42c0ebc9381854bd1ccda09d2db8f4e00d: Status 404 returned error can't find the container with id d1baea438ae164c7f25e310e6e841f42c0ebc9381854bd1ccda09d2db8f4e00d Feb 27 17:26:22 crc kubenswrapper[4700]: I0227 17:26:22.743863 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rqjvd" event={"ID":"85c1092c-f937-492c-9791-9df6150f439d","Type":"ContainerStarted","Data":"d1baea438ae164c7f25e310e6e841f42c0ebc9381854bd1ccda09d2db8f4e00d"} Feb 27 17:26:29 crc kubenswrapper[4700]: I0227 17:26:29.539712 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Feb 27 17:26:30 crc kubenswrapper[4700]: I0227 17:26:30.605664 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Feb 27 17:26:39 crc kubenswrapper[4700]: I0227 17:26:39.962640 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rqjvd" event={"ID":"85c1092c-f937-492c-9791-9df6150f439d","Type":"ContainerStarted","Data":"79feffef65bed01ea698b1e3fca36960d03d16ac512ee30145cf17be55b08278"} Feb 27 17:26:39 crc kubenswrapper[4700]: I0227 17:26:39.995408 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rqjvd" podStartSLOduration=2.172661273 podStartE2EDuration="18.995389506s" podCreationTimestamp="2026-02-27 17:26:21 +0000 UTC" firstStartedPulling="2026-02-27 17:26:22.07672743 +0000 UTC m=+1542.062040177" lastFinishedPulling="2026-02-27 17:26:38.899455643 +0000 UTC m=+1558.884768410" observedRunningTime="2026-02-27 17:26:39.98225638 +0000 UTC m=+1559.967569148" watchObservedRunningTime="2026-02-27 17:26:39.995389506 +0000 UTC m=+1559.980702263" Feb 27 17:26:51 crc kubenswrapper[4700]: I0227 17:26:51.113707 4700 generic.go:334] "Generic (PLEG): container finished" podID="85c1092c-f937-492c-9791-9df6150f439d" containerID="79feffef65bed01ea698b1e3fca36960d03d16ac512ee30145cf17be55b08278" exitCode=0 Feb 27 17:26:51 crc kubenswrapper[4700]: I0227 17:26:51.113764 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rqjvd" event={"ID":"85c1092c-f937-492c-9791-9df6150f439d","Type":"ContainerDied","Data":"79feffef65bed01ea698b1e3fca36960d03d16ac512ee30145cf17be55b08278"} Feb 27 17:26:52 crc kubenswrapper[4700]: I0227 17:26:52.656487 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rqjvd" Feb 27 17:26:52 crc kubenswrapper[4700]: I0227 17:26:52.730608 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/85c1092c-f937-492c-9791-9df6150f439d-ssh-key-openstack-edpm-ipam\") pod \"85c1092c-f937-492c-9791-9df6150f439d\" (UID: \"85c1092c-f937-492c-9791-9df6150f439d\") " Feb 27 17:26:52 crc kubenswrapper[4700]: I0227 17:26:52.730729 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pprn\" (UniqueName: \"kubernetes.io/projected/85c1092c-f937-492c-9791-9df6150f439d-kube-api-access-8pprn\") pod \"85c1092c-f937-492c-9791-9df6150f439d\" (UID: \"85c1092c-f937-492c-9791-9df6150f439d\") " Feb 27 17:26:52 crc kubenswrapper[4700]: I0227 17:26:52.731603 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/85c1092c-f937-492c-9791-9df6150f439d-inventory\") pod \"85c1092c-f937-492c-9791-9df6150f439d\" (UID: \"85c1092c-f937-492c-9791-9df6150f439d\") " Feb 27 17:26:52 crc kubenswrapper[4700]: I0227 17:26:52.731659 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85c1092c-f937-492c-9791-9df6150f439d-repo-setup-combined-ca-bundle\") pod \"85c1092c-f937-492c-9791-9df6150f439d\" (UID: \"85c1092c-f937-492c-9791-9df6150f439d\") " Feb 27 17:26:52 crc kubenswrapper[4700]: I0227 17:26:52.747748 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85c1092c-f937-492c-9791-9df6150f439d-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "85c1092c-f937-492c-9791-9df6150f439d" (UID: "85c1092c-f937-492c-9791-9df6150f439d"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:26:52 crc kubenswrapper[4700]: I0227 17:26:52.747973 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85c1092c-f937-492c-9791-9df6150f439d-kube-api-access-8pprn" (OuterVolumeSpecName: "kube-api-access-8pprn") pod "85c1092c-f937-492c-9791-9df6150f439d" (UID: "85c1092c-f937-492c-9791-9df6150f439d"). InnerVolumeSpecName "kube-api-access-8pprn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:26:52 crc kubenswrapper[4700]: I0227 17:26:52.769404 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85c1092c-f937-492c-9791-9df6150f439d-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "85c1092c-f937-492c-9791-9df6150f439d" (UID: "85c1092c-f937-492c-9791-9df6150f439d"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:26:52 crc kubenswrapper[4700]: I0227 17:26:52.797785 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85c1092c-f937-492c-9791-9df6150f439d-inventory" (OuterVolumeSpecName: "inventory") pod "85c1092c-f937-492c-9791-9df6150f439d" (UID: "85c1092c-f937-492c-9791-9df6150f439d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:26:52 crc kubenswrapper[4700]: I0227 17:26:52.835496 4700 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/85c1092c-f937-492c-9791-9df6150f439d-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 17:26:52 crc kubenswrapper[4700]: I0227 17:26:52.835792 4700 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85c1092c-f937-492c-9791-9df6150f439d-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:26:52 crc kubenswrapper[4700]: I0227 17:26:52.835810 4700 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/85c1092c-f937-492c-9791-9df6150f439d-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 17:26:52 crc kubenswrapper[4700]: I0227 17:26:52.835834 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8pprn\" (UniqueName: \"kubernetes.io/projected/85c1092c-f937-492c-9791-9df6150f439d-kube-api-access-8pprn\") on node \"crc\" DevicePath \"\"" Feb 27 17:26:53 crc kubenswrapper[4700]: I0227 17:26:53.140687 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rqjvd" event={"ID":"85c1092c-f937-492c-9791-9df6150f439d","Type":"ContainerDied","Data":"d1baea438ae164c7f25e310e6e841f42c0ebc9381854bd1ccda09d2db8f4e00d"} Feb 27 17:26:53 crc kubenswrapper[4700]: I0227 17:26:53.140770 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d1baea438ae164c7f25e310e6e841f42c0ebc9381854bd1ccda09d2db8f4e00d" Feb 27 17:26:53 crc kubenswrapper[4700]: I0227 17:26:53.140848 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rqjvd" Feb 27 17:26:53 crc kubenswrapper[4700]: I0227 17:26:53.305968 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-8b64m"] Feb 27 17:26:53 crc kubenswrapper[4700]: E0227 17:26:53.306421 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85c1092c-f937-492c-9791-9df6150f439d" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 27 17:26:53 crc kubenswrapper[4700]: I0227 17:26:53.306451 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="85c1092c-f937-492c-9791-9df6150f439d" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 27 17:26:53 crc kubenswrapper[4700]: I0227 17:26:53.307194 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="85c1092c-f937-492c-9791-9df6150f439d" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 27 17:26:53 crc kubenswrapper[4700]: I0227 17:26:53.308005 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8b64m" Feb 27 17:26:53 crc kubenswrapper[4700]: I0227 17:26:53.310888 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dtdm4" Feb 27 17:26:53 crc kubenswrapper[4700]: I0227 17:26:53.311084 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 17:26:53 crc kubenswrapper[4700]: I0227 17:26:53.311126 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 17:26:53 crc kubenswrapper[4700]: I0227 17:26:53.313384 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 17:26:53 crc kubenswrapper[4700]: I0227 17:26:53.324236 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-8b64m"] Feb 27 17:26:53 crc kubenswrapper[4700]: I0227 17:26:53.346070 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2vnx\" (UniqueName: \"kubernetes.io/projected/ee61dca5-6400-424d-b76c-53ef8e3b0434-kube-api-access-g2vnx\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-8b64m\" (UID: \"ee61dca5-6400-424d-b76c-53ef8e3b0434\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8b64m" Feb 27 17:26:53 crc kubenswrapper[4700]: I0227 17:26:53.346151 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ee61dca5-6400-424d-b76c-53ef8e3b0434-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-8b64m\" (UID: \"ee61dca5-6400-424d-b76c-53ef8e3b0434\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8b64m" Feb 27 17:26:53 crc kubenswrapper[4700]: I0227 17:26:53.346191 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee61dca5-6400-424d-b76c-53ef8e3b0434-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-8b64m\" (UID: \"ee61dca5-6400-424d-b76c-53ef8e3b0434\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8b64m" Feb 27 17:26:53 crc kubenswrapper[4700]: I0227 17:26:53.448267 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ee61dca5-6400-424d-b76c-53ef8e3b0434-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-8b64m\" (UID: \"ee61dca5-6400-424d-b76c-53ef8e3b0434\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8b64m" Feb 27 17:26:53 crc kubenswrapper[4700]: I0227 17:26:53.448357 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee61dca5-6400-424d-b76c-53ef8e3b0434-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-8b64m\" (UID: \"ee61dca5-6400-424d-b76c-53ef8e3b0434\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8b64m" Feb 27 17:26:53 crc kubenswrapper[4700]: I0227 17:26:53.448604 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2vnx\" (UniqueName: \"kubernetes.io/projected/ee61dca5-6400-424d-b76c-53ef8e3b0434-kube-api-access-g2vnx\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-8b64m\" (UID: \"ee61dca5-6400-424d-b76c-53ef8e3b0434\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8b64m" Feb 27 17:26:53 crc kubenswrapper[4700]: I0227 17:26:53.454414 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee61dca5-6400-424d-b76c-53ef8e3b0434-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-8b64m\" (UID: \"ee61dca5-6400-424d-b76c-53ef8e3b0434\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8b64m" Feb 27 17:26:53 crc kubenswrapper[4700]: I0227 17:26:53.454979 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ee61dca5-6400-424d-b76c-53ef8e3b0434-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-8b64m\" (UID: \"ee61dca5-6400-424d-b76c-53ef8e3b0434\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8b64m" Feb 27 17:26:53 crc kubenswrapper[4700]: I0227 17:26:53.469575 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2vnx\" (UniqueName: \"kubernetes.io/projected/ee61dca5-6400-424d-b76c-53ef8e3b0434-kube-api-access-g2vnx\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-8b64m\" (UID: \"ee61dca5-6400-424d-b76c-53ef8e3b0434\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8b64m" Feb 27 17:26:53 crc kubenswrapper[4700]: I0227 17:26:53.670695 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8b64m" Feb 27 17:26:54 crc kubenswrapper[4700]: I0227 17:26:54.367758 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-8b64m"] Feb 27 17:26:54 crc kubenswrapper[4700]: W0227 17:26:54.369857 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee61dca5_6400_424d_b76c_53ef8e3b0434.slice/crio-3c188a9e7a23b960389d15fd77700b0fbc403e2d679a35f7a08851c887c9c4bc WatchSource:0}: Error finding container 3c188a9e7a23b960389d15fd77700b0fbc403e2d679a35f7a08851c887c9c4bc: Status 404 returned error can't find the container with id 3c188a9e7a23b960389d15fd77700b0fbc403e2d679a35f7a08851c887c9c4bc Feb 27 17:26:55 crc kubenswrapper[4700]: I0227 17:26:55.167857 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8b64m" event={"ID":"ee61dca5-6400-424d-b76c-53ef8e3b0434","Type":"ContainerStarted","Data":"3c188a9e7a23b960389d15fd77700b0fbc403e2d679a35f7a08851c887c9c4bc"} Feb 27 17:26:56 crc kubenswrapper[4700]: I0227 17:26:56.183092 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8b64m" event={"ID":"ee61dca5-6400-424d-b76c-53ef8e3b0434","Type":"ContainerStarted","Data":"cced039e809d487db256a299220f4fff26db42a867e75026f405c80970d319aa"} Feb 27 17:26:56 crc kubenswrapper[4700]: I0227 17:26:56.209296 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8b64m" podStartSLOduration=2.716359203 podStartE2EDuration="3.209271015s" podCreationTimestamp="2026-02-27 17:26:53 +0000 UTC" firstStartedPulling="2026-02-27 17:26:54.375392177 +0000 UTC m=+1574.360704934" lastFinishedPulling="2026-02-27 17:26:54.868303999 +0000 UTC m=+1574.853616746" observedRunningTime="2026-02-27 17:26:56.201897811 +0000 UTC m=+1576.187210598" watchObservedRunningTime="2026-02-27 17:26:56.209271015 +0000 UTC m=+1576.194583762" Feb 27 17:26:58 crc kubenswrapper[4700]: I0227 17:26:58.208699 4700 generic.go:334] "Generic (PLEG): container finished" podID="ee61dca5-6400-424d-b76c-53ef8e3b0434" containerID="cced039e809d487db256a299220f4fff26db42a867e75026f405c80970d319aa" exitCode=0 Feb 27 17:26:58 crc kubenswrapper[4700]: I0227 17:26:58.208745 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8b64m" event={"ID":"ee61dca5-6400-424d-b76c-53ef8e3b0434","Type":"ContainerDied","Data":"cced039e809d487db256a299220f4fff26db42a867e75026f405c80970d319aa"} Feb 27 17:26:59 crc kubenswrapper[4700]: I0227 17:26:59.812081 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8b64m" Feb 27 17:26:59 crc kubenswrapper[4700]: I0227 17:26:59.899406 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ee61dca5-6400-424d-b76c-53ef8e3b0434-ssh-key-openstack-edpm-ipam\") pod \"ee61dca5-6400-424d-b76c-53ef8e3b0434\" (UID: \"ee61dca5-6400-424d-b76c-53ef8e3b0434\") " Feb 27 17:26:59 crc kubenswrapper[4700]: I0227 17:26:59.900336 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee61dca5-6400-424d-b76c-53ef8e3b0434-inventory\") pod \"ee61dca5-6400-424d-b76c-53ef8e3b0434\" (UID: \"ee61dca5-6400-424d-b76c-53ef8e3b0434\") " Feb 27 17:26:59 crc kubenswrapper[4700]: I0227 17:26:59.900740 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g2vnx\" (UniqueName: \"kubernetes.io/projected/ee61dca5-6400-424d-b76c-53ef8e3b0434-kube-api-access-g2vnx\") pod \"ee61dca5-6400-424d-b76c-53ef8e3b0434\" (UID: \"ee61dca5-6400-424d-b76c-53ef8e3b0434\") " Feb 27 17:26:59 crc kubenswrapper[4700]: I0227 17:26:59.907439 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee61dca5-6400-424d-b76c-53ef8e3b0434-kube-api-access-g2vnx" (OuterVolumeSpecName: "kube-api-access-g2vnx") pod "ee61dca5-6400-424d-b76c-53ef8e3b0434" (UID: "ee61dca5-6400-424d-b76c-53ef8e3b0434"). InnerVolumeSpecName "kube-api-access-g2vnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:26:59 crc kubenswrapper[4700]: I0227 17:26:59.927564 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee61dca5-6400-424d-b76c-53ef8e3b0434-inventory" (OuterVolumeSpecName: "inventory") pod "ee61dca5-6400-424d-b76c-53ef8e3b0434" (UID: "ee61dca5-6400-424d-b76c-53ef8e3b0434"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:26:59 crc kubenswrapper[4700]: I0227 17:26:59.950883 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee61dca5-6400-424d-b76c-53ef8e3b0434-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "ee61dca5-6400-424d-b76c-53ef8e3b0434" (UID: "ee61dca5-6400-424d-b76c-53ef8e3b0434"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:27:00 crc kubenswrapper[4700]: I0227 17:27:00.004641 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g2vnx\" (UniqueName: \"kubernetes.io/projected/ee61dca5-6400-424d-b76c-53ef8e3b0434-kube-api-access-g2vnx\") on node \"crc\" DevicePath \"\"" Feb 27 17:27:00 crc kubenswrapper[4700]: I0227 17:27:00.005102 4700 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ee61dca5-6400-424d-b76c-53ef8e3b0434-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 17:27:00 crc kubenswrapper[4700]: I0227 17:27:00.005278 4700 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee61dca5-6400-424d-b76c-53ef8e3b0434-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 17:27:00 crc kubenswrapper[4700]: I0227 17:27:00.245369 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8b64m" event={"ID":"ee61dca5-6400-424d-b76c-53ef8e3b0434","Type":"ContainerDied","Data":"3c188a9e7a23b960389d15fd77700b0fbc403e2d679a35f7a08851c887c9c4bc"} Feb 27 17:27:00 crc kubenswrapper[4700]: I0227 17:27:00.245416 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c188a9e7a23b960389d15fd77700b0fbc403e2d679a35f7a08851c887c9c4bc" Feb 27 17:27:00 crc kubenswrapper[4700]: I0227 17:27:00.245576 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8b64m" Feb 27 17:27:00 crc kubenswrapper[4700]: I0227 17:27:00.333673 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5f27j"] Feb 27 17:27:00 crc kubenswrapper[4700]: E0227 17:27:00.334687 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee61dca5-6400-424d-b76c-53ef8e3b0434" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 27 17:27:00 crc kubenswrapper[4700]: I0227 17:27:00.334820 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee61dca5-6400-424d-b76c-53ef8e3b0434" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 27 17:27:00 crc kubenswrapper[4700]: I0227 17:27:00.335228 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee61dca5-6400-424d-b76c-53ef8e3b0434" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 27 17:27:00 crc kubenswrapper[4700]: I0227 17:27:00.336371 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5f27j" Feb 27 17:27:00 crc kubenswrapper[4700]: I0227 17:27:00.343148 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dtdm4" Feb 27 17:27:00 crc kubenswrapper[4700]: I0227 17:27:00.343259 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 17:27:00 crc kubenswrapper[4700]: I0227 17:27:00.343308 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 17:27:00 crc kubenswrapper[4700]: I0227 17:27:00.343687 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 17:27:00 crc kubenswrapper[4700]: I0227 17:27:00.369967 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5f27j"] Feb 27 17:27:00 crc kubenswrapper[4700]: I0227 17:27:00.412578 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/583579de-19a5-4728-97f9-37398ce4a452-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5f27j\" (UID: \"583579de-19a5-4728-97f9-37398ce4a452\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5f27j" Feb 27 17:27:00 crc kubenswrapper[4700]: I0227 17:27:00.412642 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/583579de-19a5-4728-97f9-37398ce4a452-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5f27j\" (UID: \"583579de-19a5-4728-97f9-37398ce4a452\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5f27j" Feb 27 17:27:00 crc kubenswrapper[4700]: I0227 17:27:00.413017 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcbfz\" (UniqueName: \"kubernetes.io/projected/583579de-19a5-4728-97f9-37398ce4a452-kube-api-access-hcbfz\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5f27j\" (UID: \"583579de-19a5-4728-97f9-37398ce4a452\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5f27j" Feb 27 17:27:00 crc kubenswrapper[4700]: I0227 17:27:00.413150 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/583579de-19a5-4728-97f9-37398ce4a452-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5f27j\" (UID: \"583579de-19a5-4728-97f9-37398ce4a452\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5f27j" Feb 27 17:27:00 crc kubenswrapper[4700]: I0227 17:27:00.516243 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcbfz\" (UniqueName: \"kubernetes.io/projected/583579de-19a5-4728-97f9-37398ce4a452-kube-api-access-hcbfz\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5f27j\" (UID: \"583579de-19a5-4728-97f9-37398ce4a452\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5f27j" Feb 27 17:27:00 crc kubenswrapper[4700]: I0227 17:27:00.516369 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/583579de-19a5-4728-97f9-37398ce4a452-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5f27j\" (UID: \"583579de-19a5-4728-97f9-37398ce4a452\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5f27j" Feb 27 17:27:00 crc kubenswrapper[4700]: I0227 17:27:00.516455 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/583579de-19a5-4728-97f9-37398ce4a452-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5f27j\" (UID: \"583579de-19a5-4728-97f9-37398ce4a452\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5f27j" Feb 27 17:27:00 crc kubenswrapper[4700]: I0227 17:27:00.516571 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/583579de-19a5-4728-97f9-37398ce4a452-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5f27j\" (UID: \"583579de-19a5-4728-97f9-37398ce4a452\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5f27j" Feb 27 17:27:00 crc kubenswrapper[4700]: I0227 17:27:00.520787 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/583579de-19a5-4728-97f9-37398ce4a452-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5f27j\" (UID: \"583579de-19a5-4728-97f9-37398ce4a452\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5f27j" Feb 27 17:27:00 crc kubenswrapper[4700]: I0227 17:27:00.522073 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/583579de-19a5-4728-97f9-37398ce4a452-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5f27j\" (UID: \"583579de-19a5-4728-97f9-37398ce4a452\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5f27j" Feb 27 17:27:00 crc kubenswrapper[4700]: I0227 17:27:00.522523 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/583579de-19a5-4728-97f9-37398ce4a452-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5f27j\" (UID: \"583579de-19a5-4728-97f9-37398ce4a452\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5f27j" Feb 27 17:27:00 crc kubenswrapper[4700]: I0227 17:27:00.536640 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcbfz\" (UniqueName: \"kubernetes.io/projected/583579de-19a5-4728-97f9-37398ce4a452-kube-api-access-hcbfz\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5f27j\" (UID: \"583579de-19a5-4728-97f9-37398ce4a452\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5f27j" Feb 27 17:27:00 crc kubenswrapper[4700]: I0227 17:27:00.669633 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5f27j" Feb 27 17:27:01 crc kubenswrapper[4700]: I0227 17:27:01.308201 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5f27j"] Feb 27 17:27:01 crc kubenswrapper[4700]: W0227 17:27:01.310211 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod583579de_19a5_4728_97f9_37398ce4a452.slice/crio-11d3ee8ab96a449b5ebc24d08687a12c7f709aebc2e425eb9fd2fac06f6c408c WatchSource:0}: Error finding container 11d3ee8ab96a449b5ebc24d08687a12c7f709aebc2e425eb9fd2fac06f6c408c: Status 404 returned error can't find the container with id 11d3ee8ab96a449b5ebc24d08687a12c7f709aebc2e425eb9fd2fac06f6c408c Feb 27 17:27:02 crc kubenswrapper[4700]: I0227 17:27:02.269613 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5f27j" event={"ID":"583579de-19a5-4728-97f9-37398ce4a452","Type":"ContainerStarted","Data":"11d3ee8ab96a449b5ebc24d08687a12c7f709aebc2e425eb9fd2fac06f6c408c"} Feb 27 17:27:03 crc kubenswrapper[4700]: I0227 17:27:03.283397 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5f27j" event={"ID":"583579de-19a5-4728-97f9-37398ce4a452","Type":"ContainerStarted","Data":"ca2a1a851cc940e621fc247ebcc7e834f95801e00cf0a176f3b905df8d71ab43"} Feb 27 17:27:03 crc kubenswrapper[4700]: I0227 17:27:03.323919 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5f27j" podStartSLOduration=2.475201839 podStartE2EDuration="3.32388923s" podCreationTimestamp="2026-02-27 17:27:00 +0000 UTC" firstStartedPulling="2026-02-27 17:27:01.315061385 +0000 UTC m=+1581.300374132" lastFinishedPulling="2026-02-27 17:27:02.163748746 +0000 UTC m=+1582.149061523" observedRunningTime="2026-02-27 17:27:03.308467814 +0000 UTC m=+1583.293780561" watchObservedRunningTime="2026-02-27 17:27:03.32388923 +0000 UTC m=+1583.309202007" Feb 27 17:27:05 crc kubenswrapper[4700]: I0227 17:27:05.829916 4700 scope.go:117] "RemoveContainer" containerID="812a90f84653ff6064934e96d31046dc02493a88565d50828f107fbab0eaf0c1" Feb 27 17:27:05 crc kubenswrapper[4700]: I0227 17:27:05.943357 4700 scope.go:117] "RemoveContainer" containerID="faca510343c0aafdfbd58f2b0053ee3d479c24c17c45411fbcb42ad20c39d844" Feb 27 17:27:05 crc kubenswrapper[4700]: I0227 17:27:05.974759 4700 scope.go:117] "RemoveContainer" containerID="546d803ac65c71659c694790207913d1286cb34b93d6768a646ed07444bf7fca" Feb 27 17:27:06 crc kubenswrapper[4700]: I0227 17:27:06.024893 4700 scope.go:117] "RemoveContainer" containerID="7b07b3db778ef64ad19e650e40a5aaf03eb204e105f667cd3736c791ca8a4dad" Feb 27 17:27:06 crc kubenswrapper[4700]: I0227 17:27:06.411208 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:27:06 crc kubenswrapper[4700]: I0227 17:27:06.411327 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:27:36 crc kubenswrapper[4700]: I0227 17:27:36.410774 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:27:36 crc kubenswrapper[4700]: I0227 17:27:36.412138 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:28:00 crc kubenswrapper[4700]: I0227 17:28:00.145838 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536888-rhz6k"] Feb 27 17:28:00 crc kubenswrapper[4700]: I0227 17:28:00.148313 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536888-rhz6k" Feb 27 17:28:00 crc kubenswrapper[4700]: I0227 17:28:00.151241 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:28:00 crc kubenswrapper[4700]: I0227 17:28:00.151537 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:28:00 crc kubenswrapper[4700]: I0227 17:28:00.153295 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d62kc" Feb 27 17:28:00 crc kubenswrapper[4700]: I0227 17:28:00.175631 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536888-rhz6k"] Feb 27 17:28:00 crc kubenswrapper[4700]: I0227 17:28:00.250411 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhv4g\" (UniqueName: \"kubernetes.io/projected/49f627a5-028c-4f94-9bb5-122be14fda92-kube-api-access-dhv4g\") pod \"auto-csr-approver-29536888-rhz6k\" (UID: \"49f627a5-028c-4f94-9bb5-122be14fda92\") " pod="openshift-infra/auto-csr-approver-29536888-rhz6k" Feb 27 17:28:00 crc kubenswrapper[4700]: I0227 17:28:00.351855 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhv4g\" (UniqueName: \"kubernetes.io/projected/49f627a5-028c-4f94-9bb5-122be14fda92-kube-api-access-dhv4g\") pod \"auto-csr-approver-29536888-rhz6k\" (UID: \"49f627a5-028c-4f94-9bb5-122be14fda92\") " pod="openshift-infra/auto-csr-approver-29536888-rhz6k" Feb 27 17:28:00 crc kubenswrapper[4700]: I0227 17:28:00.380422 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhv4g\" (UniqueName: \"kubernetes.io/projected/49f627a5-028c-4f94-9bb5-122be14fda92-kube-api-access-dhv4g\") pod \"auto-csr-approver-29536888-rhz6k\" (UID: \"49f627a5-028c-4f94-9bb5-122be14fda92\") " pod="openshift-infra/auto-csr-approver-29536888-rhz6k" Feb 27 17:28:00 crc kubenswrapper[4700]: I0227 17:28:00.469843 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536888-rhz6k" Feb 27 17:28:00 crc kubenswrapper[4700]: I0227 17:28:00.880374 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536888-rhz6k"] Feb 27 17:28:00 crc kubenswrapper[4700]: I0227 17:28:00.993757 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536888-rhz6k" event={"ID":"49f627a5-028c-4f94-9bb5-122be14fda92","Type":"ContainerStarted","Data":"f0b10ba93953fa067159b0088c5feda8075f57f19b691bbfa16c1a120b0eacc4"} Feb 27 17:28:03 crc kubenswrapper[4700]: I0227 17:28:03.014232 4700 generic.go:334] "Generic (PLEG): container finished" podID="49f627a5-028c-4f94-9bb5-122be14fda92" containerID="dbc2a4d9396c7ae6016cf7a23809f618c11c4d98324dbad5676b18d4b1c1ebb4" exitCode=0 Feb 27 17:28:03 crc kubenswrapper[4700]: I0227 17:28:03.014297 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536888-rhz6k" event={"ID":"49f627a5-028c-4f94-9bb5-122be14fda92","Type":"ContainerDied","Data":"dbc2a4d9396c7ae6016cf7a23809f618c11c4d98324dbad5676b18d4b1c1ebb4"} Feb 27 17:28:04 crc kubenswrapper[4700]: I0227 17:28:04.474080 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536888-rhz6k" Feb 27 17:28:04 crc kubenswrapper[4700]: I0227 17:28:04.542282 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhv4g\" (UniqueName: \"kubernetes.io/projected/49f627a5-028c-4f94-9bb5-122be14fda92-kube-api-access-dhv4g\") pod \"49f627a5-028c-4f94-9bb5-122be14fda92\" (UID: \"49f627a5-028c-4f94-9bb5-122be14fda92\") " Feb 27 17:28:04 crc kubenswrapper[4700]: I0227 17:28:04.548859 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49f627a5-028c-4f94-9bb5-122be14fda92-kube-api-access-dhv4g" (OuterVolumeSpecName: "kube-api-access-dhv4g") pod "49f627a5-028c-4f94-9bb5-122be14fda92" (UID: "49f627a5-028c-4f94-9bb5-122be14fda92"). InnerVolumeSpecName "kube-api-access-dhv4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:28:04 crc kubenswrapper[4700]: I0227 17:28:04.648733 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhv4g\" (UniqueName: \"kubernetes.io/projected/49f627a5-028c-4f94-9bb5-122be14fda92-kube-api-access-dhv4g\") on node \"crc\" DevicePath \"\"" Feb 27 17:28:05 crc kubenswrapper[4700]: I0227 17:28:05.040127 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536888-rhz6k" event={"ID":"49f627a5-028c-4f94-9bb5-122be14fda92","Type":"ContainerDied","Data":"f0b10ba93953fa067159b0088c5feda8075f57f19b691bbfa16c1a120b0eacc4"} Feb 27 17:28:05 crc kubenswrapper[4700]: I0227 17:28:05.040556 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0b10ba93953fa067159b0088c5feda8075f57f19b691bbfa16c1a120b0eacc4" Feb 27 17:28:05 crc kubenswrapper[4700]: I0227 17:28:05.040316 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536888-rhz6k" Feb 27 17:28:05 crc kubenswrapper[4700]: E0227 17:28:05.091728 4700 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod49f627a5_028c_4f94_9bb5_122be14fda92.slice/crio-f0b10ba93953fa067159b0088c5feda8075f57f19b691bbfa16c1a120b0eacc4\": RecentStats: unable to find data in memory cache]" Feb 27 17:28:05 crc kubenswrapper[4700]: I0227 17:28:05.545983 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536882-csprw"] Feb 27 17:28:05 crc kubenswrapper[4700]: I0227 17:28:05.553747 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536882-csprw"] Feb 27 17:28:06 crc kubenswrapper[4700]: I0227 17:28:06.165187 4700 scope.go:117] "RemoveContainer" containerID="9d2bf4686af7c8e42b169470fe39fac12f905746ec23838a4b6615396a50674a" Feb 27 17:28:06 crc kubenswrapper[4700]: I0227 17:28:06.410257 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:28:06 crc kubenswrapper[4700]: I0227 17:28:06.410345 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:28:06 crc kubenswrapper[4700]: I0227 17:28:06.410403 4700 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" Feb 27 17:28:06 crc kubenswrapper[4700]: I0227 17:28:06.411568 4700 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"88adee85741f363c1411884023846cd0aab91e2b1c4dca4a79262d1b15c4192f"} pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 17:28:06 crc kubenswrapper[4700]: I0227 17:28:06.411672 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" containerID="cri-o://88adee85741f363c1411884023846cd0aab91e2b1c4dca4a79262d1b15c4192f" gracePeriod=600 Feb 27 17:28:06 crc kubenswrapper[4700]: E0227 17:28:06.560573 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:28:06 crc kubenswrapper[4700]: I0227 17:28:06.993278 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58023608-7e65-4290-b24c-ea314a3fb15b" path="/var/lib/kubelet/pods/58023608-7e65-4290-b24c-ea314a3fb15b/volumes" Feb 27 17:28:07 crc kubenswrapper[4700]: I0227 17:28:07.067071 4700 generic.go:334] "Generic (PLEG): container finished" podID="84b45511-b94c-479f-98db-fd2c4eceec46" containerID="88adee85741f363c1411884023846cd0aab91e2b1c4dca4a79262d1b15c4192f" exitCode=0 Feb 27 17:28:07 crc kubenswrapper[4700]: I0227 17:28:07.067137 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerDied","Data":"88adee85741f363c1411884023846cd0aab91e2b1c4dca4a79262d1b15c4192f"} Feb 27 17:28:07 crc kubenswrapper[4700]: I0227 17:28:07.067200 4700 scope.go:117] "RemoveContainer" containerID="7f3dc0b799607b839ef8d740e7569600055eb53b8ad9a51f0d8e9d8e3a22c10d" Feb 27 17:28:07 crc kubenswrapper[4700]: I0227 17:28:07.068922 4700 scope.go:117] "RemoveContainer" containerID="88adee85741f363c1411884023846cd0aab91e2b1c4dca4a79262d1b15c4192f" Feb 27 17:28:07 crc kubenswrapper[4700]: E0227 17:28:07.069380 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:28:18 crc kubenswrapper[4700]: I0227 17:28:18.981700 4700 scope.go:117] "RemoveContainer" containerID="88adee85741f363c1411884023846cd0aab91e2b1c4dca4a79262d1b15c4192f" Feb 27 17:28:18 crc kubenswrapper[4700]: E0227 17:28:18.982837 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:28:32 crc kubenswrapper[4700]: I0227 17:28:32.981290 4700 scope.go:117] "RemoveContainer" containerID="88adee85741f363c1411884023846cd0aab91e2b1c4dca4a79262d1b15c4192f" Feb 27 17:28:32 crc kubenswrapper[4700]: E0227 17:28:32.982389 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:28:46 crc kubenswrapper[4700]: I0227 17:28:46.983347 4700 scope.go:117] "RemoveContainer" containerID="88adee85741f363c1411884023846cd0aab91e2b1c4dca4a79262d1b15c4192f" Feb 27 17:28:46 crc kubenswrapper[4700]: E0227 17:28:46.990594 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:29:01 crc kubenswrapper[4700]: I0227 17:29:01.982905 4700 scope.go:117] "RemoveContainer" containerID="88adee85741f363c1411884023846cd0aab91e2b1c4dca4a79262d1b15c4192f" Feb 27 17:29:01 crc kubenswrapper[4700]: E0227 17:29:01.984347 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:29:06 crc kubenswrapper[4700]: I0227 17:29:06.261912 4700 scope.go:117] "RemoveContainer" containerID="1b851f98ecd070cf7e19c017388d8ba7cc882c7f9ed67c744e2f4332334b168f" Feb 27 17:29:12 crc kubenswrapper[4700]: I0227 17:29:12.982315 4700 scope.go:117] "RemoveContainer" containerID="88adee85741f363c1411884023846cd0aab91e2b1c4dca4a79262d1b15c4192f" Feb 27 17:29:12 crc kubenswrapper[4700]: E0227 17:29:12.983623 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:29:26 crc kubenswrapper[4700]: I0227 17:29:26.982360 4700 scope.go:117] "RemoveContainer" containerID="88adee85741f363c1411884023846cd0aab91e2b1c4dca4a79262d1b15c4192f" Feb 27 17:29:26 crc kubenswrapper[4700]: E0227 17:29:26.983840 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:29:28 crc kubenswrapper[4700]: I0227 17:29:28.216571 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wz8zc"] Feb 27 17:29:28 crc kubenswrapper[4700]: E0227 17:29:28.217393 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49f627a5-028c-4f94-9bb5-122be14fda92" containerName="oc" Feb 27 17:29:28 crc kubenswrapper[4700]: I0227 17:29:28.217409 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="49f627a5-028c-4f94-9bb5-122be14fda92" containerName="oc" Feb 27 17:29:28 crc kubenswrapper[4700]: I0227 17:29:28.217670 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="49f627a5-028c-4f94-9bb5-122be14fda92" containerName="oc" Feb 27 17:29:28 crc kubenswrapper[4700]: I0227 17:29:28.219808 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wz8zc" Feb 27 17:29:28 crc kubenswrapper[4700]: I0227 17:29:28.227726 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wz8zc"] Feb 27 17:29:28 crc kubenswrapper[4700]: I0227 17:29:28.363902 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6blc\" (UniqueName: \"kubernetes.io/projected/83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f-kube-api-access-l6blc\") pod \"redhat-marketplace-wz8zc\" (UID: \"83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f\") " pod="openshift-marketplace/redhat-marketplace-wz8zc" Feb 27 17:29:28 crc kubenswrapper[4700]: I0227 17:29:28.364312 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f-catalog-content\") pod \"redhat-marketplace-wz8zc\" (UID: \"83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f\") " pod="openshift-marketplace/redhat-marketplace-wz8zc" Feb 27 17:29:28 crc kubenswrapper[4700]: I0227 17:29:28.364554 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f-utilities\") pod \"redhat-marketplace-wz8zc\" (UID: \"83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f\") " pod="openshift-marketplace/redhat-marketplace-wz8zc" Feb 27 17:29:28 crc kubenswrapper[4700]: I0227 17:29:28.466701 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f-utilities\") pod \"redhat-marketplace-wz8zc\" (UID: \"83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f\") " pod="openshift-marketplace/redhat-marketplace-wz8zc" Feb 27 17:29:28 crc kubenswrapper[4700]: I0227 17:29:28.466785 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6blc\" (UniqueName: \"kubernetes.io/projected/83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f-kube-api-access-l6blc\") pod \"redhat-marketplace-wz8zc\" (UID: \"83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f\") " pod="openshift-marketplace/redhat-marketplace-wz8zc" Feb 27 17:29:28 crc kubenswrapper[4700]: I0227 17:29:28.466960 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f-catalog-content\") pod \"redhat-marketplace-wz8zc\" (UID: \"83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f\") " pod="openshift-marketplace/redhat-marketplace-wz8zc" Feb 27 17:29:28 crc kubenswrapper[4700]: I0227 17:29:28.467144 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f-utilities\") pod \"redhat-marketplace-wz8zc\" (UID: \"83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f\") " pod="openshift-marketplace/redhat-marketplace-wz8zc" Feb 27 17:29:28 crc kubenswrapper[4700]: I0227 17:29:28.467378 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f-catalog-content\") pod \"redhat-marketplace-wz8zc\" (UID: \"83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f\") " pod="openshift-marketplace/redhat-marketplace-wz8zc" Feb 27 17:29:28 crc kubenswrapper[4700]: I0227 17:29:28.502831 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6blc\" (UniqueName: \"kubernetes.io/projected/83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f-kube-api-access-l6blc\") pod \"redhat-marketplace-wz8zc\" (UID: \"83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f\") " pod="openshift-marketplace/redhat-marketplace-wz8zc" Feb 27 17:29:28 crc kubenswrapper[4700]: I0227 17:29:28.545385 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wz8zc" Feb 27 17:29:29 crc kubenswrapper[4700]: W0227 17:29:29.064105 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod83e2d0ca_e33b_4a00_88c5_3fae8cd7aa1f.slice/crio-165e223dd1d124ccaadbb1661853f8bc9743fc940c7fac0aa8cc055d9fdb8dfa WatchSource:0}: Error finding container 165e223dd1d124ccaadbb1661853f8bc9743fc940c7fac0aa8cc055d9fdb8dfa: Status 404 returned error can't find the container with id 165e223dd1d124ccaadbb1661853f8bc9743fc940c7fac0aa8cc055d9fdb8dfa Feb 27 17:29:29 crc kubenswrapper[4700]: I0227 17:29:29.064990 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wz8zc"] Feb 27 17:29:29 crc kubenswrapper[4700]: I0227 17:29:29.255427 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wz8zc" event={"ID":"83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f","Type":"ContainerStarted","Data":"165e223dd1d124ccaadbb1661853f8bc9743fc940c7fac0aa8cc055d9fdb8dfa"} Feb 27 17:29:30 crc kubenswrapper[4700]: I0227 17:29:30.270452 4700 generic.go:334] "Generic (PLEG): container finished" podID="83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f" containerID="e09dc5111f1a7d859b7a7189c6025aa98a6229f9a8f19472678b556e227006ad" exitCode=0 Feb 27 17:29:30 crc kubenswrapper[4700]: I0227 17:29:30.270592 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wz8zc" event={"ID":"83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f","Type":"ContainerDied","Data":"e09dc5111f1a7d859b7a7189c6025aa98a6229f9a8f19472678b556e227006ad"} Feb 27 17:29:33 crc kubenswrapper[4700]: I0227 17:29:33.316175 4700 generic.go:334] "Generic (PLEG): container finished" podID="83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f" containerID="d66399d992c1fb84fca9d1e40eac504bf41c4c768a30253b82a58c32624f1d91" exitCode=0 Feb 27 17:29:33 crc kubenswrapper[4700]: I0227 17:29:33.316528 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wz8zc" event={"ID":"83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f","Type":"ContainerDied","Data":"d66399d992c1fb84fca9d1e40eac504bf41c4c768a30253b82a58c32624f1d91"} Feb 27 17:29:34 crc kubenswrapper[4700]: I0227 17:29:34.332671 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wz8zc" event={"ID":"83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f","Type":"ContainerStarted","Data":"030fde068cd4665e51879828c63f4eafeb7f4c49824dfb77bae07c0068531aa9"} Feb 27 17:29:34 crc kubenswrapper[4700]: I0227 17:29:34.369393 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wz8zc" podStartSLOduration=2.900829373 podStartE2EDuration="6.369368987s" podCreationTimestamp="2026-02-27 17:29:28 +0000 UTC" firstStartedPulling="2026-02-27 17:29:30.27439688 +0000 UTC m=+1730.259709657" lastFinishedPulling="2026-02-27 17:29:33.742936474 +0000 UTC m=+1733.728249271" observedRunningTime="2026-02-27 17:29:34.361401027 +0000 UTC m=+1734.346713814" watchObservedRunningTime="2026-02-27 17:29:34.369368987 +0000 UTC m=+1734.354681754" Feb 27 17:29:37 crc kubenswrapper[4700]: I0227 17:29:37.982119 4700 scope.go:117] "RemoveContainer" containerID="88adee85741f363c1411884023846cd0aab91e2b1c4dca4a79262d1b15c4192f" Feb 27 17:29:37 crc kubenswrapper[4700]: E0227 17:29:37.983050 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:29:38 crc kubenswrapper[4700]: I0227 17:29:38.546295 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wz8zc" Feb 27 17:29:38 crc kubenswrapper[4700]: I0227 17:29:38.546684 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wz8zc" Feb 27 17:29:38 crc kubenswrapper[4700]: I0227 17:29:38.639417 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wz8zc" Feb 27 17:29:39 crc kubenswrapper[4700]: I0227 17:29:39.473588 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wz8zc" Feb 27 17:29:39 crc kubenswrapper[4700]: I0227 17:29:39.547689 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wz8zc"] Feb 27 17:29:41 crc kubenswrapper[4700]: I0227 17:29:41.427545 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wz8zc" podUID="83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f" containerName="registry-server" containerID="cri-o://030fde068cd4665e51879828c63f4eafeb7f4c49824dfb77bae07c0068531aa9" gracePeriod=2 Feb 27 17:29:41 crc kubenswrapper[4700]: I0227 17:29:41.969643 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wz8zc" Feb 27 17:29:42 crc kubenswrapper[4700]: I0227 17:29:42.087873 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6blc\" (UniqueName: \"kubernetes.io/projected/83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f-kube-api-access-l6blc\") pod \"83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f\" (UID: \"83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f\") " Feb 27 17:29:42 crc kubenswrapper[4700]: I0227 17:29:42.088055 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f-utilities\") pod \"83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f\" (UID: \"83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f\") " Feb 27 17:29:42 crc kubenswrapper[4700]: I0227 17:29:42.088079 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f-catalog-content\") pod \"83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f\" (UID: \"83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f\") " Feb 27 17:29:42 crc kubenswrapper[4700]: I0227 17:29:42.089313 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f-utilities" (OuterVolumeSpecName: "utilities") pod "83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f" (UID: "83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:29:42 crc kubenswrapper[4700]: I0227 17:29:42.099909 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f-kube-api-access-l6blc" (OuterVolumeSpecName: "kube-api-access-l6blc") pod "83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f" (UID: "83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f"). InnerVolumeSpecName "kube-api-access-l6blc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:29:42 crc kubenswrapper[4700]: I0227 17:29:42.121634 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f" (UID: "83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:29:42 crc kubenswrapper[4700]: I0227 17:29:42.191026 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6blc\" (UniqueName: \"kubernetes.io/projected/83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f-kube-api-access-l6blc\") on node \"crc\" DevicePath \"\"" Feb 27 17:29:42 crc kubenswrapper[4700]: I0227 17:29:42.191082 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:29:42 crc kubenswrapper[4700]: I0227 17:29:42.191106 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:29:42 crc kubenswrapper[4700]: I0227 17:29:42.440743 4700 generic.go:334] "Generic (PLEG): container finished" podID="83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f" containerID="030fde068cd4665e51879828c63f4eafeb7f4c49824dfb77bae07c0068531aa9" exitCode=0 Feb 27 17:29:42 crc kubenswrapper[4700]: I0227 17:29:42.440786 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wz8zc" event={"ID":"83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f","Type":"ContainerDied","Data":"030fde068cd4665e51879828c63f4eafeb7f4c49824dfb77bae07c0068531aa9"} Feb 27 17:29:42 crc kubenswrapper[4700]: I0227 17:29:42.440812 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wz8zc" event={"ID":"83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f","Type":"ContainerDied","Data":"165e223dd1d124ccaadbb1661853f8bc9743fc940c7fac0aa8cc055d9fdb8dfa"} Feb 27 17:29:42 crc kubenswrapper[4700]: I0227 17:29:42.440831 4700 scope.go:117] "RemoveContainer" containerID="030fde068cd4665e51879828c63f4eafeb7f4c49824dfb77bae07c0068531aa9" Feb 27 17:29:42 crc kubenswrapper[4700]: I0227 17:29:42.440828 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wz8zc" Feb 27 17:29:42 crc kubenswrapper[4700]: I0227 17:29:42.485546 4700 scope.go:117] "RemoveContainer" containerID="d66399d992c1fb84fca9d1e40eac504bf41c4c768a30253b82a58c32624f1d91" Feb 27 17:29:42 crc kubenswrapper[4700]: I0227 17:29:42.513427 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wz8zc"] Feb 27 17:29:42 crc kubenswrapper[4700]: I0227 17:29:42.526698 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wz8zc"] Feb 27 17:29:42 crc kubenswrapper[4700]: I0227 17:29:42.534008 4700 scope.go:117] "RemoveContainer" containerID="e09dc5111f1a7d859b7a7189c6025aa98a6229f9a8f19472678b556e227006ad" Feb 27 17:29:42 crc kubenswrapper[4700]: I0227 17:29:42.595611 4700 scope.go:117] "RemoveContainer" containerID="030fde068cd4665e51879828c63f4eafeb7f4c49824dfb77bae07c0068531aa9" Feb 27 17:29:42 crc kubenswrapper[4700]: E0227 17:29:42.596161 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"030fde068cd4665e51879828c63f4eafeb7f4c49824dfb77bae07c0068531aa9\": container with ID starting with 030fde068cd4665e51879828c63f4eafeb7f4c49824dfb77bae07c0068531aa9 not found: ID does not exist" containerID="030fde068cd4665e51879828c63f4eafeb7f4c49824dfb77bae07c0068531aa9" Feb 27 17:29:42 crc kubenswrapper[4700]: I0227 17:29:42.596199 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"030fde068cd4665e51879828c63f4eafeb7f4c49824dfb77bae07c0068531aa9"} err="failed to get container status \"030fde068cd4665e51879828c63f4eafeb7f4c49824dfb77bae07c0068531aa9\": rpc error: code = NotFound desc = could not find container \"030fde068cd4665e51879828c63f4eafeb7f4c49824dfb77bae07c0068531aa9\": container with ID starting with 030fde068cd4665e51879828c63f4eafeb7f4c49824dfb77bae07c0068531aa9 not found: ID does not exist" Feb 27 17:29:42 crc kubenswrapper[4700]: I0227 17:29:42.596226 4700 scope.go:117] "RemoveContainer" containerID="d66399d992c1fb84fca9d1e40eac504bf41c4c768a30253b82a58c32624f1d91" Feb 27 17:29:42 crc kubenswrapper[4700]: E0227 17:29:42.596836 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d66399d992c1fb84fca9d1e40eac504bf41c4c768a30253b82a58c32624f1d91\": container with ID starting with d66399d992c1fb84fca9d1e40eac504bf41c4c768a30253b82a58c32624f1d91 not found: ID does not exist" containerID="d66399d992c1fb84fca9d1e40eac504bf41c4c768a30253b82a58c32624f1d91" Feb 27 17:29:42 crc kubenswrapper[4700]: I0227 17:29:42.596868 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d66399d992c1fb84fca9d1e40eac504bf41c4c768a30253b82a58c32624f1d91"} err="failed to get container status \"d66399d992c1fb84fca9d1e40eac504bf41c4c768a30253b82a58c32624f1d91\": rpc error: code = NotFound desc = could not find container \"d66399d992c1fb84fca9d1e40eac504bf41c4c768a30253b82a58c32624f1d91\": container with ID starting with d66399d992c1fb84fca9d1e40eac504bf41c4c768a30253b82a58c32624f1d91 not found: ID does not exist" Feb 27 17:29:42 crc kubenswrapper[4700]: I0227 17:29:42.596889 4700 scope.go:117] "RemoveContainer" containerID="e09dc5111f1a7d859b7a7189c6025aa98a6229f9a8f19472678b556e227006ad" Feb 27 17:29:42 crc kubenswrapper[4700]: E0227 17:29:42.597320 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e09dc5111f1a7d859b7a7189c6025aa98a6229f9a8f19472678b556e227006ad\": container with ID starting with e09dc5111f1a7d859b7a7189c6025aa98a6229f9a8f19472678b556e227006ad not found: ID does not exist" containerID="e09dc5111f1a7d859b7a7189c6025aa98a6229f9a8f19472678b556e227006ad" Feb 27 17:29:42 crc kubenswrapper[4700]: I0227 17:29:42.597405 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e09dc5111f1a7d859b7a7189c6025aa98a6229f9a8f19472678b556e227006ad"} err="failed to get container status \"e09dc5111f1a7d859b7a7189c6025aa98a6229f9a8f19472678b556e227006ad\": rpc error: code = NotFound desc = could not find container \"e09dc5111f1a7d859b7a7189c6025aa98a6229f9a8f19472678b556e227006ad\": container with ID starting with e09dc5111f1a7d859b7a7189c6025aa98a6229f9a8f19472678b556e227006ad not found: ID does not exist" Feb 27 17:29:43 crc kubenswrapper[4700]: I0227 17:29:43.001496 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f" path="/var/lib/kubelet/pods/83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f/volumes" Feb 27 17:29:49 crc kubenswrapper[4700]: I0227 17:29:49.981534 4700 scope.go:117] "RemoveContainer" containerID="88adee85741f363c1411884023846cd0aab91e2b1c4dca4a79262d1b15c4192f" Feb 27 17:29:49 crc kubenswrapper[4700]: E0227 17:29:49.982726 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:30:00 crc kubenswrapper[4700]: I0227 17:30:00.172739 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536890-p9c6t"] Feb 27 17:30:00 crc kubenswrapper[4700]: E0227 17:30:00.173992 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f" containerName="extract-content" Feb 27 17:30:00 crc kubenswrapper[4700]: I0227 17:30:00.174020 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f" containerName="extract-content" Feb 27 17:30:00 crc kubenswrapper[4700]: E0227 17:30:00.174097 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f" containerName="extract-utilities" Feb 27 17:30:00 crc kubenswrapper[4700]: I0227 17:30:00.174119 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f" containerName="extract-utilities" Feb 27 17:30:00 crc kubenswrapper[4700]: E0227 17:30:00.174145 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f" containerName="registry-server" Feb 27 17:30:00 crc kubenswrapper[4700]: I0227 17:30:00.174163 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f" containerName="registry-server" Feb 27 17:30:00 crc kubenswrapper[4700]: I0227 17:30:00.174783 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="83e2d0ca-e33b-4a00-88c5-3fae8cd7aa1f" containerName="registry-server" Feb 27 17:30:00 crc kubenswrapper[4700]: I0227 17:30:00.177258 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536890-p9c6t" Feb 27 17:30:00 crc kubenswrapper[4700]: I0227 17:30:00.181177 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:30:00 crc kubenswrapper[4700]: I0227 17:30:00.181249 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:30:00 crc kubenswrapper[4700]: I0227 17:30:00.184651 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d62kc" Feb 27 17:30:00 crc kubenswrapper[4700]: I0227 17:30:00.192669 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536890-7xlw5"] Feb 27 17:30:00 crc kubenswrapper[4700]: I0227 17:30:00.195864 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536890-7xlw5" Feb 27 17:30:00 crc kubenswrapper[4700]: I0227 17:30:00.198677 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 27 17:30:00 crc kubenswrapper[4700]: I0227 17:30:00.198869 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 27 17:30:00 crc kubenswrapper[4700]: I0227 17:30:00.207205 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536890-p9c6t"] Feb 27 17:30:00 crc kubenswrapper[4700]: I0227 17:30:00.218138 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536890-7xlw5"] Feb 27 17:30:00 crc kubenswrapper[4700]: I0227 17:30:00.316218 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zlkx\" (UniqueName: \"kubernetes.io/projected/1b6d9b14-d4a9-4633-bba4-197eac2766a4-kube-api-access-6zlkx\") pod \"auto-csr-approver-29536890-p9c6t\" (UID: \"1b6d9b14-d4a9-4633-bba4-197eac2766a4\") " pod="openshift-infra/auto-csr-approver-29536890-p9c6t" Feb 27 17:30:00 crc kubenswrapper[4700]: I0227 17:30:00.316568 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a7207687-fb96-4b11-8c3b-c9f9467e0373-secret-volume\") pod \"collect-profiles-29536890-7xlw5\" (UID: \"a7207687-fb96-4b11-8c3b-c9f9467e0373\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536890-7xlw5" Feb 27 17:30:00 crc kubenswrapper[4700]: I0227 17:30:00.316787 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrgcl\" (UniqueName: \"kubernetes.io/projected/a7207687-fb96-4b11-8c3b-c9f9467e0373-kube-api-access-zrgcl\") pod \"collect-profiles-29536890-7xlw5\" (UID: \"a7207687-fb96-4b11-8c3b-c9f9467e0373\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536890-7xlw5" Feb 27 17:30:00 crc kubenswrapper[4700]: I0227 17:30:00.316844 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a7207687-fb96-4b11-8c3b-c9f9467e0373-config-volume\") pod \"collect-profiles-29536890-7xlw5\" (UID: \"a7207687-fb96-4b11-8c3b-c9f9467e0373\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536890-7xlw5" Feb 27 17:30:00 crc kubenswrapper[4700]: I0227 17:30:00.418714 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a7207687-fb96-4b11-8c3b-c9f9467e0373-secret-volume\") pod \"collect-profiles-29536890-7xlw5\" (UID: \"a7207687-fb96-4b11-8c3b-c9f9467e0373\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536890-7xlw5" Feb 27 17:30:00 crc kubenswrapper[4700]: I0227 17:30:00.418842 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrgcl\" (UniqueName: \"kubernetes.io/projected/a7207687-fb96-4b11-8c3b-c9f9467e0373-kube-api-access-zrgcl\") pod \"collect-profiles-29536890-7xlw5\" (UID: \"a7207687-fb96-4b11-8c3b-c9f9467e0373\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536890-7xlw5" Feb 27 17:30:00 crc kubenswrapper[4700]: I0227 17:30:00.418885 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a7207687-fb96-4b11-8c3b-c9f9467e0373-config-volume\") pod \"collect-profiles-29536890-7xlw5\" (UID: \"a7207687-fb96-4b11-8c3b-c9f9467e0373\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536890-7xlw5" Feb 27 17:30:00 crc kubenswrapper[4700]: I0227 17:30:00.418937 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zlkx\" (UniqueName: \"kubernetes.io/projected/1b6d9b14-d4a9-4633-bba4-197eac2766a4-kube-api-access-6zlkx\") pod \"auto-csr-approver-29536890-p9c6t\" (UID: \"1b6d9b14-d4a9-4633-bba4-197eac2766a4\") " pod="openshift-infra/auto-csr-approver-29536890-p9c6t" Feb 27 17:30:00 crc kubenswrapper[4700]: I0227 17:30:00.419989 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a7207687-fb96-4b11-8c3b-c9f9467e0373-config-volume\") pod \"collect-profiles-29536890-7xlw5\" (UID: \"a7207687-fb96-4b11-8c3b-c9f9467e0373\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536890-7xlw5" Feb 27 17:30:00 crc kubenswrapper[4700]: I0227 17:30:00.434141 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a7207687-fb96-4b11-8c3b-c9f9467e0373-secret-volume\") pod \"collect-profiles-29536890-7xlw5\" (UID: \"a7207687-fb96-4b11-8c3b-c9f9467e0373\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536890-7xlw5" Feb 27 17:30:00 crc kubenswrapper[4700]: I0227 17:30:00.438676 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrgcl\" (UniqueName: \"kubernetes.io/projected/a7207687-fb96-4b11-8c3b-c9f9467e0373-kube-api-access-zrgcl\") pod \"collect-profiles-29536890-7xlw5\" (UID: \"a7207687-fb96-4b11-8c3b-c9f9467e0373\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536890-7xlw5" Feb 27 17:30:00 crc kubenswrapper[4700]: I0227 17:30:00.455989 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zlkx\" (UniqueName: \"kubernetes.io/projected/1b6d9b14-d4a9-4633-bba4-197eac2766a4-kube-api-access-6zlkx\") pod \"auto-csr-approver-29536890-p9c6t\" (UID: \"1b6d9b14-d4a9-4633-bba4-197eac2766a4\") " pod="openshift-infra/auto-csr-approver-29536890-p9c6t" Feb 27 17:30:00 crc kubenswrapper[4700]: I0227 17:30:00.499767 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536890-p9c6t" Feb 27 17:30:00 crc kubenswrapper[4700]: I0227 17:30:00.518416 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536890-7xlw5" Feb 27 17:30:01 crc kubenswrapper[4700]: I0227 17:30:01.009279 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536890-p9c6t"] Feb 27 17:30:01 crc kubenswrapper[4700]: W0227 17:30:01.049328 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda7207687_fb96_4b11_8c3b_c9f9467e0373.slice/crio-4cfdcb00f20ab57b37049f554d89310b4d82425020022de784c5403d5530b433 WatchSource:0}: Error finding container 4cfdcb00f20ab57b37049f554d89310b4d82425020022de784c5403d5530b433: Status 404 returned error can't find the container with id 4cfdcb00f20ab57b37049f554d89310b4d82425020022de784c5403d5530b433 Feb 27 17:30:01 crc kubenswrapper[4700]: I0227 17:30:01.052701 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536890-7xlw5"] Feb 27 17:30:01 crc kubenswrapper[4700]: I0227 17:30:01.728085 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536890-7xlw5" event={"ID":"a7207687-fb96-4b11-8c3b-c9f9467e0373","Type":"ContainerStarted","Data":"512b48dbe4a24e5db07e29406ee97abf45ae51a6445f58b4122ad3375c0a48a9"} Feb 27 17:30:01 crc kubenswrapper[4700]: I0227 17:30:01.729583 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536890-7xlw5" event={"ID":"a7207687-fb96-4b11-8c3b-c9f9467e0373","Type":"ContainerStarted","Data":"4cfdcb00f20ab57b37049f554d89310b4d82425020022de784c5403d5530b433"} Feb 27 17:30:01 crc kubenswrapper[4700]: I0227 17:30:01.732164 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536890-p9c6t" event={"ID":"1b6d9b14-d4a9-4633-bba4-197eac2766a4","Type":"ContainerStarted","Data":"eb4862ee0e96fdca59bbe8eaf9b8aa3a33d8af8a9de30fb143684f04366f5045"} Feb 27 17:30:01 crc kubenswrapper[4700]: I0227 17:30:01.753602 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29536890-7xlw5" podStartSLOduration=1.753577528 podStartE2EDuration="1.753577528s" podCreationTimestamp="2026-02-27 17:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:30:01.745981917 +0000 UTC m=+1761.731294664" watchObservedRunningTime="2026-02-27 17:30:01.753577528 +0000 UTC m=+1761.738890285" Feb 27 17:30:02 crc kubenswrapper[4700]: I0227 17:30:02.745842 4700 generic.go:334] "Generic (PLEG): container finished" podID="a7207687-fb96-4b11-8c3b-c9f9467e0373" containerID="512b48dbe4a24e5db07e29406ee97abf45ae51a6445f58b4122ad3375c0a48a9" exitCode=0 Feb 27 17:30:02 crc kubenswrapper[4700]: I0227 17:30:02.745920 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536890-7xlw5" event={"ID":"a7207687-fb96-4b11-8c3b-c9f9467e0373","Type":"ContainerDied","Data":"512b48dbe4a24e5db07e29406ee97abf45ae51a6445f58b4122ad3375c0a48a9"} Feb 27 17:30:02 crc kubenswrapper[4700]: I0227 17:30:02.981199 4700 scope.go:117] "RemoveContainer" containerID="88adee85741f363c1411884023846cd0aab91e2b1c4dca4a79262d1b15c4192f" Feb 27 17:30:02 crc kubenswrapper[4700]: E0227 17:30:02.981483 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:30:03 crc kubenswrapper[4700]: I0227 17:30:03.765203 4700 generic.go:334] "Generic (PLEG): container finished" podID="1b6d9b14-d4a9-4633-bba4-197eac2766a4" containerID="2bb6222ce1bf85664559dcaf0292653e231e4d9e5541158eb0296b51394c0ed4" exitCode=0 Feb 27 17:30:03 crc kubenswrapper[4700]: I0227 17:30:03.765365 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536890-p9c6t" event={"ID":"1b6d9b14-d4a9-4633-bba4-197eac2766a4","Type":"ContainerDied","Data":"2bb6222ce1bf85664559dcaf0292653e231e4d9e5541158eb0296b51394c0ed4"} Feb 27 17:30:04 crc kubenswrapper[4700]: I0227 17:30:04.162305 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536890-7xlw5" Feb 27 17:30:04 crc kubenswrapper[4700]: I0227 17:30:04.204765 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a7207687-fb96-4b11-8c3b-c9f9467e0373-config-volume\") pod \"a7207687-fb96-4b11-8c3b-c9f9467e0373\" (UID: \"a7207687-fb96-4b11-8c3b-c9f9467e0373\") " Feb 27 17:30:04 crc kubenswrapper[4700]: I0227 17:30:04.204843 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrgcl\" (UniqueName: \"kubernetes.io/projected/a7207687-fb96-4b11-8c3b-c9f9467e0373-kube-api-access-zrgcl\") pod \"a7207687-fb96-4b11-8c3b-c9f9467e0373\" (UID: \"a7207687-fb96-4b11-8c3b-c9f9467e0373\") " Feb 27 17:30:04 crc kubenswrapper[4700]: I0227 17:30:04.205167 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a7207687-fb96-4b11-8c3b-c9f9467e0373-secret-volume\") pod \"a7207687-fb96-4b11-8c3b-c9f9467e0373\" (UID: \"a7207687-fb96-4b11-8c3b-c9f9467e0373\") " Feb 27 17:30:04 crc kubenswrapper[4700]: I0227 17:30:04.205500 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7207687-fb96-4b11-8c3b-c9f9467e0373-config-volume" (OuterVolumeSpecName: "config-volume") pod "a7207687-fb96-4b11-8c3b-c9f9467e0373" (UID: "a7207687-fb96-4b11-8c3b-c9f9467e0373"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:30:04 crc kubenswrapper[4700]: I0227 17:30:04.205883 4700 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a7207687-fb96-4b11-8c3b-c9f9467e0373-config-volume\") on node \"crc\" DevicePath \"\"" Feb 27 17:30:04 crc kubenswrapper[4700]: I0227 17:30:04.214267 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7207687-fb96-4b11-8c3b-c9f9467e0373-kube-api-access-zrgcl" (OuterVolumeSpecName: "kube-api-access-zrgcl") pod "a7207687-fb96-4b11-8c3b-c9f9467e0373" (UID: "a7207687-fb96-4b11-8c3b-c9f9467e0373"). InnerVolumeSpecName "kube-api-access-zrgcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:30:04 crc kubenswrapper[4700]: I0227 17:30:04.214793 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7207687-fb96-4b11-8c3b-c9f9467e0373-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a7207687-fb96-4b11-8c3b-c9f9467e0373" (UID: "a7207687-fb96-4b11-8c3b-c9f9467e0373"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:30:04 crc kubenswrapper[4700]: I0227 17:30:04.307565 4700 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a7207687-fb96-4b11-8c3b-c9f9467e0373-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 27 17:30:04 crc kubenswrapper[4700]: I0227 17:30:04.307596 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrgcl\" (UniqueName: \"kubernetes.io/projected/a7207687-fb96-4b11-8c3b-c9f9467e0373-kube-api-access-zrgcl\") on node \"crc\" DevicePath \"\"" Feb 27 17:30:04 crc kubenswrapper[4700]: I0227 17:30:04.781896 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536890-7xlw5" event={"ID":"a7207687-fb96-4b11-8c3b-c9f9467e0373","Type":"ContainerDied","Data":"4cfdcb00f20ab57b37049f554d89310b4d82425020022de784c5403d5530b433"} Feb 27 17:30:04 crc kubenswrapper[4700]: I0227 17:30:04.781962 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4cfdcb00f20ab57b37049f554d89310b4d82425020022de784c5403d5530b433" Feb 27 17:30:04 crc kubenswrapper[4700]: I0227 17:30:04.782050 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536890-7xlw5" Feb 27 17:30:05 crc kubenswrapper[4700]: I0227 17:30:05.226142 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536890-p9c6t" Feb 27 17:30:05 crc kubenswrapper[4700]: I0227 17:30:05.337308 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6zlkx\" (UniqueName: \"kubernetes.io/projected/1b6d9b14-d4a9-4633-bba4-197eac2766a4-kube-api-access-6zlkx\") pod \"1b6d9b14-d4a9-4633-bba4-197eac2766a4\" (UID: \"1b6d9b14-d4a9-4633-bba4-197eac2766a4\") " Feb 27 17:30:05 crc kubenswrapper[4700]: I0227 17:30:05.350115 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b6d9b14-d4a9-4633-bba4-197eac2766a4-kube-api-access-6zlkx" (OuterVolumeSpecName: "kube-api-access-6zlkx") pod "1b6d9b14-d4a9-4633-bba4-197eac2766a4" (UID: "1b6d9b14-d4a9-4633-bba4-197eac2766a4"). InnerVolumeSpecName "kube-api-access-6zlkx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:30:05 crc kubenswrapper[4700]: I0227 17:30:05.438925 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6zlkx\" (UniqueName: \"kubernetes.io/projected/1b6d9b14-d4a9-4633-bba4-197eac2766a4-kube-api-access-6zlkx\") on node \"crc\" DevicePath \"\"" Feb 27 17:30:05 crc kubenswrapper[4700]: I0227 17:30:05.797400 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536890-p9c6t" event={"ID":"1b6d9b14-d4a9-4633-bba4-197eac2766a4","Type":"ContainerDied","Data":"eb4862ee0e96fdca59bbe8eaf9b8aa3a33d8af8a9de30fb143684f04366f5045"} Feb 27 17:30:05 crc kubenswrapper[4700]: I0227 17:30:05.797498 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536890-p9c6t" Feb 27 17:30:05 crc kubenswrapper[4700]: I0227 17:30:05.797510 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb4862ee0e96fdca59bbe8eaf9b8aa3a33d8af8a9de30fb143684f04366f5045" Feb 27 17:30:06 crc kubenswrapper[4700]: I0227 17:30:06.320272 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536884-25wnt"] Feb 27 17:30:06 crc kubenswrapper[4700]: I0227 17:30:06.339614 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536884-25wnt"] Feb 27 17:30:06 crc kubenswrapper[4700]: I0227 17:30:06.354797 4700 scope.go:117] "RemoveContainer" containerID="311929f6d9e48ad9e0020e036a90ed5783e7cbce50894a14a3efda6d9a10cc5b" Feb 27 17:30:07 crc kubenswrapper[4700]: I0227 17:30:07.000864 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="950c16a0-025a-44bb-8ebd-1b69dea4d6b8" path="/var/lib/kubelet/pods/950c16a0-025a-44bb-8ebd-1b69dea4d6b8/volumes" Feb 27 17:30:08 crc kubenswrapper[4700]: I0227 17:30:08.843363 4700 generic.go:334] "Generic (PLEG): container finished" podID="583579de-19a5-4728-97f9-37398ce4a452" containerID="ca2a1a851cc940e621fc247ebcc7e834f95801e00cf0a176f3b905df8d71ab43" exitCode=0 Feb 27 17:30:08 crc kubenswrapper[4700]: I0227 17:30:08.843641 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5f27j" event={"ID":"583579de-19a5-4728-97f9-37398ce4a452","Type":"ContainerDied","Data":"ca2a1a851cc940e621fc247ebcc7e834f95801e00cf0a176f3b905df8d71ab43"} Feb 27 17:30:10 crc kubenswrapper[4700]: I0227 17:30:10.362175 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5f27j" Feb 27 17:30:10 crc kubenswrapper[4700]: I0227 17:30:10.458982 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/583579de-19a5-4728-97f9-37398ce4a452-inventory\") pod \"583579de-19a5-4728-97f9-37398ce4a452\" (UID: \"583579de-19a5-4728-97f9-37398ce4a452\") " Feb 27 17:30:10 crc kubenswrapper[4700]: I0227 17:30:10.459070 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/583579de-19a5-4728-97f9-37398ce4a452-ssh-key-openstack-edpm-ipam\") pod \"583579de-19a5-4728-97f9-37398ce4a452\" (UID: \"583579de-19a5-4728-97f9-37398ce4a452\") " Feb 27 17:30:10 crc kubenswrapper[4700]: I0227 17:30:10.459107 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/583579de-19a5-4728-97f9-37398ce4a452-bootstrap-combined-ca-bundle\") pod \"583579de-19a5-4728-97f9-37398ce4a452\" (UID: \"583579de-19a5-4728-97f9-37398ce4a452\") " Feb 27 17:30:10 crc kubenswrapper[4700]: I0227 17:30:10.459159 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcbfz\" (UniqueName: \"kubernetes.io/projected/583579de-19a5-4728-97f9-37398ce4a452-kube-api-access-hcbfz\") pod \"583579de-19a5-4728-97f9-37398ce4a452\" (UID: \"583579de-19a5-4728-97f9-37398ce4a452\") " Feb 27 17:30:10 crc kubenswrapper[4700]: I0227 17:30:10.480777 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/583579de-19a5-4728-97f9-37398ce4a452-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "583579de-19a5-4728-97f9-37398ce4a452" (UID: "583579de-19a5-4728-97f9-37398ce4a452"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:30:10 crc kubenswrapper[4700]: I0227 17:30:10.482219 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/583579de-19a5-4728-97f9-37398ce4a452-kube-api-access-hcbfz" (OuterVolumeSpecName: "kube-api-access-hcbfz") pod "583579de-19a5-4728-97f9-37398ce4a452" (UID: "583579de-19a5-4728-97f9-37398ce4a452"). InnerVolumeSpecName "kube-api-access-hcbfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:30:10 crc kubenswrapper[4700]: I0227 17:30:10.496016 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/583579de-19a5-4728-97f9-37398ce4a452-inventory" (OuterVolumeSpecName: "inventory") pod "583579de-19a5-4728-97f9-37398ce4a452" (UID: "583579de-19a5-4728-97f9-37398ce4a452"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:30:10 crc kubenswrapper[4700]: I0227 17:30:10.497685 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/583579de-19a5-4728-97f9-37398ce4a452-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "583579de-19a5-4728-97f9-37398ce4a452" (UID: "583579de-19a5-4728-97f9-37398ce4a452"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:30:10 crc kubenswrapper[4700]: I0227 17:30:10.561549 4700 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/583579de-19a5-4728-97f9-37398ce4a452-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 17:30:10 crc kubenswrapper[4700]: I0227 17:30:10.561591 4700 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/583579de-19a5-4728-97f9-37398ce4a452-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 17:30:10 crc kubenswrapper[4700]: I0227 17:30:10.561606 4700 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/583579de-19a5-4728-97f9-37398ce4a452-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:30:10 crc kubenswrapper[4700]: I0227 17:30:10.561619 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcbfz\" (UniqueName: \"kubernetes.io/projected/583579de-19a5-4728-97f9-37398ce4a452-kube-api-access-hcbfz\") on node \"crc\" DevicePath \"\"" Feb 27 17:30:10 crc kubenswrapper[4700]: I0227 17:30:10.874520 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5f27j" event={"ID":"583579de-19a5-4728-97f9-37398ce4a452","Type":"ContainerDied","Data":"11d3ee8ab96a449b5ebc24d08687a12c7f709aebc2e425eb9fd2fac06f6c408c"} Feb 27 17:30:10 crc kubenswrapper[4700]: I0227 17:30:10.874595 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11d3ee8ab96a449b5ebc24d08687a12c7f709aebc2e425eb9fd2fac06f6c408c" Feb 27 17:30:10 crc kubenswrapper[4700]: I0227 17:30:10.874766 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5f27j" Feb 27 17:30:11 crc kubenswrapper[4700]: I0227 17:30:11.043631 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t8clg"] Feb 27 17:30:11 crc kubenswrapper[4700]: E0227 17:30:11.044108 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7207687-fb96-4b11-8c3b-c9f9467e0373" containerName="collect-profiles" Feb 27 17:30:11 crc kubenswrapper[4700]: I0227 17:30:11.044129 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7207687-fb96-4b11-8c3b-c9f9467e0373" containerName="collect-profiles" Feb 27 17:30:11 crc kubenswrapper[4700]: E0227 17:30:11.044144 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="583579de-19a5-4728-97f9-37398ce4a452" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 27 17:30:11 crc kubenswrapper[4700]: I0227 17:30:11.044153 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="583579de-19a5-4728-97f9-37398ce4a452" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 27 17:30:11 crc kubenswrapper[4700]: E0227 17:30:11.044171 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b6d9b14-d4a9-4633-bba4-197eac2766a4" containerName="oc" Feb 27 17:30:11 crc kubenswrapper[4700]: I0227 17:30:11.044177 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b6d9b14-d4a9-4633-bba4-197eac2766a4" containerName="oc" Feb 27 17:30:11 crc kubenswrapper[4700]: I0227 17:30:11.044382 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="583579de-19a5-4728-97f9-37398ce4a452" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 27 17:30:11 crc kubenswrapper[4700]: I0227 17:30:11.044395 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7207687-fb96-4b11-8c3b-c9f9467e0373" containerName="collect-profiles" Feb 27 17:30:11 crc kubenswrapper[4700]: I0227 17:30:11.044408 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b6d9b14-d4a9-4633-bba4-197eac2766a4" containerName="oc" Feb 27 17:30:11 crc kubenswrapper[4700]: I0227 17:30:11.045118 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t8clg"] Feb 27 17:30:11 crc kubenswrapper[4700]: I0227 17:30:11.045207 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t8clg" Feb 27 17:30:11 crc kubenswrapper[4700]: I0227 17:30:11.049394 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 17:30:11 crc kubenswrapper[4700]: I0227 17:30:11.049697 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 17:30:11 crc kubenswrapper[4700]: I0227 17:30:11.049425 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dtdm4" Feb 27 17:30:11 crc kubenswrapper[4700]: I0227 17:30:11.049815 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 17:30:11 crc kubenswrapper[4700]: I0227 17:30:11.073000 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77wgv\" (UniqueName: \"kubernetes.io/projected/e72836a3-fb67-48cf-971a-e90210a3d43d-kube-api-access-77wgv\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-t8clg\" (UID: \"e72836a3-fb67-48cf-971a-e90210a3d43d\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t8clg" Feb 27 17:30:11 crc kubenswrapper[4700]: I0227 17:30:11.073257 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e72836a3-fb67-48cf-971a-e90210a3d43d-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-t8clg\" (UID: \"e72836a3-fb67-48cf-971a-e90210a3d43d\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t8clg" Feb 27 17:30:11 crc kubenswrapper[4700]: I0227 17:30:11.073358 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e72836a3-fb67-48cf-971a-e90210a3d43d-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-t8clg\" (UID: \"e72836a3-fb67-48cf-971a-e90210a3d43d\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t8clg" Feb 27 17:30:11 crc kubenswrapper[4700]: I0227 17:30:11.176174 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e72836a3-fb67-48cf-971a-e90210a3d43d-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-t8clg\" (UID: \"e72836a3-fb67-48cf-971a-e90210a3d43d\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t8clg" Feb 27 17:30:11 crc kubenswrapper[4700]: I0227 17:30:11.176665 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e72836a3-fb67-48cf-971a-e90210a3d43d-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-t8clg\" (UID: \"e72836a3-fb67-48cf-971a-e90210a3d43d\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t8clg" Feb 27 17:30:11 crc kubenswrapper[4700]: I0227 17:30:11.177007 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77wgv\" (UniqueName: \"kubernetes.io/projected/e72836a3-fb67-48cf-971a-e90210a3d43d-kube-api-access-77wgv\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-t8clg\" (UID: \"e72836a3-fb67-48cf-971a-e90210a3d43d\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t8clg" Feb 27 17:30:11 crc kubenswrapper[4700]: I0227 17:30:11.181333 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e72836a3-fb67-48cf-971a-e90210a3d43d-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-t8clg\" (UID: \"e72836a3-fb67-48cf-971a-e90210a3d43d\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t8clg" Feb 27 17:30:11 crc kubenswrapper[4700]: I0227 17:30:11.194088 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e72836a3-fb67-48cf-971a-e90210a3d43d-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-t8clg\" (UID: \"e72836a3-fb67-48cf-971a-e90210a3d43d\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t8clg" Feb 27 17:30:11 crc kubenswrapper[4700]: I0227 17:30:11.202986 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77wgv\" (UniqueName: \"kubernetes.io/projected/e72836a3-fb67-48cf-971a-e90210a3d43d-kube-api-access-77wgv\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-t8clg\" (UID: \"e72836a3-fb67-48cf-971a-e90210a3d43d\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t8clg" Feb 27 17:30:11 crc kubenswrapper[4700]: I0227 17:30:11.369056 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t8clg" Feb 27 17:30:11 crc kubenswrapper[4700]: I0227 17:30:11.983243 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t8clg"] Feb 27 17:30:12 crc kubenswrapper[4700]: I0227 17:30:12.904889 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t8clg" event={"ID":"e72836a3-fb67-48cf-971a-e90210a3d43d","Type":"ContainerStarted","Data":"f33b7194714ff971026cccd17e2abdb5246a0d937a064dde466bcdf241424cee"} Feb 27 17:30:12 crc kubenswrapper[4700]: I0227 17:30:12.905315 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t8clg" event={"ID":"e72836a3-fb67-48cf-971a-e90210a3d43d","Type":"ContainerStarted","Data":"66e4785834987b0c07196555a2db78fd09ecc13b2f0afba9d7e654cb27c534b8"} Feb 27 17:30:12 crc kubenswrapper[4700]: I0227 17:30:12.926844 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t8clg" podStartSLOduration=2.471080208 podStartE2EDuration="2.926823037s" podCreationTimestamp="2026-02-27 17:30:10 +0000 UTC" firstStartedPulling="2026-02-27 17:30:11.985433832 +0000 UTC m=+1771.970746619" lastFinishedPulling="2026-02-27 17:30:12.441176701 +0000 UTC m=+1772.426489448" observedRunningTime="2026-02-27 17:30:12.918781815 +0000 UTC m=+1772.904094592" watchObservedRunningTime="2026-02-27 17:30:12.926823037 +0000 UTC m=+1772.912135784" Feb 27 17:30:15 crc kubenswrapper[4700]: I0227 17:30:15.983042 4700 scope.go:117] "RemoveContainer" containerID="88adee85741f363c1411884023846cd0aab91e2b1c4dca4a79262d1b15c4192f" Feb 27 17:30:15 crc kubenswrapper[4700]: E0227 17:30:15.984063 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:30:26 crc kubenswrapper[4700]: I0227 17:30:26.981296 4700 scope.go:117] "RemoveContainer" containerID="88adee85741f363c1411884023846cd0aab91e2b1c4dca4a79262d1b15c4192f" Feb 27 17:30:26 crc kubenswrapper[4700]: E0227 17:30:26.982185 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:30:29 crc kubenswrapper[4700]: I0227 17:30:29.052432 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-db-create-gx4c9"] Feb 27 17:30:29 crc kubenswrapper[4700]: I0227 17:30:29.064763 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-0076-account-create-update-hwxxq"] Feb 27 17:30:29 crc kubenswrapper[4700]: I0227 17:30:29.074491 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-0076-account-create-update-hwxxq"] Feb 27 17:30:29 crc kubenswrapper[4700]: I0227 17:30:29.084947 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-db-create-gx4c9"] Feb 27 17:30:31 crc kubenswrapper[4700]: I0227 17:30:31.008085 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18a5e20e-5c10-4966-9adc-c10670ddd4e3" path="/var/lib/kubelet/pods/18a5e20e-5c10-4966-9adc-c10670ddd4e3/volumes" Feb 27 17:30:31 crc kubenswrapper[4700]: I0227 17:30:31.010543 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0612657-780a-4388-9b79-ae606ed152e3" path="/var/lib/kubelet/pods/d0612657-780a-4388-9b79-ae606ed152e3/volumes" Feb 27 17:30:36 crc kubenswrapper[4700]: I0227 17:30:36.030904 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-2ppp5"] Feb 27 17:30:36 crc kubenswrapper[4700]: I0227 17:30:36.040542 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-2ppp5"] Feb 27 17:30:37 crc kubenswrapper[4700]: I0227 17:30:37.001413 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42f288bd-13a4-4e0c-a616-b09462acd2bf" path="/var/lib/kubelet/pods/42f288bd-13a4-4e0c-a616-b09462acd2bf/volumes" Feb 27 17:30:37 crc kubenswrapper[4700]: I0227 17:30:37.063547 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-b838-account-create-update-g6qs7"] Feb 27 17:30:37 crc kubenswrapper[4700]: I0227 17:30:37.075828 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-glgwk"] Feb 27 17:30:37 crc kubenswrapper[4700]: I0227 17:30:37.092097 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-4fb2-account-create-update-h494r"] Feb 27 17:30:37 crc kubenswrapper[4700]: I0227 17:30:37.104956 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-94be-account-create-update-snkzp"] Feb 27 17:30:37 crc kubenswrapper[4700]: I0227 17:30:37.114620 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-vfjlj"] Feb 27 17:30:37 crc kubenswrapper[4700]: I0227 17:30:37.124067 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-94be-account-create-update-snkzp"] Feb 27 17:30:37 crc kubenswrapper[4700]: I0227 17:30:37.132594 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-b838-account-create-update-g6qs7"] Feb 27 17:30:37 crc kubenswrapper[4700]: I0227 17:30:37.142166 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-glgwk"] Feb 27 17:30:37 crc kubenswrapper[4700]: I0227 17:30:37.150274 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-4fb2-account-create-update-h494r"] Feb 27 17:30:37 crc kubenswrapper[4700]: I0227 17:30:37.157993 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-vfjlj"] Feb 27 17:30:38 crc kubenswrapper[4700]: I0227 17:30:38.996178 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0af77e17-d33d-4d3b-8246-e38499ea68f7" path="/var/lib/kubelet/pods/0af77e17-d33d-4d3b-8246-e38499ea68f7/volumes" Feb 27 17:30:38 crc kubenswrapper[4700]: I0227 17:30:38.999524 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a889ce63-746b-45bc-9f99-487f7416a80c" path="/var/lib/kubelet/pods/a889ce63-746b-45bc-9f99-487f7416a80c/volumes" Feb 27 17:30:39 crc kubenswrapper[4700]: I0227 17:30:39.000673 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b393dd07-e44c-499b-8a95-81c50f78efe8" path="/var/lib/kubelet/pods/b393dd07-e44c-499b-8a95-81c50f78efe8/volumes" Feb 27 17:30:39 crc kubenswrapper[4700]: I0227 17:30:39.001809 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7812092-9d78-409a-80f8-4a4db92af41e" path="/var/lib/kubelet/pods/c7812092-9d78-409a-80f8-4a4db92af41e/volumes" Feb 27 17:30:39 crc kubenswrapper[4700]: I0227 17:30:39.002524 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9e7190d-50d9-4287-8f66-5aa298c5cdc7" path="/var/lib/kubelet/pods/c9e7190d-50d9-4287-8f66-5aa298c5cdc7/volumes" Feb 27 17:30:41 crc kubenswrapper[4700]: I0227 17:30:41.982635 4700 scope.go:117] "RemoveContainer" containerID="88adee85741f363c1411884023846cd0aab91e2b1c4dca4a79262d1b15c4192f" Feb 27 17:30:41 crc kubenswrapper[4700]: E0227 17:30:41.983172 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:30:43 crc kubenswrapper[4700]: I0227 17:30:43.038113 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-27j58"] Feb 27 17:30:43 crc kubenswrapper[4700]: I0227 17:30:43.051093 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-27j58"] Feb 27 17:30:44 crc kubenswrapper[4700]: I0227 17:30:44.996777 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5145ef3e-2efa-4f76-b395-179abf1b1818" path="/var/lib/kubelet/pods/5145ef3e-2efa-4f76-b395-179abf1b1818/volumes" Feb 27 17:30:52 crc kubenswrapper[4700]: I0227 17:30:52.982057 4700 scope.go:117] "RemoveContainer" containerID="88adee85741f363c1411884023846cd0aab91e2b1c4dca4a79262d1b15c4192f" Feb 27 17:30:52 crc kubenswrapper[4700]: E0227 17:30:52.983250 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:31:04 crc kubenswrapper[4700]: I0227 17:31:04.065467 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-9s7hk"] Feb 27 17:31:04 crc kubenswrapper[4700]: I0227 17:31:04.082250 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-9s7hk"] Feb 27 17:31:04 crc kubenswrapper[4700]: I0227 17:31:04.981498 4700 scope.go:117] "RemoveContainer" containerID="88adee85741f363c1411884023846cd0aab91e2b1c4dca4a79262d1b15c4192f" Feb 27 17:31:04 crc kubenswrapper[4700]: E0227 17:31:04.982116 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:31:05 crc kubenswrapper[4700]: I0227 17:31:05.002126 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3d01029-07d5-414c-badd-0d28f0db9730" path="/var/lib/kubelet/pods/c3d01029-07d5-414c-badd-0d28f0db9730/volumes" Feb 27 17:31:06 crc kubenswrapper[4700]: I0227 17:31:06.423352 4700 scope.go:117] "RemoveContainer" containerID="d493357cc638860117531b3c4db652280791a199e851e8bf7c70f533308e5f06" Feb 27 17:31:06 crc kubenswrapper[4700]: I0227 17:31:06.446305 4700 scope.go:117] "RemoveContainer" containerID="45057e85f95f295f9914750cfa486175715ffae747822c8d4025a5d9b9b95ad6" Feb 27 17:31:06 crc kubenswrapper[4700]: I0227 17:31:06.513721 4700 scope.go:117] "RemoveContainer" containerID="7dbfbad6525b94cb92f5314afed9b77c4b093833511b704ca4881f88a6e32547" Feb 27 17:31:06 crc kubenswrapper[4700]: I0227 17:31:06.552189 4700 scope.go:117] "RemoveContainer" containerID="4de383a1ce255a5abdcc76689987552575b674f8ee3b8a58361c39735e81964a" Feb 27 17:31:06 crc kubenswrapper[4700]: I0227 17:31:06.598726 4700 scope.go:117] "RemoveContainer" containerID="c513a087e1e918ce5827b7e53435b261f731d6a8770be071707b9a006ffa3bd9" Feb 27 17:31:06 crc kubenswrapper[4700]: I0227 17:31:06.702261 4700 scope.go:117] "RemoveContainer" containerID="dc2763324c06b0b064ffa165970464ae60e3ecfda6de3837813bddacda25008a" Feb 27 17:31:06 crc kubenswrapper[4700]: I0227 17:31:06.751296 4700 scope.go:117] "RemoveContainer" containerID="aa2b25a6011304a5a553000d459adb9829a298cf83d72a61403e200eafccaa22" Feb 27 17:31:06 crc kubenswrapper[4700]: I0227 17:31:06.774244 4700 scope.go:117] "RemoveContainer" containerID="a18083759d87601fd4c2f06b18c7f74979cf39abb7274d034a6781bb0f2eebc4" Feb 27 17:31:06 crc kubenswrapper[4700]: I0227 17:31:06.799858 4700 scope.go:117] "RemoveContainer" containerID="c69cc325940dde65fa7783b1d92d27f073d11b12f2222b9e713d3414cb599515" Feb 27 17:31:06 crc kubenswrapper[4700]: I0227 17:31:06.830347 4700 scope.go:117] "RemoveContainer" containerID="3e260f40afc484c8c52c1b94079e2bccc68fd3802ca10b847bcaf2e2dcaf288d" Feb 27 17:31:06 crc kubenswrapper[4700]: I0227 17:31:06.858609 4700 scope.go:117] "RemoveContainer" containerID="3c2c275c3cc206ac97d16d9dd3bc253b01082dc180beb82d9a0f98a67dae1c47" Feb 27 17:31:06 crc kubenswrapper[4700]: I0227 17:31:06.884542 4700 scope.go:117] "RemoveContainer" containerID="f90f21dfd58e8098c3e9c424632b9b719bc730c56acfc19e81fcfda0344ecaeb" Feb 27 17:31:06 crc kubenswrapper[4700]: I0227 17:31:06.918820 4700 scope.go:117] "RemoveContainer" containerID="edb119d383758b51fad5555188e081aba816283333b85f5639fad1cfbaf98c58" Feb 27 17:31:11 crc kubenswrapper[4700]: I0227 17:31:11.043564 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-7xvrx"] Feb 27 17:31:11 crc kubenswrapper[4700]: I0227 17:31:11.056336 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-1943-account-create-update-h8f9l"] Feb 27 17:31:11 crc kubenswrapper[4700]: I0227 17:31:11.070333 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-7xvrx"] Feb 27 17:31:11 crc kubenswrapper[4700]: I0227 17:31:11.080271 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-1943-account-create-update-h8f9l"] Feb 27 17:31:12 crc kubenswrapper[4700]: I0227 17:31:12.042339 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-rc9w6"] Feb 27 17:31:12 crc kubenswrapper[4700]: I0227 17:31:12.059033 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-2c01-account-create-update-crm9x"] Feb 27 17:31:12 crc kubenswrapper[4700]: I0227 17:31:12.073400 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-rc9w6"] Feb 27 17:31:12 crc kubenswrapper[4700]: I0227 17:31:12.085027 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-2c01-account-create-update-crm9x"] Feb 27 17:31:13 crc kubenswrapper[4700]: I0227 17:31:13.004692 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2062608c-6759-45bd-9879-2a96af693b31" path="/var/lib/kubelet/pods/2062608c-6759-45bd-9879-2a96af693b31/volumes" Feb 27 17:31:13 crc kubenswrapper[4700]: I0227 17:31:13.009566 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e67eb82-300d-46bd-b1c6-d2607bf45ccc" path="/var/lib/kubelet/pods/2e67eb82-300d-46bd-b1c6-d2607bf45ccc/volumes" Feb 27 17:31:13 crc kubenswrapper[4700]: I0227 17:31:13.012869 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5d89c26-6133-4987-8c61-cc19c80567e0" path="/var/lib/kubelet/pods/f5d89c26-6133-4987-8c61-cc19c80567e0/volumes" Feb 27 17:31:13 crc kubenswrapper[4700]: I0227 17:31:13.016124 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe426647-f5e6-46bf-b187-8f8cb62b06e1" path="/var/lib/kubelet/pods/fe426647-f5e6-46bf-b187-8f8cb62b06e1/volumes" Feb 27 17:31:16 crc kubenswrapper[4700]: I0227 17:31:16.039748 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-f5xkq"] Feb 27 17:31:16 crc kubenswrapper[4700]: I0227 17:31:16.052506 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-f5xkq"] Feb 27 17:31:16 crc kubenswrapper[4700]: I0227 17:31:16.981679 4700 scope.go:117] "RemoveContainer" containerID="88adee85741f363c1411884023846cd0aab91e2b1c4dca4a79262d1b15c4192f" Feb 27 17:31:16 crc kubenswrapper[4700]: E0227 17:31:16.982735 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:31:16 crc kubenswrapper[4700]: I0227 17:31:16.997426 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a11aafed-1003-4121-9e6e-e5f20c40389f" path="/var/lib/kubelet/pods/a11aafed-1003-4121-9e6e-e5f20c40389f/volumes" Feb 27 17:31:22 crc kubenswrapper[4700]: I0227 17:31:22.055828 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-582c-account-create-update-jqxth"] Feb 27 17:31:22 crc kubenswrapper[4700]: I0227 17:31:22.071961 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-582c-account-create-update-jqxth"] Feb 27 17:31:22 crc kubenswrapper[4700]: I0227 17:31:22.994833 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ffc04905-1f57-4888-bb88-db5187753b5f" path="/var/lib/kubelet/pods/ffc04905-1f57-4888-bb88-db5187753b5f/volumes" Feb 27 17:31:27 crc kubenswrapper[4700]: I0227 17:31:27.037220 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-db-sync-5zzh7"] Feb 27 17:31:27 crc kubenswrapper[4700]: I0227 17:31:27.048897 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-db-sync-5zzh7"] Feb 27 17:31:27 crc kubenswrapper[4700]: I0227 17:31:27.981304 4700 scope.go:117] "RemoveContainer" containerID="88adee85741f363c1411884023846cd0aab91e2b1c4dca4a79262d1b15c4192f" Feb 27 17:31:27 crc kubenswrapper[4700]: E0227 17:31:27.981722 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:31:28 crc kubenswrapper[4700]: I0227 17:31:28.043690 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-qm46k"] Feb 27 17:31:28 crc kubenswrapper[4700]: I0227 17:31:28.063449 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-qm46k"] Feb 27 17:31:28 crc kubenswrapper[4700]: I0227 17:31:28.995699 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16189a3f-0a2b-453c-8bb1-8e8c4d79c13f" path="/var/lib/kubelet/pods/16189a3f-0a2b-453c-8bb1-8e8c4d79c13f/volumes" Feb 27 17:31:28 crc kubenswrapper[4700]: I0227 17:31:28.997839 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a238beaf-451f-452e-a30f-09779ae217c1" path="/var/lib/kubelet/pods/a238beaf-451f-452e-a30f-09779ae217c1/volumes" Feb 27 17:31:40 crc kubenswrapper[4700]: I0227 17:31:40.982884 4700 scope.go:117] "RemoveContainer" containerID="88adee85741f363c1411884023846cd0aab91e2b1c4dca4a79262d1b15c4192f" Feb 27 17:31:40 crc kubenswrapper[4700]: E0227 17:31:40.983971 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:31:43 crc kubenswrapper[4700]: I0227 17:31:43.044652 4700 generic.go:334] "Generic (PLEG): container finished" podID="e72836a3-fb67-48cf-971a-e90210a3d43d" containerID="f33b7194714ff971026cccd17e2abdb5246a0d937a064dde466bcdf241424cee" exitCode=0 Feb 27 17:31:43 crc kubenswrapper[4700]: I0227 17:31:43.044771 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t8clg" event={"ID":"e72836a3-fb67-48cf-971a-e90210a3d43d","Type":"ContainerDied","Data":"f33b7194714ff971026cccd17e2abdb5246a0d937a064dde466bcdf241424cee"} Feb 27 17:31:44 crc kubenswrapper[4700]: I0227 17:31:44.507188 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t8clg" Feb 27 17:31:44 crc kubenswrapper[4700]: I0227 17:31:44.597833 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e72836a3-fb67-48cf-971a-e90210a3d43d-ssh-key-openstack-edpm-ipam\") pod \"e72836a3-fb67-48cf-971a-e90210a3d43d\" (UID: \"e72836a3-fb67-48cf-971a-e90210a3d43d\") " Feb 27 17:31:44 crc kubenswrapper[4700]: I0227 17:31:44.597951 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77wgv\" (UniqueName: \"kubernetes.io/projected/e72836a3-fb67-48cf-971a-e90210a3d43d-kube-api-access-77wgv\") pod \"e72836a3-fb67-48cf-971a-e90210a3d43d\" (UID: \"e72836a3-fb67-48cf-971a-e90210a3d43d\") " Feb 27 17:31:44 crc kubenswrapper[4700]: I0227 17:31:44.598013 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e72836a3-fb67-48cf-971a-e90210a3d43d-inventory\") pod \"e72836a3-fb67-48cf-971a-e90210a3d43d\" (UID: \"e72836a3-fb67-48cf-971a-e90210a3d43d\") " Feb 27 17:31:44 crc kubenswrapper[4700]: I0227 17:31:44.604798 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e72836a3-fb67-48cf-971a-e90210a3d43d-kube-api-access-77wgv" (OuterVolumeSpecName: "kube-api-access-77wgv") pod "e72836a3-fb67-48cf-971a-e90210a3d43d" (UID: "e72836a3-fb67-48cf-971a-e90210a3d43d"). InnerVolumeSpecName "kube-api-access-77wgv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:31:44 crc kubenswrapper[4700]: I0227 17:31:44.626971 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e72836a3-fb67-48cf-971a-e90210a3d43d-inventory" (OuterVolumeSpecName: "inventory") pod "e72836a3-fb67-48cf-971a-e90210a3d43d" (UID: "e72836a3-fb67-48cf-971a-e90210a3d43d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:31:44 crc kubenswrapper[4700]: I0227 17:31:44.643336 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e72836a3-fb67-48cf-971a-e90210a3d43d-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "e72836a3-fb67-48cf-971a-e90210a3d43d" (UID: "e72836a3-fb67-48cf-971a-e90210a3d43d"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:31:44 crc kubenswrapper[4700]: I0227 17:31:44.700900 4700 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e72836a3-fb67-48cf-971a-e90210a3d43d-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 17:31:44 crc kubenswrapper[4700]: I0227 17:31:44.700936 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77wgv\" (UniqueName: \"kubernetes.io/projected/e72836a3-fb67-48cf-971a-e90210a3d43d-kube-api-access-77wgv\") on node \"crc\" DevicePath \"\"" Feb 27 17:31:44 crc kubenswrapper[4700]: I0227 17:31:44.700949 4700 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e72836a3-fb67-48cf-971a-e90210a3d43d-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 17:31:45 crc kubenswrapper[4700]: I0227 17:31:45.070654 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t8clg" event={"ID":"e72836a3-fb67-48cf-971a-e90210a3d43d","Type":"ContainerDied","Data":"66e4785834987b0c07196555a2db78fd09ecc13b2f0afba9d7e654cb27c534b8"} Feb 27 17:31:45 crc kubenswrapper[4700]: I0227 17:31:45.070730 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="66e4785834987b0c07196555a2db78fd09ecc13b2f0afba9d7e654cb27c534b8" Feb 27 17:31:45 crc kubenswrapper[4700]: I0227 17:31:45.070830 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t8clg" Feb 27 17:31:45 crc kubenswrapper[4700]: I0227 17:31:45.168529 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xndbg"] Feb 27 17:31:45 crc kubenswrapper[4700]: E0227 17:31:45.168924 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e72836a3-fb67-48cf-971a-e90210a3d43d" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 27 17:31:45 crc kubenswrapper[4700]: I0227 17:31:45.168942 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="e72836a3-fb67-48cf-971a-e90210a3d43d" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 27 17:31:45 crc kubenswrapper[4700]: I0227 17:31:45.169137 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="e72836a3-fb67-48cf-971a-e90210a3d43d" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 27 17:31:45 crc kubenswrapper[4700]: I0227 17:31:45.169836 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xndbg" Feb 27 17:31:45 crc kubenswrapper[4700]: I0227 17:31:45.171493 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dtdm4" Feb 27 17:31:45 crc kubenswrapper[4700]: I0227 17:31:45.171884 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 17:31:45 crc kubenswrapper[4700]: I0227 17:31:45.173976 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 17:31:45 crc kubenswrapper[4700]: I0227 17:31:45.177056 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 17:31:45 crc kubenswrapper[4700]: I0227 17:31:45.191199 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xndbg"] Feb 27 17:31:45 crc kubenswrapper[4700]: I0227 17:31:45.312121 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/493471ba-3317-48f3-accc-a2ddb143984c-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xndbg\" (UID: \"493471ba-3317-48f3-accc-a2ddb143984c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xndbg" Feb 27 17:31:45 crc kubenswrapper[4700]: I0227 17:31:45.312226 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/493471ba-3317-48f3-accc-a2ddb143984c-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xndbg\" (UID: \"493471ba-3317-48f3-accc-a2ddb143984c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xndbg" Feb 27 17:31:45 crc kubenswrapper[4700]: I0227 17:31:45.312276 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wxtc\" (UniqueName: \"kubernetes.io/projected/493471ba-3317-48f3-accc-a2ddb143984c-kube-api-access-9wxtc\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xndbg\" (UID: \"493471ba-3317-48f3-accc-a2ddb143984c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xndbg" Feb 27 17:31:45 crc kubenswrapper[4700]: I0227 17:31:45.414343 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/493471ba-3317-48f3-accc-a2ddb143984c-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xndbg\" (UID: \"493471ba-3317-48f3-accc-a2ddb143984c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xndbg" Feb 27 17:31:45 crc kubenswrapper[4700]: I0227 17:31:45.414755 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wxtc\" (UniqueName: \"kubernetes.io/projected/493471ba-3317-48f3-accc-a2ddb143984c-kube-api-access-9wxtc\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xndbg\" (UID: \"493471ba-3317-48f3-accc-a2ddb143984c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xndbg" Feb 27 17:31:45 crc kubenswrapper[4700]: I0227 17:31:45.414922 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/493471ba-3317-48f3-accc-a2ddb143984c-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xndbg\" (UID: \"493471ba-3317-48f3-accc-a2ddb143984c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xndbg" Feb 27 17:31:45 crc kubenswrapper[4700]: I0227 17:31:45.419572 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/493471ba-3317-48f3-accc-a2ddb143984c-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xndbg\" (UID: \"493471ba-3317-48f3-accc-a2ddb143984c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xndbg" Feb 27 17:31:45 crc kubenswrapper[4700]: I0227 17:31:45.420346 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/493471ba-3317-48f3-accc-a2ddb143984c-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xndbg\" (UID: \"493471ba-3317-48f3-accc-a2ddb143984c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xndbg" Feb 27 17:31:45 crc kubenswrapper[4700]: I0227 17:31:45.439340 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wxtc\" (UniqueName: \"kubernetes.io/projected/493471ba-3317-48f3-accc-a2ddb143984c-kube-api-access-9wxtc\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xndbg\" (UID: \"493471ba-3317-48f3-accc-a2ddb143984c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xndbg" Feb 27 17:31:45 crc kubenswrapper[4700]: I0227 17:31:45.490318 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xndbg" Feb 27 17:31:46 crc kubenswrapper[4700]: I0227 17:31:46.217547 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xndbg"] Feb 27 17:31:46 crc kubenswrapper[4700]: W0227 17:31:46.234926 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod493471ba_3317_48f3_accc_a2ddb143984c.slice/crio-8759bd966cc0bec3320edc3293fc1bf9ff72a0fb5f06f6ea747ffece2b89f39d WatchSource:0}: Error finding container 8759bd966cc0bec3320edc3293fc1bf9ff72a0fb5f06f6ea747ffece2b89f39d: Status 404 returned error can't find the container with id 8759bd966cc0bec3320edc3293fc1bf9ff72a0fb5f06f6ea747ffece2b89f39d Feb 27 17:31:46 crc kubenswrapper[4700]: I0227 17:31:46.238877 4700 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 17:31:47 crc kubenswrapper[4700]: I0227 17:31:47.094396 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xndbg" event={"ID":"493471ba-3317-48f3-accc-a2ddb143984c","Type":"ContainerStarted","Data":"ebe86de496afb20bdfa27886e72b4965808f83e86a48659f31d810bc45b70151"} Feb 27 17:31:47 crc kubenswrapper[4700]: I0227 17:31:47.094889 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xndbg" event={"ID":"493471ba-3317-48f3-accc-a2ddb143984c","Type":"ContainerStarted","Data":"8759bd966cc0bec3320edc3293fc1bf9ff72a0fb5f06f6ea747ffece2b89f39d"} Feb 27 17:31:47 crc kubenswrapper[4700]: I0227 17:31:47.118511 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xndbg" podStartSLOduration=1.6479838789999999 podStartE2EDuration="2.118484237s" podCreationTimestamp="2026-02-27 17:31:45 +0000 UTC" firstStartedPulling="2026-02-27 17:31:46.238489752 +0000 UTC m=+1866.223802509" lastFinishedPulling="2026-02-27 17:31:46.70899012 +0000 UTC m=+1866.694302867" observedRunningTime="2026-02-27 17:31:47.116072644 +0000 UTC m=+1867.101385391" watchObservedRunningTime="2026-02-27 17:31:47.118484237 +0000 UTC m=+1867.103796994" Feb 27 17:31:51 crc kubenswrapper[4700]: I0227 17:31:51.981095 4700 scope.go:117] "RemoveContainer" containerID="88adee85741f363c1411884023846cd0aab91e2b1c4dca4a79262d1b15c4192f" Feb 27 17:31:51 crc kubenswrapper[4700]: E0227 17:31:51.982147 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:32:00 crc kubenswrapper[4700]: I0227 17:32:00.142714 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536892-ndtg4"] Feb 27 17:32:00 crc kubenswrapper[4700]: I0227 17:32:00.146592 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536892-ndtg4" Feb 27 17:32:00 crc kubenswrapper[4700]: I0227 17:32:00.149767 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:32:00 crc kubenswrapper[4700]: I0227 17:32:00.149915 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:32:00 crc kubenswrapper[4700]: I0227 17:32:00.151073 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d62kc" Feb 27 17:32:00 crc kubenswrapper[4700]: I0227 17:32:00.159312 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536892-ndtg4"] Feb 27 17:32:00 crc kubenswrapper[4700]: I0227 17:32:00.253559 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbfgd\" (UniqueName: \"kubernetes.io/projected/5a6c1496-be0e-4733-8176-19c62bbc329e-kube-api-access-rbfgd\") pod \"auto-csr-approver-29536892-ndtg4\" (UID: \"5a6c1496-be0e-4733-8176-19c62bbc329e\") " pod="openshift-infra/auto-csr-approver-29536892-ndtg4" Feb 27 17:32:00 crc kubenswrapper[4700]: I0227 17:32:00.354706 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbfgd\" (UniqueName: \"kubernetes.io/projected/5a6c1496-be0e-4733-8176-19c62bbc329e-kube-api-access-rbfgd\") pod \"auto-csr-approver-29536892-ndtg4\" (UID: \"5a6c1496-be0e-4733-8176-19c62bbc329e\") " pod="openshift-infra/auto-csr-approver-29536892-ndtg4" Feb 27 17:32:00 crc kubenswrapper[4700]: I0227 17:32:00.385337 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbfgd\" (UniqueName: \"kubernetes.io/projected/5a6c1496-be0e-4733-8176-19c62bbc329e-kube-api-access-rbfgd\") pod \"auto-csr-approver-29536892-ndtg4\" (UID: \"5a6c1496-be0e-4733-8176-19c62bbc329e\") " pod="openshift-infra/auto-csr-approver-29536892-ndtg4" Feb 27 17:32:00 crc kubenswrapper[4700]: I0227 17:32:00.472872 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536892-ndtg4" Feb 27 17:32:00 crc kubenswrapper[4700]: I0227 17:32:00.942303 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536892-ndtg4"] Feb 27 17:32:00 crc kubenswrapper[4700]: W0227 17:32:00.966295 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a6c1496_be0e_4733_8176_19c62bbc329e.slice/crio-046d7478649e5ca966762a742316d0d1ffb77c129a2821f1bfe09f34f37878b0 WatchSource:0}: Error finding container 046d7478649e5ca966762a742316d0d1ffb77c129a2821f1bfe09f34f37878b0: Status 404 returned error can't find the container with id 046d7478649e5ca966762a742316d0d1ffb77c129a2821f1bfe09f34f37878b0 Feb 27 17:32:01 crc kubenswrapper[4700]: I0227 17:32:01.256526 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536892-ndtg4" event={"ID":"5a6c1496-be0e-4733-8176-19c62bbc329e","Type":"ContainerStarted","Data":"046d7478649e5ca966762a742316d0d1ffb77c129a2821f1bfe09f34f37878b0"} Feb 27 17:32:02 crc kubenswrapper[4700]: I0227 17:32:02.282848 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536892-ndtg4" podStartSLOduration=1.36880844 podStartE2EDuration="2.282817633s" podCreationTimestamp="2026-02-27 17:32:00 +0000 UTC" firstStartedPulling="2026-02-27 17:32:00.972614693 +0000 UTC m=+1880.957927480" lastFinishedPulling="2026-02-27 17:32:01.886623926 +0000 UTC m=+1881.871936673" observedRunningTime="2026-02-27 17:32:02.277438471 +0000 UTC m=+1882.262751218" watchObservedRunningTime="2026-02-27 17:32:02.282817633 +0000 UTC m=+1882.268130400" Feb 27 17:32:03 crc kubenswrapper[4700]: I0227 17:32:03.283281 4700 generic.go:334] "Generic (PLEG): container finished" podID="5a6c1496-be0e-4733-8176-19c62bbc329e" containerID="3ed6861afca6d101aa226387a12b0eeb3f4daadb50454b739b39d920a10f0cf2" exitCode=0 Feb 27 17:32:03 crc kubenswrapper[4700]: I0227 17:32:03.283346 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536892-ndtg4" event={"ID":"5a6c1496-be0e-4733-8176-19c62bbc329e","Type":"ContainerDied","Data":"3ed6861afca6d101aa226387a12b0eeb3f4daadb50454b739b39d920a10f0cf2"} Feb 27 17:32:04 crc kubenswrapper[4700]: I0227 17:32:04.798182 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536892-ndtg4" Feb 27 17:32:04 crc kubenswrapper[4700]: I0227 17:32:04.963678 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbfgd\" (UniqueName: \"kubernetes.io/projected/5a6c1496-be0e-4733-8176-19c62bbc329e-kube-api-access-rbfgd\") pod \"5a6c1496-be0e-4733-8176-19c62bbc329e\" (UID: \"5a6c1496-be0e-4733-8176-19c62bbc329e\") " Feb 27 17:32:04 crc kubenswrapper[4700]: I0227 17:32:04.974074 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a6c1496-be0e-4733-8176-19c62bbc329e-kube-api-access-rbfgd" (OuterVolumeSpecName: "kube-api-access-rbfgd") pod "5a6c1496-be0e-4733-8176-19c62bbc329e" (UID: "5a6c1496-be0e-4733-8176-19c62bbc329e"). InnerVolumeSpecName "kube-api-access-rbfgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:32:04 crc kubenswrapper[4700]: I0227 17:32:04.982163 4700 scope.go:117] "RemoveContainer" containerID="88adee85741f363c1411884023846cd0aab91e2b1c4dca4a79262d1b15c4192f" Feb 27 17:32:04 crc kubenswrapper[4700]: E0227 17:32:04.982620 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:32:05 crc kubenswrapper[4700]: I0227 17:32:05.067138 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbfgd\" (UniqueName: \"kubernetes.io/projected/5a6c1496-be0e-4733-8176-19c62bbc329e-kube-api-access-rbfgd\") on node \"crc\" DevicePath \"\"" Feb 27 17:32:05 crc kubenswrapper[4700]: I0227 17:32:05.325441 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536892-ndtg4" event={"ID":"5a6c1496-be0e-4733-8176-19c62bbc329e","Type":"ContainerDied","Data":"046d7478649e5ca966762a742316d0d1ffb77c129a2821f1bfe09f34f37878b0"} Feb 27 17:32:05 crc kubenswrapper[4700]: I0227 17:32:05.325544 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="046d7478649e5ca966762a742316d0d1ffb77c129a2821f1bfe09f34f37878b0" Feb 27 17:32:05 crc kubenswrapper[4700]: I0227 17:32:05.325583 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536892-ndtg4" Feb 27 17:32:05 crc kubenswrapper[4700]: I0227 17:32:05.375370 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536886-87g88"] Feb 27 17:32:05 crc kubenswrapper[4700]: I0227 17:32:05.387001 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536886-87g88"] Feb 27 17:32:06 crc kubenswrapper[4700]: I0227 17:32:06.997850 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4b01082-c066-4fbd-aef5-169e9ec31fff" path="/var/lib/kubelet/pods/a4b01082-c066-4fbd-aef5-169e9ec31fff/volumes" Feb 27 17:32:07 crc kubenswrapper[4700]: I0227 17:32:07.294994 4700 scope.go:117] "RemoveContainer" containerID="05e0c2020c9571a38a99efb4ec606145ffe49be0926c426091baeabfdfac356b" Feb 27 17:32:07 crc kubenswrapper[4700]: I0227 17:32:07.338493 4700 scope.go:117] "RemoveContainer" containerID="4b7f93b578e644781c33d2b3aa82faa1979c30b013b09b4efe7e74cb9564891a" Feb 27 17:32:07 crc kubenswrapper[4700]: I0227 17:32:07.421231 4700 scope.go:117] "RemoveContainer" containerID="89b673903cbfb2a6d5ce565347af43128ab227341122662ec2b9e57c4cb16400" Feb 27 17:32:07 crc kubenswrapper[4700]: I0227 17:32:07.457897 4700 scope.go:117] "RemoveContainer" containerID="66ff4804ed8e0b99ba4fee444456685d8955deb6348ffb8f52c49d90997aa8b6" Feb 27 17:32:07 crc kubenswrapper[4700]: I0227 17:32:07.504645 4700 scope.go:117] "RemoveContainer" containerID="715c57270ee2cdfd98da501c87de458e60122b453b9ff5f05607477c5dfcdc44" Feb 27 17:32:07 crc kubenswrapper[4700]: I0227 17:32:07.573419 4700 scope.go:117] "RemoveContainer" containerID="c743a30c2358f798b11a6ad8bb7d06f0869f65ded2d52c2eb9d11e6bbafa0117" Feb 27 17:32:07 crc kubenswrapper[4700]: I0227 17:32:07.600843 4700 scope.go:117] "RemoveContainer" containerID="a1755da87ed07a85b745e9409058de03e937db1f3700433133f4bcecfb9f0aa4" Feb 27 17:32:07 crc kubenswrapper[4700]: I0227 17:32:07.631425 4700 scope.go:117] "RemoveContainer" containerID="3cc4589589d2611cf5d42961ae0bae28311f2045634dd9d7392d4f4d2a990485" Feb 27 17:32:07 crc kubenswrapper[4700]: I0227 17:32:07.714419 4700 scope.go:117] "RemoveContainer" containerID="008e6edb1f631b4dc678019ab290f43fec55bb8f1539b3845f7fd56bbede54c2" Feb 27 17:32:09 crc kubenswrapper[4700]: I0227 17:32:09.038868 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-6b5ck"] Feb 27 17:32:09 crc kubenswrapper[4700]: I0227 17:32:09.056833 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-6b5ck"] Feb 27 17:32:11 crc kubenswrapper[4700]: I0227 17:32:10.999943 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f209c04-79ea-4048-9d31-6222874de091" path="/var/lib/kubelet/pods/3f209c04-79ea-4048-9d31-6222874de091/volumes" Feb 27 17:32:16 crc kubenswrapper[4700]: I0227 17:32:16.063098 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-dmf4m"] Feb 27 17:32:16 crc kubenswrapper[4700]: I0227 17:32:16.081753 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-dmf4m"] Feb 27 17:32:17 crc kubenswrapper[4700]: I0227 17:32:16.999721 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="989eacab-9c48-4375-b7a0-f2c7f40d17ca" path="/var/lib/kubelet/pods/989eacab-9c48-4375-b7a0-f2c7f40d17ca/volumes" Feb 27 17:32:17 crc kubenswrapper[4700]: I0227 17:32:17.048785 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-vrb9h"] Feb 27 17:32:17 crc kubenswrapper[4700]: I0227 17:32:17.064551 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-vrb9h"] Feb 27 17:32:17 crc kubenswrapper[4700]: I0227 17:32:17.075285 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-8xqhb"] Feb 27 17:32:17 crc kubenswrapper[4700]: I0227 17:32:17.111116 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-8xqhb"] Feb 27 17:32:18 crc kubenswrapper[4700]: I0227 17:32:18.981408 4700 scope.go:117] "RemoveContainer" containerID="88adee85741f363c1411884023846cd0aab91e2b1c4dca4a79262d1b15c4192f" Feb 27 17:32:18 crc kubenswrapper[4700]: E0227 17:32:18.982037 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:32:18 crc kubenswrapper[4700]: I0227 17:32:18.996919 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84901b1e-0055-4489-8009-711c1676e51f" path="/var/lib/kubelet/pods/84901b1e-0055-4489-8009-711c1676e51f/volumes" Feb 27 17:32:18 crc kubenswrapper[4700]: I0227 17:32:18.997829 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f72833f0-cf5f-4bf8-bb5f-1f65477c4b76" path="/var/lib/kubelet/pods/f72833f0-cf5f-4bf8-bb5f-1f65477c4b76/volumes" Feb 27 17:32:31 crc kubenswrapper[4700]: I0227 17:32:31.982258 4700 scope.go:117] "RemoveContainer" containerID="88adee85741f363c1411884023846cd0aab91e2b1c4dca4a79262d1b15c4192f" Feb 27 17:32:31 crc kubenswrapper[4700]: E0227 17:32:31.983643 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:32:33 crc kubenswrapper[4700]: I0227 17:32:33.061569 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-mhcf5"] Feb 27 17:32:33 crc kubenswrapper[4700]: I0227 17:32:33.073574 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-mhcf5"] Feb 27 17:32:35 crc kubenswrapper[4700]: I0227 17:32:35.001385 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd3817a7-c2ab-43f0-9c74-b477ac59fa52" path="/var/lib/kubelet/pods/dd3817a7-c2ab-43f0-9c74-b477ac59fa52/volumes" Feb 27 17:32:42 crc kubenswrapper[4700]: I0227 17:32:42.981544 4700 scope.go:117] "RemoveContainer" containerID="88adee85741f363c1411884023846cd0aab91e2b1c4dca4a79262d1b15c4192f" Feb 27 17:32:42 crc kubenswrapper[4700]: E0227 17:32:42.982585 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:32:57 crc kubenswrapper[4700]: I0227 17:32:57.939395 4700 generic.go:334] "Generic (PLEG): container finished" podID="493471ba-3317-48f3-accc-a2ddb143984c" containerID="ebe86de496afb20bdfa27886e72b4965808f83e86a48659f31d810bc45b70151" exitCode=0 Feb 27 17:32:57 crc kubenswrapper[4700]: I0227 17:32:57.939515 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xndbg" event={"ID":"493471ba-3317-48f3-accc-a2ddb143984c","Type":"ContainerDied","Data":"ebe86de496afb20bdfa27886e72b4965808f83e86a48659f31d810bc45b70151"} Feb 27 17:32:57 crc kubenswrapper[4700]: I0227 17:32:57.981662 4700 scope.go:117] "RemoveContainer" containerID="88adee85741f363c1411884023846cd0aab91e2b1c4dca4a79262d1b15c4192f" Feb 27 17:32:57 crc kubenswrapper[4700]: E0227 17:32:57.982454 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:32:59 crc kubenswrapper[4700]: I0227 17:32:59.469982 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xndbg" Feb 27 17:32:59 crc kubenswrapper[4700]: I0227 17:32:59.605159 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/493471ba-3317-48f3-accc-a2ddb143984c-ssh-key-openstack-edpm-ipam\") pod \"493471ba-3317-48f3-accc-a2ddb143984c\" (UID: \"493471ba-3317-48f3-accc-a2ddb143984c\") " Feb 27 17:32:59 crc kubenswrapper[4700]: I0227 17:32:59.605432 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wxtc\" (UniqueName: \"kubernetes.io/projected/493471ba-3317-48f3-accc-a2ddb143984c-kube-api-access-9wxtc\") pod \"493471ba-3317-48f3-accc-a2ddb143984c\" (UID: \"493471ba-3317-48f3-accc-a2ddb143984c\") " Feb 27 17:32:59 crc kubenswrapper[4700]: I0227 17:32:59.605525 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/493471ba-3317-48f3-accc-a2ddb143984c-inventory\") pod \"493471ba-3317-48f3-accc-a2ddb143984c\" (UID: \"493471ba-3317-48f3-accc-a2ddb143984c\") " Feb 27 17:32:59 crc kubenswrapper[4700]: I0227 17:32:59.616128 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/493471ba-3317-48f3-accc-a2ddb143984c-kube-api-access-9wxtc" (OuterVolumeSpecName: "kube-api-access-9wxtc") pod "493471ba-3317-48f3-accc-a2ddb143984c" (UID: "493471ba-3317-48f3-accc-a2ddb143984c"). InnerVolumeSpecName "kube-api-access-9wxtc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:32:59 crc kubenswrapper[4700]: I0227 17:32:59.656181 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/493471ba-3317-48f3-accc-a2ddb143984c-inventory" (OuterVolumeSpecName: "inventory") pod "493471ba-3317-48f3-accc-a2ddb143984c" (UID: "493471ba-3317-48f3-accc-a2ddb143984c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:32:59 crc kubenswrapper[4700]: I0227 17:32:59.662633 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/493471ba-3317-48f3-accc-a2ddb143984c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "493471ba-3317-48f3-accc-a2ddb143984c" (UID: "493471ba-3317-48f3-accc-a2ddb143984c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:32:59 crc kubenswrapper[4700]: I0227 17:32:59.708807 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wxtc\" (UniqueName: \"kubernetes.io/projected/493471ba-3317-48f3-accc-a2ddb143984c-kube-api-access-9wxtc\") on node \"crc\" DevicePath \"\"" Feb 27 17:32:59 crc kubenswrapper[4700]: I0227 17:32:59.708851 4700 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/493471ba-3317-48f3-accc-a2ddb143984c-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 17:32:59 crc kubenswrapper[4700]: I0227 17:32:59.708871 4700 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/493471ba-3317-48f3-accc-a2ddb143984c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 17:32:59 crc kubenswrapper[4700]: I0227 17:32:59.963735 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xndbg" event={"ID":"493471ba-3317-48f3-accc-a2ddb143984c","Type":"ContainerDied","Data":"8759bd966cc0bec3320edc3293fc1bf9ff72a0fb5f06f6ea747ffece2b89f39d"} Feb 27 17:32:59 crc kubenswrapper[4700]: I0227 17:32:59.963839 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xndbg" Feb 27 17:32:59 crc kubenswrapper[4700]: I0227 17:32:59.963858 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8759bd966cc0bec3320edc3293fc1bf9ff72a0fb5f06f6ea747ffece2b89f39d" Feb 27 17:33:00 crc kubenswrapper[4700]: I0227 17:33:00.107052 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgj8b"] Feb 27 17:33:00 crc kubenswrapper[4700]: E0227 17:33:00.107627 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a6c1496-be0e-4733-8176-19c62bbc329e" containerName="oc" Feb 27 17:33:00 crc kubenswrapper[4700]: I0227 17:33:00.107655 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a6c1496-be0e-4733-8176-19c62bbc329e" containerName="oc" Feb 27 17:33:00 crc kubenswrapper[4700]: E0227 17:33:00.107693 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="493471ba-3317-48f3-accc-a2ddb143984c" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 27 17:33:00 crc kubenswrapper[4700]: I0227 17:33:00.107706 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="493471ba-3317-48f3-accc-a2ddb143984c" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 27 17:33:00 crc kubenswrapper[4700]: I0227 17:33:00.108072 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="493471ba-3317-48f3-accc-a2ddb143984c" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 27 17:33:00 crc kubenswrapper[4700]: I0227 17:33:00.108113 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a6c1496-be0e-4733-8176-19c62bbc329e" containerName="oc" Feb 27 17:33:00 crc kubenswrapper[4700]: I0227 17:33:00.109235 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgj8b" Feb 27 17:33:00 crc kubenswrapper[4700]: I0227 17:33:00.122938 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 17:33:00 crc kubenswrapper[4700]: I0227 17:33:00.123004 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dtdm4" Feb 27 17:33:00 crc kubenswrapper[4700]: I0227 17:33:00.123022 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 17:33:00 crc kubenswrapper[4700]: I0227 17:33:00.123443 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 17:33:00 crc kubenswrapper[4700]: I0227 17:33:00.128435 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgj8b"] Feb 27 17:33:00 crc kubenswrapper[4700]: I0227 17:33:00.221838 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ssrl\" (UniqueName: \"kubernetes.io/projected/2855040c-fd3a-4dae-b6c0-626cc01ee321-kube-api-access-9ssrl\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dgj8b\" (UID: \"2855040c-fd3a-4dae-b6c0-626cc01ee321\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgj8b" Feb 27 17:33:00 crc kubenswrapper[4700]: I0227 17:33:00.222084 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2855040c-fd3a-4dae-b6c0-626cc01ee321-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dgj8b\" (UID: \"2855040c-fd3a-4dae-b6c0-626cc01ee321\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgj8b" Feb 27 17:33:00 crc kubenswrapper[4700]: I0227 17:33:00.222360 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2855040c-fd3a-4dae-b6c0-626cc01ee321-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dgj8b\" (UID: \"2855040c-fd3a-4dae-b6c0-626cc01ee321\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgj8b" Feb 27 17:33:00 crc kubenswrapper[4700]: I0227 17:33:00.324423 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2855040c-fd3a-4dae-b6c0-626cc01ee321-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dgj8b\" (UID: \"2855040c-fd3a-4dae-b6c0-626cc01ee321\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgj8b" Feb 27 17:33:00 crc kubenswrapper[4700]: I0227 17:33:00.324594 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2855040c-fd3a-4dae-b6c0-626cc01ee321-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dgj8b\" (UID: \"2855040c-fd3a-4dae-b6c0-626cc01ee321\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgj8b" Feb 27 17:33:00 crc kubenswrapper[4700]: I0227 17:33:00.324805 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ssrl\" (UniqueName: \"kubernetes.io/projected/2855040c-fd3a-4dae-b6c0-626cc01ee321-kube-api-access-9ssrl\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dgj8b\" (UID: \"2855040c-fd3a-4dae-b6c0-626cc01ee321\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgj8b" Feb 27 17:33:00 crc kubenswrapper[4700]: I0227 17:33:00.329440 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2855040c-fd3a-4dae-b6c0-626cc01ee321-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dgj8b\" (UID: \"2855040c-fd3a-4dae-b6c0-626cc01ee321\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgj8b" Feb 27 17:33:00 crc kubenswrapper[4700]: I0227 17:33:00.329685 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2855040c-fd3a-4dae-b6c0-626cc01ee321-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dgj8b\" (UID: \"2855040c-fd3a-4dae-b6c0-626cc01ee321\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgj8b" Feb 27 17:33:00 crc kubenswrapper[4700]: I0227 17:33:00.360054 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ssrl\" (UniqueName: \"kubernetes.io/projected/2855040c-fd3a-4dae-b6c0-626cc01ee321-kube-api-access-9ssrl\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dgj8b\" (UID: \"2855040c-fd3a-4dae-b6c0-626cc01ee321\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgj8b" Feb 27 17:33:00 crc kubenswrapper[4700]: I0227 17:33:00.438006 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgj8b" Feb 27 17:33:00 crc kubenswrapper[4700]: I0227 17:33:00.842420 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgj8b"] Feb 27 17:33:00 crc kubenswrapper[4700]: I0227 17:33:00.995995 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgj8b" event={"ID":"2855040c-fd3a-4dae-b6c0-626cc01ee321","Type":"ContainerStarted","Data":"878e0886c7734411159c73b3f30d85497b97a715068d1871e3827500c5e7f0c6"} Feb 27 17:33:01 crc kubenswrapper[4700]: I0227 17:33:01.992489 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgj8b" event={"ID":"2855040c-fd3a-4dae-b6c0-626cc01ee321","Type":"ContainerStarted","Data":"0b10d8cabacf52c62e3ab6ca88a2a0e93582e1d7e74109bd450b904730aaa6aa"} Feb 27 17:33:02 crc kubenswrapper[4700]: I0227 17:33:02.017598 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgj8b" podStartSLOduration=1.583349985 podStartE2EDuration="2.017566925s" podCreationTimestamp="2026-02-27 17:33:00 +0000 UTC" firstStartedPulling="2026-02-27 17:33:00.848041948 +0000 UTC m=+1940.833354715" lastFinishedPulling="2026-02-27 17:33:01.282258898 +0000 UTC m=+1941.267571655" observedRunningTime="2026-02-27 17:33:02.011970877 +0000 UTC m=+1941.997283654" watchObservedRunningTime="2026-02-27 17:33:02.017566925 +0000 UTC m=+1942.002879702" Feb 27 17:33:06 crc kubenswrapper[4700]: I0227 17:33:06.069718 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-xf6ff"] Feb 27 17:33:06 crc kubenswrapper[4700]: I0227 17:33:06.083037 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-6728-account-create-update-swj84"] Feb 27 17:33:06 crc kubenswrapper[4700]: I0227 17:33:06.093342 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-xf6ff"] Feb 27 17:33:06 crc kubenswrapper[4700]: I0227 17:33:06.101020 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-6728-account-create-update-swj84"] Feb 27 17:33:07 crc kubenswrapper[4700]: I0227 17:33:07.030227 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26947d30-e5b4-4a9f-9a14-c41185ff5a4f" path="/var/lib/kubelet/pods/26947d30-e5b4-4a9f-9a14-c41185ff5a4f/volumes" Feb 27 17:33:07 crc kubenswrapper[4700]: I0227 17:33:07.033220 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b931c6bc-3ec1-493d-ad2e-21008ce1d1f5" path="/var/lib/kubelet/pods/b931c6bc-3ec1-493d-ad2e-21008ce1d1f5/volumes" Feb 27 17:33:07 crc kubenswrapper[4700]: I0227 17:33:07.072686 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-829r9"] Feb 27 17:33:07 crc kubenswrapper[4700]: I0227 17:33:07.074998 4700 generic.go:334] "Generic (PLEG): container finished" podID="2855040c-fd3a-4dae-b6c0-626cc01ee321" containerID="0b10d8cabacf52c62e3ab6ca88a2a0e93582e1d7e74109bd450b904730aaa6aa" exitCode=0 Feb 27 17:33:07 crc kubenswrapper[4700]: I0227 17:33:07.075051 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgj8b" event={"ID":"2855040c-fd3a-4dae-b6c0-626cc01ee321","Type":"ContainerDied","Data":"0b10d8cabacf52c62e3ab6ca88a2a0e93582e1d7e74109bd450b904730aaa6aa"} Feb 27 17:33:07 crc kubenswrapper[4700]: I0227 17:33:07.083421 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-4603-account-create-update-4tz7h"] Feb 27 17:33:07 crc kubenswrapper[4700]: I0227 17:33:07.091894 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-xr75f"] Feb 27 17:33:07 crc kubenswrapper[4700]: I0227 17:33:07.100856 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-xr75f"] Feb 27 17:33:07 crc kubenswrapper[4700]: I0227 17:33:07.113279 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-7178-account-create-update-ttwhv"] Feb 27 17:33:07 crc kubenswrapper[4700]: I0227 17:33:07.124905 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-4603-account-create-update-4tz7h"] Feb 27 17:33:07 crc kubenswrapper[4700]: I0227 17:33:07.133914 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-829r9"] Feb 27 17:33:07 crc kubenswrapper[4700]: I0227 17:33:07.143790 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-7178-account-create-update-ttwhv"] Feb 27 17:33:07 crc kubenswrapper[4700]: I0227 17:33:07.962195 4700 scope.go:117] "RemoveContainer" containerID="0f2f183bfd914f6252dd78997763889b1e5cc74a29e4d9af31c6456aceb88f96" Feb 27 17:33:08 crc kubenswrapper[4700]: I0227 17:33:08.010912 4700 scope.go:117] "RemoveContainer" containerID="eba32c7c133cd06b0c577d3b9b1e8c3f433015ecdecbd878fa857cb650a8df98" Feb 27 17:33:08 crc kubenswrapper[4700]: I0227 17:33:08.107302 4700 scope.go:117] "RemoveContainer" containerID="4bcb91689f3aad19cad5a2592cf44a074600f816d8f0720250702e244a8da0db" Feb 27 17:33:08 crc kubenswrapper[4700]: I0227 17:33:08.176486 4700 scope.go:117] "RemoveContainer" containerID="1995dc0741ace72e33bf0776b3003b5630b6888aacbfc518cefa2974898a9358" Feb 27 17:33:08 crc kubenswrapper[4700]: I0227 17:33:08.236612 4700 scope.go:117] "RemoveContainer" containerID="3caf20a4365af9022601cc805ec79dcb3f0bc006f30ea14208f9a3981739664e" Feb 27 17:33:08 crc kubenswrapper[4700]: I0227 17:33:08.277430 4700 scope.go:117] "RemoveContainer" containerID="7275f5c66342d314054d0067e1ee2df3fa76dc8fe893971d7e1ed2db1f2a95c5" Feb 27 17:33:08 crc kubenswrapper[4700]: I0227 17:33:08.317738 4700 scope.go:117] "RemoveContainer" containerID="6e0eeaaf9cba985e7619f89de8cc8165762e556f8d16817b3a9a8b3b42c601cc" Feb 27 17:33:08 crc kubenswrapper[4700]: I0227 17:33:08.473034 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgj8b" Feb 27 17:33:08 crc kubenswrapper[4700]: I0227 17:33:08.637793 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2855040c-fd3a-4dae-b6c0-626cc01ee321-ssh-key-openstack-edpm-ipam\") pod \"2855040c-fd3a-4dae-b6c0-626cc01ee321\" (UID: \"2855040c-fd3a-4dae-b6c0-626cc01ee321\") " Feb 27 17:33:08 crc kubenswrapper[4700]: I0227 17:33:08.638349 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2855040c-fd3a-4dae-b6c0-626cc01ee321-inventory\") pod \"2855040c-fd3a-4dae-b6c0-626cc01ee321\" (UID: \"2855040c-fd3a-4dae-b6c0-626cc01ee321\") " Feb 27 17:33:08 crc kubenswrapper[4700]: I0227 17:33:08.638441 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ssrl\" (UniqueName: \"kubernetes.io/projected/2855040c-fd3a-4dae-b6c0-626cc01ee321-kube-api-access-9ssrl\") pod \"2855040c-fd3a-4dae-b6c0-626cc01ee321\" (UID: \"2855040c-fd3a-4dae-b6c0-626cc01ee321\") " Feb 27 17:33:08 crc kubenswrapper[4700]: I0227 17:33:08.645613 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2855040c-fd3a-4dae-b6c0-626cc01ee321-kube-api-access-9ssrl" (OuterVolumeSpecName: "kube-api-access-9ssrl") pod "2855040c-fd3a-4dae-b6c0-626cc01ee321" (UID: "2855040c-fd3a-4dae-b6c0-626cc01ee321"). InnerVolumeSpecName "kube-api-access-9ssrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:33:08 crc kubenswrapper[4700]: I0227 17:33:08.671033 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2855040c-fd3a-4dae-b6c0-626cc01ee321-inventory" (OuterVolumeSpecName: "inventory") pod "2855040c-fd3a-4dae-b6c0-626cc01ee321" (UID: "2855040c-fd3a-4dae-b6c0-626cc01ee321"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:33:08 crc kubenswrapper[4700]: I0227 17:33:08.675435 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2855040c-fd3a-4dae-b6c0-626cc01ee321-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "2855040c-fd3a-4dae-b6c0-626cc01ee321" (UID: "2855040c-fd3a-4dae-b6c0-626cc01ee321"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:33:08 crc kubenswrapper[4700]: I0227 17:33:08.741341 4700 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2855040c-fd3a-4dae-b6c0-626cc01ee321-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 17:33:08 crc kubenswrapper[4700]: I0227 17:33:08.741381 4700 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2855040c-fd3a-4dae-b6c0-626cc01ee321-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 17:33:08 crc kubenswrapper[4700]: I0227 17:33:08.741394 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ssrl\" (UniqueName: \"kubernetes.io/projected/2855040c-fd3a-4dae-b6c0-626cc01ee321-kube-api-access-9ssrl\") on node \"crc\" DevicePath \"\"" Feb 27 17:33:08 crc kubenswrapper[4700]: I0227 17:33:08.994635 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04cd0229-5bc9-424e-ad73-9b3f9eb3de8d" path="/var/lib/kubelet/pods/04cd0229-5bc9-424e-ad73-9b3f9eb3de8d/volumes" Feb 27 17:33:08 crc kubenswrapper[4700]: I0227 17:33:08.995422 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05664c1a-c691-4925-a4e6-6fdc12b7998a" path="/var/lib/kubelet/pods/05664c1a-c691-4925-a4e6-6fdc12b7998a/volumes" Feb 27 17:33:08 crc kubenswrapper[4700]: I0227 17:33:08.996018 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c22fb154-217c-462c-a0a6-4b67b5bb11d2" path="/var/lib/kubelet/pods/c22fb154-217c-462c-a0a6-4b67b5bb11d2/volumes" Feb 27 17:33:08 crc kubenswrapper[4700]: I0227 17:33:08.996620 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d646da10-fce6-4960-aedd-6aebd88f075b" path="/var/lib/kubelet/pods/d646da10-fce6-4960-aedd-6aebd88f075b/volumes" Feb 27 17:33:09 crc kubenswrapper[4700]: I0227 17:33:09.112556 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgj8b" event={"ID":"2855040c-fd3a-4dae-b6c0-626cc01ee321","Type":"ContainerDied","Data":"878e0886c7734411159c73b3f30d85497b97a715068d1871e3827500c5e7f0c6"} Feb 27 17:33:09 crc kubenswrapper[4700]: I0227 17:33:09.112598 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="878e0886c7734411159c73b3f30d85497b97a715068d1871e3827500c5e7f0c6" Feb 27 17:33:09 crc kubenswrapper[4700]: I0227 17:33:09.112672 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgj8b" Feb 27 17:33:09 crc kubenswrapper[4700]: I0227 17:33:09.196652 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-r7dmx"] Feb 27 17:33:09 crc kubenswrapper[4700]: E0227 17:33:09.197182 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2855040c-fd3a-4dae-b6c0-626cc01ee321" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 27 17:33:09 crc kubenswrapper[4700]: I0227 17:33:09.197204 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="2855040c-fd3a-4dae-b6c0-626cc01ee321" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 27 17:33:09 crc kubenswrapper[4700]: I0227 17:33:09.198741 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="2855040c-fd3a-4dae-b6c0-626cc01ee321" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 27 17:33:09 crc kubenswrapper[4700]: I0227 17:33:09.199549 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r7dmx" Feb 27 17:33:09 crc kubenswrapper[4700]: I0227 17:33:09.203281 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 17:33:09 crc kubenswrapper[4700]: I0227 17:33:09.203318 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 17:33:09 crc kubenswrapper[4700]: I0227 17:33:09.203472 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 17:33:09 crc kubenswrapper[4700]: I0227 17:33:09.203554 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dtdm4" Feb 27 17:33:09 crc kubenswrapper[4700]: I0227 17:33:09.216501 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-r7dmx"] Feb 27 17:33:09 crc kubenswrapper[4700]: I0227 17:33:09.356231 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bfd90117-4b1d-45ed-9c0b-51a98fd914a4-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-r7dmx\" (UID: \"bfd90117-4b1d-45ed-9c0b-51a98fd914a4\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r7dmx" Feb 27 17:33:09 crc kubenswrapper[4700]: I0227 17:33:09.356553 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sw5br\" (UniqueName: \"kubernetes.io/projected/bfd90117-4b1d-45ed-9c0b-51a98fd914a4-kube-api-access-sw5br\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-r7dmx\" (UID: \"bfd90117-4b1d-45ed-9c0b-51a98fd914a4\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r7dmx" Feb 27 17:33:09 crc kubenswrapper[4700]: I0227 17:33:09.356900 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bfd90117-4b1d-45ed-9c0b-51a98fd914a4-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-r7dmx\" (UID: \"bfd90117-4b1d-45ed-9c0b-51a98fd914a4\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r7dmx" Feb 27 17:33:09 crc kubenswrapper[4700]: I0227 17:33:09.460372 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bfd90117-4b1d-45ed-9c0b-51a98fd914a4-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-r7dmx\" (UID: \"bfd90117-4b1d-45ed-9c0b-51a98fd914a4\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r7dmx" Feb 27 17:33:09 crc kubenswrapper[4700]: I0227 17:33:09.460907 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sw5br\" (UniqueName: \"kubernetes.io/projected/bfd90117-4b1d-45ed-9c0b-51a98fd914a4-kube-api-access-sw5br\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-r7dmx\" (UID: \"bfd90117-4b1d-45ed-9c0b-51a98fd914a4\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r7dmx" Feb 27 17:33:09 crc kubenswrapper[4700]: I0227 17:33:09.461074 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bfd90117-4b1d-45ed-9c0b-51a98fd914a4-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-r7dmx\" (UID: \"bfd90117-4b1d-45ed-9c0b-51a98fd914a4\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r7dmx" Feb 27 17:33:09 crc kubenswrapper[4700]: I0227 17:33:09.466006 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bfd90117-4b1d-45ed-9c0b-51a98fd914a4-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-r7dmx\" (UID: \"bfd90117-4b1d-45ed-9c0b-51a98fd914a4\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r7dmx" Feb 27 17:33:09 crc kubenswrapper[4700]: I0227 17:33:09.469257 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bfd90117-4b1d-45ed-9c0b-51a98fd914a4-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-r7dmx\" (UID: \"bfd90117-4b1d-45ed-9c0b-51a98fd914a4\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r7dmx" Feb 27 17:33:09 crc kubenswrapper[4700]: I0227 17:33:09.493861 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sw5br\" (UniqueName: \"kubernetes.io/projected/bfd90117-4b1d-45ed-9c0b-51a98fd914a4-kube-api-access-sw5br\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-r7dmx\" (UID: \"bfd90117-4b1d-45ed-9c0b-51a98fd914a4\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r7dmx" Feb 27 17:33:09 crc kubenswrapper[4700]: I0227 17:33:09.514822 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r7dmx" Feb 27 17:33:10 crc kubenswrapper[4700]: I0227 17:33:10.117894 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-r7dmx"] Feb 27 17:33:11 crc kubenswrapper[4700]: I0227 17:33:11.137921 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r7dmx" event={"ID":"bfd90117-4b1d-45ed-9c0b-51a98fd914a4","Type":"ContainerStarted","Data":"5ba5df4770955f4af8637ec83b372932fb7e7c254b0d3097017461d59e04712f"} Feb 27 17:33:11 crc kubenswrapper[4700]: I0227 17:33:11.138338 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r7dmx" event={"ID":"bfd90117-4b1d-45ed-9c0b-51a98fd914a4","Type":"ContainerStarted","Data":"58ed414733116f9f08457ad41dcdb38457b9069f58e9da45206b8cefeac8288e"} Feb 27 17:33:11 crc kubenswrapper[4700]: I0227 17:33:11.172349 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r7dmx" podStartSLOduration=1.744745027 podStartE2EDuration="2.172320539s" podCreationTimestamp="2026-02-27 17:33:09 +0000 UTC" firstStartedPulling="2026-02-27 17:33:10.139358263 +0000 UTC m=+1950.124671010" lastFinishedPulling="2026-02-27 17:33:10.566933765 +0000 UTC m=+1950.552246522" observedRunningTime="2026-02-27 17:33:11.160292063 +0000 UTC m=+1951.145604850" watchObservedRunningTime="2026-02-27 17:33:11.172320539 +0000 UTC m=+1951.157633316" Feb 27 17:33:11 crc kubenswrapper[4700]: I0227 17:33:11.980943 4700 scope.go:117] "RemoveContainer" containerID="88adee85741f363c1411884023846cd0aab91e2b1c4dca4a79262d1b15c4192f" Feb 27 17:33:13 crc kubenswrapper[4700]: I0227 17:33:13.169425 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerStarted","Data":"d14a07f9a7a8a0cecf47fa738d13cde159084a95d4b48612e05d84dc5f14de89"} Feb 27 17:33:51 crc kubenswrapper[4700]: I0227 17:33:51.664341 4700 generic.go:334] "Generic (PLEG): container finished" podID="bfd90117-4b1d-45ed-9c0b-51a98fd914a4" containerID="5ba5df4770955f4af8637ec83b372932fb7e7c254b0d3097017461d59e04712f" exitCode=0 Feb 27 17:33:51 crc kubenswrapper[4700]: I0227 17:33:51.664491 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r7dmx" event={"ID":"bfd90117-4b1d-45ed-9c0b-51a98fd914a4","Type":"ContainerDied","Data":"5ba5df4770955f4af8637ec83b372932fb7e7c254b0d3097017461d59e04712f"} Feb 27 17:33:52 crc kubenswrapper[4700]: I0227 17:33:52.067981 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-lf45g"] Feb 27 17:33:52 crc kubenswrapper[4700]: I0227 17:33:52.080867 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-lf45g"] Feb 27 17:33:53 crc kubenswrapper[4700]: I0227 17:33:53.000553 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb8c359-c338-4350-acb2-b302786f082a" path="/var/lib/kubelet/pods/3cb8c359-c338-4350-acb2-b302786f082a/volumes" Feb 27 17:33:53 crc kubenswrapper[4700]: I0227 17:33:53.182144 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r7dmx" Feb 27 17:33:53 crc kubenswrapper[4700]: I0227 17:33:53.271371 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bfd90117-4b1d-45ed-9c0b-51a98fd914a4-ssh-key-openstack-edpm-ipam\") pod \"bfd90117-4b1d-45ed-9c0b-51a98fd914a4\" (UID: \"bfd90117-4b1d-45ed-9c0b-51a98fd914a4\") " Feb 27 17:33:53 crc kubenswrapper[4700]: I0227 17:33:53.271545 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bfd90117-4b1d-45ed-9c0b-51a98fd914a4-inventory\") pod \"bfd90117-4b1d-45ed-9c0b-51a98fd914a4\" (UID: \"bfd90117-4b1d-45ed-9c0b-51a98fd914a4\") " Feb 27 17:33:53 crc kubenswrapper[4700]: I0227 17:33:53.271701 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sw5br\" (UniqueName: \"kubernetes.io/projected/bfd90117-4b1d-45ed-9c0b-51a98fd914a4-kube-api-access-sw5br\") pod \"bfd90117-4b1d-45ed-9c0b-51a98fd914a4\" (UID: \"bfd90117-4b1d-45ed-9c0b-51a98fd914a4\") " Feb 27 17:33:53 crc kubenswrapper[4700]: I0227 17:33:53.281286 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfd90117-4b1d-45ed-9c0b-51a98fd914a4-kube-api-access-sw5br" (OuterVolumeSpecName: "kube-api-access-sw5br") pod "bfd90117-4b1d-45ed-9c0b-51a98fd914a4" (UID: "bfd90117-4b1d-45ed-9c0b-51a98fd914a4"). InnerVolumeSpecName "kube-api-access-sw5br". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:33:53 crc kubenswrapper[4700]: I0227 17:33:53.311136 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfd90117-4b1d-45ed-9c0b-51a98fd914a4-inventory" (OuterVolumeSpecName: "inventory") pod "bfd90117-4b1d-45ed-9c0b-51a98fd914a4" (UID: "bfd90117-4b1d-45ed-9c0b-51a98fd914a4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:33:53 crc kubenswrapper[4700]: I0227 17:33:53.322964 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfd90117-4b1d-45ed-9c0b-51a98fd914a4-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "bfd90117-4b1d-45ed-9c0b-51a98fd914a4" (UID: "bfd90117-4b1d-45ed-9c0b-51a98fd914a4"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:33:53 crc kubenswrapper[4700]: I0227 17:33:53.374065 4700 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bfd90117-4b1d-45ed-9c0b-51a98fd914a4-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 17:33:53 crc kubenswrapper[4700]: I0227 17:33:53.374119 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sw5br\" (UniqueName: \"kubernetes.io/projected/bfd90117-4b1d-45ed-9c0b-51a98fd914a4-kube-api-access-sw5br\") on node \"crc\" DevicePath \"\"" Feb 27 17:33:53 crc kubenswrapper[4700]: I0227 17:33:53.374142 4700 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bfd90117-4b1d-45ed-9c0b-51a98fd914a4-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 17:33:53 crc kubenswrapper[4700]: I0227 17:33:53.692934 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r7dmx" event={"ID":"bfd90117-4b1d-45ed-9c0b-51a98fd914a4","Type":"ContainerDied","Data":"58ed414733116f9f08457ad41dcdb38457b9069f58e9da45206b8cefeac8288e"} Feb 27 17:33:53 crc kubenswrapper[4700]: I0227 17:33:53.693011 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58ed414733116f9f08457ad41dcdb38457b9069f58e9da45206b8cefeac8288e" Feb 27 17:33:53 crc kubenswrapper[4700]: I0227 17:33:53.693039 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r7dmx" Feb 27 17:33:53 crc kubenswrapper[4700]: I0227 17:33:53.812030 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wd8dw"] Feb 27 17:33:53 crc kubenswrapper[4700]: E0227 17:33:53.812614 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfd90117-4b1d-45ed-9c0b-51a98fd914a4" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 27 17:33:53 crc kubenswrapper[4700]: I0227 17:33:53.812645 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfd90117-4b1d-45ed-9c0b-51a98fd914a4" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 27 17:33:53 crc kubenswrapper[4700]: I0227 17:33:53.813013 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfd90117-4b1d-45ed-9c0b-51a98fd914a4" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 27 17:33:53 crc kubenswrapper[4700]: I0227 17:33:53.814099 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wd8dw" Feb 27 17:33:53 crc kubenswrapper[4700]: I0227 17:33:53.817302 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 17:33:53 crc kubenswrapper[4700]: I0227 17:33:53.817523 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 17:33:53 crc kubenswrapper[4700]: I0227 17:33:53.817555 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 17:33:53 crc kubenswrapper[4700]: I0227 17:33:53.818200 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dtdm4" Feb 27 17:33:53 crc kubenswrapper[4700]: I0227 17:33:53.825419 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wd8dw"] Feb 27 17:33:53 crc kubenswrapper[4700]: I0227 17:33:53.987957 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4gqs\" (UniqueName: \"kubernetes.io/projected/d3872381-519a-4292-bf34-fc008f8828c1-kube-api-access-z4gqs\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wd8dw\" (UID: \"d3872381-519a-4292-bf34-fc008f8828c1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wd8dw" Feb 27 17:33:53 crc kubenswrapper[4700]: I0227 17:33:53.988094 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d3872381-519a-4292-bf34-fc008f8828c1-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wd8dw\" (UID: \"d3872381-519a-4292-bf34-fc008f8828c1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wd8dw" Feb 27 17:33:53 crc kubenswrapper[4700]: I0227 17:33:53.988254 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d3872381-519a-4292-bf34-fc008f8828c1-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wd8dw\" (UID: \"d3872381-519a-4292-bf34-fc008f8828c1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wd8dw" Feb 27 17:33:54 crc kubenswrapper[4700]: I0227 17:33:54.090926 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4gqs\" (UniqueName: \"kubernetes.io/projected/d3872381-519a-4292-bf34-fc008f8828c1-kube-api-access-z4gqs\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wd8dw\" (UID: \"d3872381-519a-4292-bf34-fc008f8828c1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wd8dw" Feb 27 17:33:54 crc kubenswrapper[4700]: I0227 17:33:54.091083 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d3872381-519a-4292-bf34-fc008f8828c1-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wd8dw\" (UID: \"d3872381-519a-4292-bf34-fc008f8828c1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wd8dw" Feb 27 17:33:54 crc kubenswrapper[4700]: I0227 17:33:54.091223 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d3872381-519a-4292-bf34-fc008f8828c1-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wd8dw\" (UID: \"d3872381-519a-4292-bf34-fc008f8828c1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wd8dw" Feb 27 17:33:54 crc kubenswrapper[4700]: I0227 17:33:54.099513 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d3872381-519a-4292-bf34-fc008f8828c1-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wd8dw\" (UID: \"d3872381-519a-4292-bf34-fc008f8828c1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wd8dw" Feb 27 17:33:54 crc kubenswrapper[4700]: I0227 17:33:54.115847 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d3872381-519a-4292-bf34-fc008f8828c1-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wd8dw\" (UID: \"d3872381-519a-4292-bf34-fc008f8828c1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wd8dw" Feb 27 17:33:54 crc kubenswrapper[4700]: I0227 17:33:54.123778 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4gqs\" (UniqueName: \"kubernetes.io/projected/d3872381-519a-4292-bf34-fc008f8828c1-kube-api-access-z4gqs\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wd8dw\" (UID: \"d3872381-519a-4292-bf34-fc008f8828c1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wd8dw" Feb 27 17:33:54 crc kubenswrapper[4700]: I0227 17:33:54.148385 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wd8dw" Feb 27 17:33:54 crc kubenswrapper[4700]: I0227 17:33:54.520175 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wd8dw"] Feb 27 17:33:54 crc kubenswrapper[4700]: I0227 17:33:54.707571 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wd8dw" event={"ID":"d3872381-519a-4292-bf34-fc008f8828c1","Type":"ContainerStarted","Data":"7df0d8e9ed5200b7d84dd9787e6c578aa64399380f7a73ddce39b454d4b0ed73"} Feb 27 17:33:55 crc kubenswrapper[4700]: I0227 17:33:55.725192 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wd8dw" event={"ID":"d3872381-519a-4292-bf34-fc008f8828c1","Type":"ContainerStarted","Data":"9c9de8035db26d325a87b793be55c9e2ad99dba47b78bf06dc2d0d1caa44a864"} Feb 27 17:33:55 crc kubenswrapper[4700]: I0227 17:33:55.769786 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wd8dw" podStartSLOduration=2.368759194 podStartE2EDuration="2.769758518s" podCreationTimestamp="2026-02-27 17:33:53 +0000 UTC" firstStartedPulling="2026-02-27 17:33:54.521103328 +0000 UTC m=+1994.506416105" lastFinishedPulling="2026-02-27 17:33:54.922102652 +0000 UTC m=+1994.907415429" observedRunningTime="2026-02-27 17:33:55.753540723 +0000 UTC m=+1995.738853530" watchObservedRunningTime="2026-02-27 17:33:55.769758518 +0000 UTC m=+1995.755071265" Feb 27 17:34:00 crc kubenswrapper[4700]: I0227 17:34:00.140697 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536894-59kkv"] Feb 27 17:34:00 crc kubenswrapper[4700]: I0227 17:34:00.143919 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536894-59kkv" Feb 27 17:34:00 crc kubenswrapper[4700]: I0227 17:34:00.146897 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d62kc" Feb 27 17:34:00 crc kubenswrapper[4700]: I0227 17:34:00.146963 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:34:00 crc kubenswrapper[4700]: I0227 17:34:00.152333 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:34:00 crc kubenswrapper[4700]: I0227 17:34:00.188048 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536894-59kkv"] Feb 27 17:34:00 crc kubenswrapper[4700]: I0227 17:34:00.241253 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6chvw\" (UniqueName: \"kubernetes.io/projected/3e3527f7-bc5c-4321-bc5c-0fe5c8284b4b-kube-api-access-6chvw\") pod \"auto-csr-approver-29536894-59kkv\" (UID: \"3e3527f7-bc5c-4321-bc5c-0fe5c8284b4b\") " pod="openshift-infra/auto-csr-approver-29536894-59kkv" Feb 27 17:34:00 crc kubenswrapper[4700]: I0227 17:34:00.343541 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6chvw\" (UniqueName: \"kubernetes.io/projected/3e3527f7-bc5c-4321-bc5c-0fe5c8284b4b-kube-api-access-6chvw\") pod \"auto-csr-approver-29536894-59kkv\" (UID: \"3e3527f7-bc5c-4321-bc5c-0fe5c8284b4b\") " pod="openshift-infra/auto-csr-approver-29536894-59kkv" Feb 27 17:34:00 crc kubenswrapper[4700]: I0227 17:34:00.375414 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6chvw\" (UniqueName: \"kubernetes.io/projected/3e3527f7-bc5c-4321-bc5c-0fe5c8284b4b-kube-api-access-6chvw\") pod \"auto-csr-approver-29536894-59kkv\" (UID: \"3e3527f7-bc5c-4321-bc5c-0fe5c8284b4b\") " pod="openshift-infra/auto-csr-approver-29536894-59kkv" Feb 27 17:34:00 crc kubenswrapper[4700]: I0227 17:34:00.465307 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536894-59kkv" Feb 27 17:34:00 crc kubenswrapper[4700]: I0227 17:34:00.948668 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536894-59kkv"] Feb 27 17:34:00 crc kubenswrapper[4700]: W0227 17:34:00.961501 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e3527f7_bc5c_4321_bc5c_0fe5c8284b4b.slice/crio-f928be7b6d25b4bc77f87495729536fbf469f4b4aa5e2bfcd973dc4e64516189 WatchSource:0}: Error finding container f928be7b6d25b4bc77f87495729536fbf469f4b4aa5e2bfcd973dc4e64516189: Status 404 returned error can't find the container with id f928be7b6d25b4bc77f87495729536fbf469f4b4aa5e2bfcd973dc4e64516189 Feb 27 17:34:01 crc kubenswrapper[4700]: I0227 17:34:01.811940 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536894-59kkv" event={"ID":"3e3527f7-bc5c-4321-bc5c-0fe5c8284b4b","Type":"ContainerStarted","Data":"f928be7b6d25b4bc77f87495729536fbf469f4b4aa5e2bfcd973dc4e64516189"} Feb 27 17:34:02 crc kubenswrapper[4700]: I0227 17:34:02.824908 4700 generic.go:334] "Generic (PLEG): container finished" podID="3e3527f7-bc5c-4321-bc5c-0fe5c8284b4b" containerID="c681a8e61f5a02078b4b5673c622a07ce6ededd6ccc76440bac388e04d9d87d1" exitCode=0 Feb 27 17:34:02 crc kubenswrapper[4700]: I0227 17:34:02.825109 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536894-59kkv" event={"ID":"3e3527f7-bc5c-4321-bc5c-0fe5c8284b4b","Type":"ContainerDied","Data":"c681a8e61f5a02078b4b5673c622a07ce6ededd6ccc76440bac388e04d9d87d1"} Feb 27 17:34:04 crc kubenswrapper[4700]: I0227 17:34:04.255287 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536894-59kkv" Feb 27 17:34:04 crc kubenswrapper[4700]: I0227 17:34:04.437563 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6chvw\" (UniqueName: \"kubernetes.io/projected/3e3527f7-bc5c-4321-bc5c-0fe5c8284b4b-kube-api-access-6chvw\") pod \"3e3527f7-bc5c-4321-bc5c-0fe5c8284b4b\" (UID: \"3e3527f7-bc5c-4321-bc5c-0fe5c8284b4b\") " Feb 27 17:34:04 crc kubenswrapper[4700]: I0227 17:34:04.446042 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e3527f7-bc5c-4321-bc5c-0fe5c8284b4b-kube-api-access-6chvw" (OuterVolumeSpecName: "kube-api-access-6chvw") pod "3e3527f7-bc5c-4321-bc5c-0fe5c8284b4b" (UID: "3e3527f7-bc5c-4321-bc5c-0fe5c8284b4b"). InnerVolumeSpecName "kube-api-access-6chvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:34:04 crc kubenswrapper[4700]: I0227 17:34:04.540867 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6chvw\" (UniqueName: \"kubernetes.io/projected/3e3527f7-bc5c-4321-bc5c-0fe5c8284b4b-kube-api-access-6chvw\") on node \"crc\" DevicePath \"\"" Feb 27 17:34:04 crc kubenswrapper[4700]: I0227 17:34:04.848029 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536894-59kkv" event={"ID":"3e3527f7-bc5c-4321-bc5c-0fe5c8284b4b","Type":"ContainerDied","Data":"f928be7b6d25b4bc77f87495729536fbf469f4b4aa5e2bfcd973dc4e64516189"} Feb 27 17:34:04 crc kubenswrapper[4700]: I0227 17:34:04.848073 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f928be7b6d25b4bc77f87495729536fbf469f4b4aa5e2bfcd973dc4e64516189" Feb 27 17:34:04 crc kubenswrapper[4700]: I0227 17:34:04.848102 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536894-59kkv" Feb 27 17:34:05 crc kubenswrapper[4700]: I0227 17:34:05.332573 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536888-rhz6k"] Feb 27 17:34:05 crc kubenswrapper[4700]: I0227 17:34:05.343167 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536888-rhz6k"] Feb 27 17:34:07 crc kubenswrapper[4700]: I0227 17:34:07.001581 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49f627a5-028c-4f94-9bb5-122be14fda92" path="/var/lib/kubelet/pods/49f627a5-028c-4f94-9bb5-122be14fda92/volumes" Feb 27 17:34:08 crc kubenswrapper[4700]: I0227 17:34:08.500217 4700 scope.go:117] "RemoveContainer" containerID="117d54daaae734326ad4742ce2bd14a1fb2c2af5276ef54492ec642e6f432eca" Feb 27 17:34:08 crc kubenswrapper[4700]: I0227 17:34:08.573259 4700 scope.go:117] "RemoveContainer" containerID="9e561378214aa9078b85a2a26e03e0b80ebcd6e117cee4356fbc8010ac8b4944" Feb 27 17:34:08 crc kubenswrapper[4700]: I0227 17:34:08.608170 4700 scope.go:117] "RemoveContainer" containerID="dbc2a4d9396c7ae6016cf7a23809f618c11c4d98324dbad5676b18d4b1c1ebb4" Feb 27 17:34:08 crc kubenswrapper[4700]: I0227 17:34:08.682436 4700 scope.go:117] "RemoveContainer" containerID="04297c3017efd2ccd0fcdc65482cf6fd54aa9e25a3275b3de77fd7e7e227baa9" Feb 27 17:34:08 crc kubenswrapper[4700]: I0227 17:34:08.712572 4700 scope.go:117] "RemoveContainer" containerID="428a86c77496a1ef355760ae818554a4470338024f8e15f6a0eecae75188b776" Feb 27 17:34:08 crc kubenswrapper[4700]: I0227 17:34:08.770975 4700 scope.go:117] "RemoveContainer" containerID="f72590160ae6e9739a856399d7858272e98b055deb042e5960a406b9535a918b" Feb 27 17:34:15 crc kubenswrapper[4700]: I0227 17:34:15.051138 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-xc2m6"] Feb 27 17:34:15 crc kubenswrapper[4700]: I0227 17:34:15.065430 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-xc2m6"] Feb 27 17:34:17 crc kubenswrapper[4700]: I0227 17:34:17.000065 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e412b441-8c49-41fe-b795-7c50d4ef09d0" path="/var/lib/kubelet/pods/e412b441-8c49-41fe-b795-7c50d4ef09d0/volumes" Feb 27 17:34:20 crc kubenswrapper[4700]: I0227 17:34:20.026570 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8djkd"] Feb 27 17:34:20 crc kubenswrapper[4700]: I0227 17:34:20.036626 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8djkd"] Feb 27 17:34:20 crc kubenswrapper[4700]: I0227 17:34:20.993383 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c828f754-e6cd-46ff-9c5b-b3db50639317" path="/var/lib/kubelet/pods/c828f754-e6cd-46ff-9c5b-b3db50639317/volumes" Feb 27 17:34:49 crc kubenswrapper[4700]: I0227 17:34:49.366208 4700 generic.go:334] "Generic (PLEG): container finished" podID="d3872381-519a-4292-bf34-fc008f8828c1" containerID="9c9de8035db26d325a87b793be55c9e2ad99dba47b78bf06dc2d0d1caa44a864" exitCode=0 Feb 27 17:34:49 crc kubenswrapper[4700]: I0227 17:34:49.366331 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wd8dw" event={"ID":"d3872381-519a-4292-bf34-fc008f8828c1","Type":"ContainerDied","Data":"9c9de8035db26d325a87b793be55c9e2ad99dba47b78bf06dc2d0d1caa44a864"} Feb 27 17:34:50 crc kubenswrapper[4700]: I0227 17:34:50.893184 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wd8dw" Feb 27 17:34:51 crc kubenswrapper[4700]: I0227 17:34:51.010359 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4gqs\" (UniqueName: \"kubernetes.io/projected/d3872381-519a-4292-bf34-fc008f8828c1-kube-api-access-z4gqs\") pod \"d3872381-519a-4292-bf34-fc008f8828c1\" (UID: \"d3872381-519a-4292-bf34-fc008f8828c1\") " Feb 27 17:34:51 crc kubenswrapper[4700]: I0227 17:34:51.010570 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d3872381-519a-4292-bf34-fc008f8828c1-inventory\") pod \"d3872381-519a-4292-bf34-fc008f8828c1\" (UID: \"d3872381-519a-4292-bf34-fc008f8828c1\") " Feb 27 17:34:51 crc kubenswrapper[4700]: I0227 17:34:51.010605 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d3872381-519a-4292-bf34-fc008f8828c1-ssh-key-openstack-edpm-ipam\") pod \"d3872381-519a-4292-bf34-fc008f8828c1\" (UID: \"d3872381-519a-4292-bf34-fc008f8828c1\") " Feb 27 17:34:51 crc kubenswrapper[4700]: I0227 17:34:51.018211 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3872381-519a-4292-bf34-fc008f8828c1-kube-api-access-z4gqs" (OuterVolumeSpecName: "kube-api-access-z4gqs") pod "d3872381-519a-4292-bf34-fc008f8828c1" (UID: "d3872381-519a-4292-bf34-fc008f8828c1"). InnerVolumeSpecName "kube-api-access-z4gqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:34:51 crc kubenswrapper[4700]: I0227 17:34:51.046057 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3872381-519a-4292-bf34-fc008f8828c1-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "d3872381-519a-4292-bf34-fc008f8828c1" (UID: "d3872381-519a-4292-bf34-fc008f8828c1"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:34:51 crc kubenswrapper[4700]: I0227 17:34:51.055861 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3872381-519a-4292-bf34-fc008f8828c1-inventory" (OuterVolumeSpecName: "inventory") pod "d3872381-519a-4292-bf34-fc008f8828c1" (UID: "d3872381-519a-4292-bf34-fc008f8828c1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:34:51 crc kubenswrapper[4700]: I0227 17:34:51.112271 4700 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d3872381-519a-4292-bf34-fc008f8828c1-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 17:34:51 crc kubenswrapper[4700]: I0227 17:34:51.112304 4700 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d3872381-519a-4292-bf34-fc008f8828c1-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 17:34:51 crc kubenswrapper[4700]: I0227 17:34:51.112313 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4gqs\" (UniqueName: \"kubernetes.io/projected/d3872381-519a-4292-bf34-fc008f8828c1-kube-api-access-z4gqs\") on node \"crc\" DevicePath \"\"" Feb 27 17:34:51 crc kubenswrapper[4700]: I0227 17:34:51.398393 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wd8dw" event={"ID":"d3872381-519a-4292-bf34-fc008f8828c1","Type":"ContainerDied","Data":"7df0d8e9ed5200b7d84dd9787e6c578aa64399380f7a73ddce39b454d4b0ed73"} Feb 27 17:34:51 crc kubenswrapper[4700]: I0227 17:34:51.398443 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7df0d8e9ed5200b7d84dd9787e6c578aa64399380f7a73ddce39b454d4b0ed73" Feb 27 17:34:51 crc kubenswrapper[4700]: I0227 17:34:51.398531 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wd8dw" Feb 27 17:34:51 crc kubenswrapper[4700]: I0227 17:34:51.494639 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-sznqn"] Feb 27 17:34:51 crc kubenswrapper[4700]: E0227 17:34:51.495274 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3872381-519a-4292-bf34-fc008f8828c1" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 27 17:34:51 crc kubenswrapper[4700]: I0227 17:34:51.495306 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3872381-519a-4292-bf34-fc008f8828c1" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 27 17:34:51 crc kubenswrapper[4700]: E0227 17:34:51.495563 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e3527f7-bc5c-4321-bc5c-0fe5c8284b4b" containerName="oc" Feb 27 17:34:51 crc kubenswrapper[4700]: I0227 17:34:51.495577 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e3527f7-bc5c-4321-bc5c-0fe5c8284b4b" containerName="oc" Feb 27 17:34:51 crc kubenswrapper[4700]: I0227 17:34:51.495893 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3872381-519a-4292-bf34-fc008f8828c1" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 27 17:34:51 crc kubenswrapper[4700]: I0227 17:34:51.495953 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e3527f7-bc5c-4321-bc5c-0fe5c8284b4b" containerName="oc" Feb 27 17:34:51 crc kubenswrapper[4700]: I0227 17:34:51.497134 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-sznqn" Feb 27 17:34:51 crc kubenswrapper[4700]: I0227 17:34:51.502012 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 17:34:51 crc kubenswrapper[4700]: I0227 17:34:51.502120 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 17:34:51 crc kubenswrapper[4700]: I0227 17:34:51.503962 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dtdm4" Feb 27 17:34:51 crc kubenswrapper[4700]: I0227 17:34:51.504353 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 17:34:51 crc kubenswrapper[4700]: I0227 17:34:51.504922 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-sznqn"] Feb 27 17:34:51 crc kubenswrapper[4700]: I0227 17:34:51.624487 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57s8x\" (UniqueName: \"kubernetes.io/projected/0115f2bd-d709-44d2-8a30-0c8debd51afa-kube-api-access-57s8x\") pod \"ssh-known-hosts-edpm-deployment-sznqn\" (UID: \"0115f2bd-d709-44d2-8a30-0c8debd51afa\") " pod="openstack/ssh-known-hosts-edpm-deployment-sznqn" Feb 27 17:34:51 crc kubenswrapper[4700]: I0227 17:34:51.624638 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0115f2bd-d709-44d2-8a30-0c8debd51afa-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-sznqn\" (UID: \"0115f2bd-d709-44d2-8a30-0c8debd51afa\") " pod="openstack/ssh-known-hosts-edpm-deployment-sznqn" Feb 27 17:34:51 crc kubenswrapper[4700]: I0227 17:34:51.624737 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/0115f2bd-d709-44d2-8a30-0c8debd51afa-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-sznqn\" (UID: \"0115f2bd-d709-44d2-8a30-0c8debd51afa\") " pod="openstack/ssh-known-hosts-edpm-deployment-sznqn" Feb 27 17:34:51 crc kubenswrapper[4700]: I0227 17:34:51.726957 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0115f2bd-d709-44d2-8a30-0c8debd51afa-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-sznqn\" (UID: \"0115f2bd-d709-44d2-8a30-0c8debd51afa\") " pod="openstack/ssh-known-hosts-edpm-deployment-sznqn" Feb 27 17:34:51 crc kubenswrapper[4700]: I0227 17:34:51.727096 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/0115f2bd-d709-44d2-8a30-0c8debd51afa-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-sznqn\" (UID: \"0115f2bd-d709-44d2-8a30-0c8debd51afa\") " pod="openstack/ssh-known-hosts-edpm-deployment-sznqn" Feb 27 17:34:51 crc kubenswrapper[4700]: I0227 17:34:51.727391 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57s8x\" (UniqueName: \"kubernetes.io/projected/0115f2bd-d709-44d2-8a30-0c8debd51afa-kube-api-access-57s8x\") pod \"ssh-known-hosts-edpm-deployment-sznqn\" (UID: \"0115f2bd-d709-44d2-8a30-0c8debd51afa\") " pod="openstack/ssh-known-hosts-edpm-deployment-sznqn" Feb 27 17:34:51 crc kubenswrapper[4700]: I0227 17:34:51.733035 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/0115f2bd-d709-44d2-8a30-0c8debd51afa-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-sznqn\" (UID: \"0115f2bd-d709-44d2-8a30-0c8debd51afa\") " pod="openstack/ssh-known-hosts-edpm-deployment-sznqn" Feb 27 17:34:51 crc kubenswrapper[4700]: I0227 17:34:51.741769 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0115f2bd-d709-44d2-8a30-0c8debd51afa-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-sznqn\" (UID: \"0115f2bd-d709-44d2-8a30-0c8debd51afa\") " pod="openstack/ssh-known-hosts-edpm-deployment-sznqn" Feb 27 17:34:51 crc kubenswrapper[4700]: I0227 17:34:51.759372 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57s8x\" (UniqueName: \"kubernetes.io/projected/0115f2bd-d709-44d2-8a30-0c8debd51afa-kube-api-access-57s8x\") pod \"ssh-known-hosts-edpm-deployment-sznqn\" (UID: \"0115f2bd-d709-44d2-8a30-0c8debd51afa\") " pod="openstack/ssh-known-hosts-edpm-deployment-sznqn" Feb 27 17:34:51 crc kubenswrapper[4700]: I0227 17:34:51.823223 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-sznqn" Feb 27 17:34:52 crc kubenswrapper[4700]: I0227 17:34:52.473270 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-sznqn"] Feb 27 17:34:52 crc kubenswrapper[4700]: W0227 17:34:52.482149 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0115f2bd_d709_44d2_8a30_0c8debd51afa.slice/crio-992cb7313382ea46f144e20312e87bc6ce149bcbdf04accad23694cdadc5e18d WatchSource:0}: Error finding container 992cb7313382ea46f144e20312e87bc6ce149bcbdf04accad23694cdadc5e18d: Status 404 returned error can't find the container with id 992cb7313382ea46f144e20312e87bc6ce149bcbdf04accad23694cdadc5e18d Feb 27 17:34:53 crc kubenswrapper[4700]: I0227 17:34:53.422031 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-sznqn" event={"ID":"0115f2bd-d709-44d2-8a30-0c8debd51afa","Type":"ContainerStarted","Data":"0952a606c0105dcf932afb01e568979ab7f9fdcf84e4479e51dbd78b4dd99f21"} Feb 27 17:34:53 crc kubenswrapper[4700]: I0227 17:34:53.422537 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-sznqn" event={"ID":"0115f2bd-d709-44d2-8a30-0c8debd51afa","Type":"ContainerStarted","Data":"992cb7313382ea46f144e20312e87bc6ce149bcbdf04accad23694cdadc5e18d"} Feb 27 17:34:53 crc kubenswrapper[4700]: I0227 17:34:53.448964 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-sznqn" podStartSLOduration=2.024549433 podStartE2EDuration="2.44891586s" podCreationTimestamp="2026-02-27 17:34:51 +0000 UTC" firstStartedPulling="2026-02-27 17:34:52.485822607 +0000 UTC m=+2052.471135364" lastFinishedPulling="2026-02-27 17:34:52.910189024 +0000 UTC m=+2052.895501791" observedRunningTime="2026-02-27 17:34:53.441082495 +0000 UTC m=+2053.426395272" watchObservedRunningTime="2026-02-27 17:34:53.44891586 +0000 UTC m=+2053.434228647" Feb 27 17:34:59 crc kubenswrapper[4700]: I0227 17:34:59.056341 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-9s776"] Feb 27 17:34:59 crc kubenswrapper[4700]: I0227 17:34:59.070257 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-9s776"] Feb 27 17:35:00 crc kubenswrapper[4700]: I0227 17:35:00.511971 4700 generic.go:334] "Generic (PLEG): container finished" podID="0115f2bd-d709-44d2-8a30-0c8debd51afa" containerID="0952a606c0105dcf932afb01e568979ab7f9fdcf84e4479e51dbd78b4dd99f21" exitCode=0 Feb 27 17:35:00 crc kubenswrapper[4700]: I0227 17:35:00.512033 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-sznqn" event={"ID":"0115f2bd-d709-44d2-8a30-0c8debd51afa","Type":"ContainerDied","Data":"0952a606c0105dcf932afb01e568979ab7f9fdcf84e4479e51dbd78b4dd99f21"} Feb 27 17:35:01 crc kubenswrapper[4700]: I0227 17:35:01.015122 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5e54324-9f69-4bc9-a6af-02b5ca6445e8" path="/var/lib/kubelet/pods/e5e54324-9f69-4bc9-a6af-02b5ca6445e8/volumes" Feb 27 17:35:02 crc kubenswrapper[4700]: I0227 17:35:02.006744 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-sznqn" Feb 27 17:35:02 crc kubenswrapper[4700]: I0227 17:35:02.064514 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57s8x\" (UniqueName: \"kubernetes.io/projected/0115f2bd-d709-44d2-8a30-0c8debd51afa-kube-api-access-57s8x\") pod \"0115f2bd-d709-44d2-8a30-0c8debd51afa\" (UID: \"0115f2bd-d709-44d2-8a30-0c8debd51afa\") " Feb 27 17:35:02 crc kubenswrapper[4700]: I0227 17:35:02.064801 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0115f2bd-d709-44d2-8a30-0c8debd51afa-ssh-key-openstack-edpm-ipam\") pod \"0115f2bd-d709-44d2-8a30-0c8debd51afa\" (UID: \"0115f2bd-d709-44d2-8a30-0c8debd51afa\") " Feb 27 17:35:02 crc kubenswrapper[4700]: I0227 17:35:02.064925 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/0115f2bd-d709-44d2-8a30-0c8debd51afa-inventory-0\") pod \"0115f2bd-d709-44d2-8a30-0c8debd51afa\" (UID: \"0115f2bd-d709-44d2-8a30-0c8debd51afa\") " Feb 27 17:35:02 crc kubenswrapper[4700]: I0227 17:35:02.074274 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0115f2bd-d709-44d2-8a30-0c8debd51afa-kube-api-access-57s8x" (OuterVolumeSpecName: "kube-api-access-57s8x") pod "0115f2bd-d709-44d2-8a30-0c8debd51afa" (UID: "0115f2bd-d709-44d2-8a30-0c8debd51afa"). InnerVolumeSpecName "kube-api-access-57s8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:35:02 crc kubenswrapper[4700]: I0227 17:35:02.121057 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0115f2bd-d709-44d2-8a30-0c8debd51afa-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "0115f2bd-d709-44d2-8a30-0c8debd51afa" (UID: "0115f2bd-d709-44d2-8a30-0c8debd51afa"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:35:02 crc kubenswrapper[4700]: I0227 17:35:02.127638 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0115f2bd-d709-44d2-8a30-0c8debd51afa-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "0115f2bd-d709-44d2-8a30-0c8debd51afa" (UID: "0115f2bd-d709-44d2-8a30-0c8debd51afa"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:35:02 crc kubenswrapper[4700]: I0227 17:35:02.167170 4700 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0115f2bd-d709-44d2-8a30-0c8debd51afa-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 17:35:02 crc kubenswrapper[4700]: I0227 17:35:02.167210 4700 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/0115f2bd-d709-44d2-8a30-0c8debd51afa-inventory-0\") on node \"crc\" DevicePath \"\"" Feb 27 17:35:02 crc kubenswrapper[4700]: I0227 17:35:02.167225 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57s8x\" (UniqueName: \"kubernetes.io/projected/0115f2bd-d709-44d2-8a30-0c8debd51afa-kube-api-access-57s8x\") on node \"crc\" DevicePath \"\"" Feb 27 17:35:02 crc kubenswrapper[4700]: I0227 17:35:02.538029 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-sznqn" event={"ID":"0115f2bd-d709-44d2-8a30-0c8debd51afa","Type":"ContainerDied","Data":"992cb7313382ea46f144e20312e87bc6ce149bcbdf04accad23694cdadc5e18d"} Feb 27 17:35:02 crc kubenswrapper[4700]: I0227 17:35:02.538115 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="992cb7313382ea46f144e20312e87bc6ce149bcbdf04accad23694cdadc5e18d" Feb 27 17:35:02 crc kubenswrapper[4700]: I0227 17:35:02.538133 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-sznqn" Feb 27 17:35:02 crc kubenswrapper[4700]: I0227 17:35:02.634579 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-dmfwb"] Feb 27 17:35:02 crc kubenswrapper[4700]: E0227 17:35:02.635129 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0115f2bd-d709-44d2-8a30-0c8debd51afa" containerName="ssh-known-hosts-edpm-deployment" Feb 27 17:35:02 crc kubenswrapper[4700]: I0227 17:35:02.635152 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="0115f2bd-d709-44d2-8a30-0c8debd51afa" containerName="ssh-known-hosts-edpm-deployment" Feb 27 17:35:02 crc kubenswrapper[4700]: I0227 17:35:02.635409 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="0115f2bd-d709-44d2-8a30-0c8debd51afa" containerName="ssh-known-hosts-edpm-deployment" Feb 27 17:35:02 crc kubenswrapper[4700]: I0227 17:35:02.636220 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dmfwb" Feb 27 17:35:02 crc kubenswrapper[4700]: I0227 17:35:02.639589 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 17:35:02 crc kubenswrapper[4700]: I0227 17:35:02.639871 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 17:35:02 crc kubenswrapper[4700]: I0227 17:35:02.640200 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 17:35:02 crc kubenswrapper[4700]: I0227 17:35:02.640281 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dtdm4" Feb 27 17:35:02 crc kubenswrapper[4700]: I0227 17:35:02.646984 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-dmfwb"] Feb 27 17:35:02 crc kubenswrapper[4700]: I0227 17:35:02.676353 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6t5wr\" (UniqueName: \"kubernetes.io/projected/7ee8fa48-c17e-43a3-aef8-d5ac737e0b36-kube-api-access-6t5wr\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dmfwb\" (UID: \"7ee8fa48-c17e-43a3-aef8-d5ac737e0b36\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dmfwb" Feb 27 17:35:02 crc kubenswrapper[4700]: I0227 17:35:02.676453 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ee8fa48-c17e-43a3-aef8-d5ac737e0b36-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dmfwb\" (UID: \"7ee8fa48-c17e-43a3-aef8-d5ac737e0b36\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dmfwb" Feb 27 17:35:02 crc kubenswrapper[4700]: I0227 17:35:02.676506 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7ee8fa48-c17e-43a3-aef8-d5ac737e0b36-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dmfwb\" (UID: \"7ee8fa48-c17e-43a3-aef8-d5ac737e0b36\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dmfwb" Feb 27 17:35:02 crc kubenswrapper[4700]: I0227 17:35:02.778118 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ee8fa48-c17e-43a3-aef8-d5ac737e0b36-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dmfwb\" (UID: \"7ee8fa48-c17e-43a3-aef8-d5ac737e0b36\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dmfwb" Feb 27 17:35:02 crc kubenswrapper[4700]: I0227 17:35:02.778215 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7ee8fa48-c17e-43a3-aef8-d5ac737e0b36-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dmfwb\" (UID: \"7ee8fa48-c17e-43a3-aef8-d5ac737e0b36\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dmfwb" Feb 27 17:35:02 crc kubenswrapper[4700]: I0227 17:35:02.778454 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6t5wr\" (UniqueName: \"kubernetes.io/projected/7ee8fa48-c17e-43a3-aef8-d5ac737e0b36-kube-api-access-6t5wr\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dmfwb\" (UID: \"7ee8fa48-c17e-43a3-aef8-d5ac737e0b36\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dmfwb" Feb 27 17:35:02 crc kubenswrapper[4700]: I0227 17:35:02.783912 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7ee8fa48-c17e-43a3-aef8-d5ac737e0b36-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dmfwb\" (UID: \"7ee8fa48-c17e-43a3-aef8-d5ac737e0b36\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dmfwb" Feb 27 17:35:02 crc kubenswrapper[4700]: I0227 17:35:02.784052 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ee8fa48-c17e-43a3-aef8-d5ac737e0b36-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dmfwb\" (UID: \"7ee8fa48-c17e-43a3-aef8-d5ac737e0b36\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dmfwb" Feb 27 17:35:02 crc kubenswrapper[4700]: I0227 17:35:02.806723 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6t5wr\" (UniqueName: \"kubernetes.io/projected/7ee8fa48-c17e-43a3-aef8-d5ac737e0b36-kube-api-access-6t5wr\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dmfwb\" (UID: \"7ee8fa48-c17e-43a3-aef8-d5ac737e0b36\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dmfwb" Feb 27 17:35:02 crc kubenswrapper[4700]: I0227 17:35:02.955973 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dmfwb" Feb 27 17:35:03 crc kubenswrapper[4700]: I0227 17:35:03.550226 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-dmfwb"] Feb 27 17:35:04 crc kubenswrapper[4700]: I0227 17:35:04.567034 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dmfwb" event={"ID":"7ee8fa48-c17e-43a3-aef8-d5ac737e0b36","Type":"ContainerStarted","Data":"07b01c4478c7404615fb6f17a9601caaf93ceffc2c075ac45c85cbbd76bcf3e8"} Feb 27 17:35:04 crc kubenswrapper[4700]: I0227 17:35:04.567724 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dmfwb" event={"ID":"7ee8fa48-c17e-43a3-aef8-d5ac737e0b36","Type":"ContainerStarted","Data":"a41a468f5ca3aa0ddc4bf70ae06be709a33db9abb89e11c0e98956f315d515e0"} Feb 27 17:35:04 crc kubenswrapper[4700]: I0227 17:35:04.600851 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dmfwb" podStartSLOduration=1.961620571 podStartE2EDuration="2.600827372s" podCreationTimestamp="2026-02-27 17:35:02 +0000 UTC" firstStartedPulling="2026-02-27 17:35:03.559951169 +0000 UTC m=+2063.545263926" lastFinishedPulling="2026-02-27 17:35:04.19915794 +0000 UTC m=+2064.184470727" observedRunningTime="2026-02-27 17:35:04.595651556 +0000 UTC m=+2064.580964343" watchObservedRunningTime="2026-02-27 17:35:04.600827372 +0000 UTC m=+2064.586140149" Feb 27 17:35:08 crc kubenswrapper[4700]: I0227 17:35:08.967775 4700 scope.go:117] "RemoveContainer" containerID="8009591760c8c226bb147d170c36329029be33c21271f32af0d70481c9d9cc6e" Feb 27 17:35:09 crc kubenswrapper[4700]: I0227 17:35:09.013593 4700 scope.go:117] "RemoveContainer" containerID="76cd2335a654dbd228865b5835dd72e39a3fe261f701ccccfeed3f61f4f4fbf2" Feb 27 17:35:09 crc kubenswrapper[4700]: I0227 17:35:09.070403 4700 scope.go:117] "RemoveContainer" containerID="90785a0c53b789b80497313e888385c92d2327f8d5a03457ee7ccc9658666729" Feb 27 17:35:13 crc kubenswrapper[4700]: I0227 17:35:13.674090 4700 generic.go:334] "Generic (PLEG): container finished" podID="7ee8fa48-c17e-43a3-aef8-d5ac737e0b36" containerID="07b01c4478c7404615fb6f17a9601caaf93ceffc2c075ac45c85cbbd76bcf3e8" exitCode=0 Feb 27 17:35:13 crc kubenswrapper[4700]: I0227 17:35:13.674212 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dmfwb" event={"ID":"7ee8fa48-c17e-43a3-aef8-d5ac737e0b36","Type":"ContainerDied","Data":"07b01c4478c7404615fb6f17a9601caaf93ceffc2c075ac45c85cbbd76bcf3e8"} Feb 27 17:35:15 crc kubenswrapper[4700]: I0227 17:35:15.152805 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dmfwb" Feb 27 17:35:15 crc kubenswrapper[4700]: I0227 17:35:15.286110 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ee8fa48-c17e-43a3-aef8-d5ac737e0b36-inventory\") pod \"7ee8fa48-c17e-43a3-aef8-d5ac737e0b36\" (UID: \"7ee8fa48-c17e-43a3-aef8-d5ac737e0b36\") " Feb 27 17:35:15 crc kubenswrapper[4700]: I0227 17:35:15.286245 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7ee8fa48-c17e-43a3-aef8-d5ac737e0b36-ssh-key-openstack-edpm-ipam\") pod \"7ee8fa48-c17e-43a3-aef8-d5ac737e0b36\" (UID: \"7ee8fa48-c17e-43a3-aef8-d5ac737e0b36\") " Feb 27 17:35:15 crc kubenswrapper[4700]: I0227 17:35:15.286398 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6t5wr\" (UniqueName: \"kubernetes.io/projected/7ee8fa48-c17e-43a3-aef8-d5ac737e0b36-kube-api-access-6t5wr\") pod \"7ee8fa48-c17e-43a3-aef8-d5ac737e0b36\" (UID: \"7ee8fa48-c17e-43a3-aef8-d5ac737e0b36\") " Feb 27 17:35:15 crc kubenswrapper[4700]: I0227 17:35:15.298831 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ee8fa48-c17e-43a3-aef8-d5ac737e0b36-kube-api-access-6t5wr" (OuterVolumeSpecName: "kube-api-access-6t5wr") pod "7ee8fa48-c17e-43a3-aef8-d5ac737e0b36" (UID: "7ee8fa48-c17e-43a3-aef8-d5ac737e0b36"). InnerVolumeSpecName "kube-api-access-6t5wr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:35:15 crc kubenswrapper[4700]: I0227 17:35:15.323601 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ee8fa48-c17e-43a3-aef8-d5ac737e0b36-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "7ee8fa48-c17e-43a3-aef8-d5ac737e0b36" (UID: "7ee8fa48-c17e-43a3-aef8-d5ac737e0b36"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:35:15 crc kubenswrapper[4700]: I0227 17:35:15.356339 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ee8fa48-c17e-43a3-aef8-d5ac737e0b36-inventory" (OuterVolumeSpecName: "inventory") pod "7ee8fa48-c17e-43a3-aef8-d5ac737e0b36" (UID: "7ee8fa48-c17e-43a3-aef8-d5ac737e0b36"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:35:15 crc kubenswrapper[4700]: I0227 17:35:15.389426 4700 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ee8fa48-c17e-43a3-aef8-d5ac737e0b36-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 17:35:15 crc kubenswrapper[4700]: I0227 17:35:15.389494 4700 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7ee8fa48-c17e-43a3-aef8-d5ac737e0b36-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 17:35:15 crc kubenswrapper[4700]: I0227 17:35:15.389511 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6t5wr\" (UniqueName: \"kubernetes.io/projected/7ee8fa48-c17e-43a3-aef8-d5ac737e0b36-kube-api-access-6t5wr\") on node \"crc\" DevicePath \"\"" Feb 27 17:35:15 crc kubenswrapper[4700]: I0227 17:35:15.699917 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dmfwb" event={"ID":"7ee8fa48-c17e-43a3-aef8-d5ac737e0b36","Type":"ContainerDied","Data":"a41a468f5ca3aa0ddc4bf70ae06be709a33db9abb89e11c0e98956f315d515e0"} Feb 27 17:35:15 crc kubenswrapper[4700]: I0227 17:35:15.700480 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a41a468f5ca3aa0ddc4bf70ae06be709a33db9abb89e11c0e98956f315d515e0" Feb 27 17:35:15 crc kubenswrapper[4700]: I0227 17:35:15.700558 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dmfwb" Feb 27 17:35:15 crc kubenswrapper[4700]: I0227 17:35:15.821668 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-274bw"] Feb 27 17:35:15 crc kubenswrapper[4700]: E0227 17:35:15.822507 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ee8fa48-c17e-43a3-aef8-d5ac737e0b36" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 27 17:35:15 crc kubenswrapper[4700]: I0227 17:35:15.822544 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ee8fa48-c17e-43a3-aef8-d5ac737e0b36" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 27 17:35:15 crc kubenswrapper[4700]: I0227 17:35:15.823037 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ee8fa48-c17e-43a3-aef8-d5ac737e0b36" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 27 17:35:15 crc kubenswrapper[4700]: I0227 17:35:15.824290 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-274bw" Feb 27 17:35:15 crc kubenswrapper[4700]: I0227 17:35:15.827985 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 17:35:15 crc kubenswrapper[4700]: I0227 17:35:15.828094 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 17:35:15 crc kubenswrapper[4700]: I0227 17:35:15.828160 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dtdm4" Feb 27 17:35:15 crc kubenswrapper[4700]: I0227 17:35:15.829554 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 17:35:15 crc kubenswrapper[4700]: I0227 17:35:15.858964 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-274bw"] Feb 27 17:35:16 crc kubenswrapper[4700]: I0227 17:35:16.004714 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvdcj\" (UniqueName: \"kubernetes.io/projected/f3d2123c-c2d1-44d0-90c2-2fa65b9d062c-kube-api-access-nvdcj\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-274bw\" (UID: \"f3d2123c-c2d1-44d0-90c2-2fa65b9d062c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-274bw" Feb 27 17:35:16 crc kubenswrapper[4700]: I0227 17:35:16.004908 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f3d2123c-c2d1-44d0-90c2-2fa65b9d062c-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-274bw\" (UID: \"f3d2123c-c2d1-44d0-90c2-2fa65b9d062c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-274bw" Feb 27 17:35:16 crc kubenswrapper[4700]: I0227 17:35:16.004931 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3d2123c-c2d1-44d0-90c2-2fa65b9d062c-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-274bw\" (UID: \"f3d2123c-c2d1-44d0-90c2-2fa65b9d062c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-274bw" Feb 27 17:35:16 crc kubenswrapper[4700]: I0227 17:35:16.106878 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvdcj\" (UniqueName: \"kubernetes.io/projected/f3d2123c-c2d1-44d0-90c2-2fa65b9d062c-kube-api-access-nvdcj\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-274bw\" (UID: \"f3d2123c-c2d1-44d0-90c2-2fa65b9d062c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-274bw" Feb 27 17:35:16 crc kubenswrapper[4700]: I0227 17:35:16.107114 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f3d2123c-c2d1-44d0-90c2-2fa65b9d062c-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-274bw\" (UID: \"f3d2123c-c2d1-44d0-90c2-2fa65b9d062c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-274bw" Feb 27 17:35:16 crc kubenswrapper[4700]: I0227 17:35:16.107139 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3d2123c-c2d1-44d0-90c2-2fa65b9d062c-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-274bw\" (UID: \"f3d2123c-c2d1-44d0-90c2-2fa65b9d062c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-274bw" Feb 27 17:35:16 crc kubenswrapper[4700]: I0227 17:35:16.112312 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3d2123c-c2d1-44d0-90c2-2fa65b9d062c-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-274bw\" (UID: \"f3d2123c-c2d1-44d0-90c2-2fa65b9d062c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-274bw" Feb 27 17:35:16 crc kubenswrapper[4700]: I0227 17:35:16.126521 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f3d2123c-c2d1-44d0-90c2-2fa65b9d062c-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-274bw\" (UID: \"f3d2123c-c2d1-44d0-90c2-2fa65b9d062c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-274bw" Feb 27 17:35:16 crc kubenswrapper[4700]: I0227 17:35:16.129657 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvdcj\" (UniqueName: \"kubernetes.io/projected/f3d2123c-c2d1-44d0-90c2-2fa65b9d062c-kube-api-access-nvdcj\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-274bw\" (UID: \"f3d2123c-c2d1-44d0-90c2-2fa65b9d062c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-274bw" Feb 27 17:35:16 crc kubenswrapper[4700]: I0227 17:35:16.148065 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-274bw" Feb 27 17:35:16 crc kubenswrapper[4700]: W0227 17:35:16.775000 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf3d2123c_c2d1_44d0_90c2_2fa65b9d062c.slice/crio-57ba5717d356538c8c7d7ff3442fa7392f86820199b9dc30d514a4da7a79d18a WatchSource:0}: Error finding container 57ba5717d356538c8c7d7ff3442fa7392f86820199b9dc30d514a4da7a79d18a: Status 404 returned error can't find the container with id 57ba5717d356538c8c7d7ff3442fa7392f86820199b9dc30d514a4da7a79d18a Feb 27 17:35:16 crc kubenswrapper[4700]: I0227 17:35:16.776177 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-274bw"] Feb 27 17:35:17 crc kubenswrapper[4700]: I0227 17:35:17.728155 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-274bw" event={"ID":"f3d2123c-c2d1-44d0-90c2-2fa65b9d062c","Type":"ContainerStarted","Data":"b0cd81eea5a2cecc0f006ffca10068713c89a1c847984c301e2358a8e443eb5a"} Feb 27 17:35:17 crc kubenswrapper[4700]: I0227 17:35:17.728993 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-274bw" event={"ID":"f3d2123c-c2d1-44d0-90c2-2fa65b9d062c","Type":"ContainerStarted","Data":"57ba5717d356538c8c7d7ff3442fa7392f86820199b9dc30d514a4da7a79d18a"} Feb 27 17:35:17 crc kubenswrapper[4700]: I0227 17:35:17.752721 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-274bw" podStartSLOduration=2.267196975 podStartE2EDuration="2.752695754s" podCreationTimestamp="2026-02-27 17:35:15 +0000 UTC" firstStartedPulling="2026-02-27 17:35:16.77734145 +0000 UTC m=+2076.762654197" lastFinishedPulling="2026-02-27 17:35:17.262840229 +0000 UTC m=+2077.248152976" observedRunningTime="2026-02-27 17:35:17.75102188 +0000 UTC m=+2077.736334627" watchObservedRunningTime="2026-02-27 17:35:17.752695754 +0000 UTC m=+2077.738008531" Feb 27 17:35:27 crc kubenswrapper[4700]: I0227 17:35:27.860760 4700 generic.go:334] "Generic (PLEG): container finished" podID="f3d2123c-c2d1-44d0-90c2-2fa65b9d062c" containerID="b0cd81eea5a2cecc0f006ffca10068713c89a1c847984c301e2358a8e443eb5a" exitCode=0 Feb 27 17:35:27 crc kubenswrapper[4700]: I0227 17:35:27.860933 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-274bw" event={"ID":"f3d2123c-c2d1-44d0-90c2-2fa65b9d062c","Type":"ContainerDied","Data":"b0cd81eea5a2cecc0f006ffca10068713c89a1c847984c301e2358a8e443eb5a"} Feb 27 17:35:29 crc kubenswrapper[4700]: I0227 17:35:29.465390 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-274bw" Feb 27 17:35:29 crc kubenswrapper[4700]: I0227 17:35:29.607680 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f3d2123c-c2d1-44d0-90c2-2fa65b9d062c-ssh-key-openstack-edpm-ipam\") pod \"f3d2123c-c2d1-44d0-90c2-2fa65b9d062c\" (UID: \"f3d2123c-c2d1-44d0-90c2-2fa65b9d062c\") " Feb 27 17:35:29 crc kubenswrapper[4700]: I0227 17:35:29.607834 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3d2123c-c2d1-44d0-90c2-2fa65b9d062c-inventory\") pod \"f3d2123c-c2d1-44d0-90c2-2fa65b9d062c\" (UID: \"f3d2123c-c2d1-44d0-90c2-2fa65b9d062c\") " Feb 27 17:35:29 crc kubenswrapper[4700]: I0227 17:35:29.607957 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvdcj\" (UniqueName: \"kubernetes.io/projected/f3d2123c-c2d1-44d0-90c2-2fa65b9d062c-kube-api-access-nvdcj\") pod \"f3d2123c-c2d1-44d0-90c2-2fa65b9d062c\" (UID: \"f3d2123c-c2d1-44d0-90c2-2fa65b9d062c\") " Feb 27 17:35:29 crc kubenswrapper[4700]: I0227 17:35:29.622636 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3d2123c-c2d1-44d0-90c2-2fa65b9d062c-kube-api-access-nvdcj" (OuterVolumeSpecName: "kube-api-access-nvdcj") pod "f3d2123c-c2d1-44d0-90c2-2fa65b9d062c" (UID: "f3d2123c-c2d1-44d0-90c2-2fa65b9d062c"). InnerVolumeSpecName "kube-api-access-nvdcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:35:29 crc kubenswrapper[4700]: I0227 17:35:29.649534 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3d2123c-c2d1-44d0-90c2-2fa65b9d062c-inventory" (OuterVolumeSpecName: "inventory") pod "f3d2123c-c2d1-44d0-90c2-2fa65b9d062c" (UID: "f3d2123c-c2d1-44d0-90c2-2fa65b9d062c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:35:29 crc kubenswrapper[4700]: I0227 17:35:29.657640 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3d2123c-c2d1-44d0-90c2-2fa65b9d062c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "f3d2123c-c2d1-44d0-90c2-2fa65b9d062c" (UID: "f3d2123c-c2d1-44d0-90c2-2fa65b9d062c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:35:29 crc kubenswrapper[4700]: I0227 17:35:29.710253 4700 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f3d2123c-c2d1-44d0-90c2-2fa65b9d062c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 17:35:29 crc kubenswrapper[4700]: I0227 17:35:29.710292 4700 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3d2123c-c2d1-44d0-90c2-2fa65b9d062c-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 17:35:29 crc kubenswrapper[4700]: I0227 17:35:29.710304 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvdcj\" (UniqueName: \"kubernetes.io/projected/f3d2123c-c2d1-44d0-90c2-2fa65b9d062c-kube-api-access-nvdcj\") on node \"crc\" DevicePath \"\"" Feb 27 17:35:29 crc kubenswrapper[4700]: I0227 17:35:29.895421 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-274bw" event={"ID":"f3d2123c-c2d1-44d0-90c2-2fa65b9d062c","Type":"ContainerDied","Data":"57ba5717d356538c8c7d7ff3442fa7392f86820199b9dc30d514a4da7a79d18a"} Feb 27 17:35:29 crc kubenswrapper[4700]: I0227 17:35:29.895579 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57ba5717d356538c8c7d7ff3442fa7392f86820199b9dc30d514a4da7a79d18a" Feb 27 17:35:29 crc kubenswrapper[4700]: I0227 17:35:29.895666 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-274bw" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.115410 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz"] Feb 27 17:35:30 crc kubenswrapper[4700]: E0227 17:35:30.120405 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3d2123c-c2d1-44d0-90c2-2fa65b9d062c" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.120429 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3d2123c-c2d1-44d0-90c2-2fa65b9d062c" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.120646 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3d2123c-c2d1-44d0-90c2-2fa65b9d062c" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.121370 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.123788 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.124252 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.124430 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.129768 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.129860 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.130045 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.130642 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.149896 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dtdm4" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.150333 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz"] Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.220840 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.220888 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.221017 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.221148 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.221192 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.221283 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.221387 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.221474 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.221559 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.221704 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.221815 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7p4wk\" (UniqueName: \"kubernetes.io/projected/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-kube-api-access-7p4wk\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.221848 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.221877 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.221938 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.324349 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.324420 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.324490 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.324572 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.324638 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7p4wk\" (UniqueName: \"kubernetes.io/projected/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-kube-api-access-7p4wk\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.324667 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.324692 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.324739 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.324769 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.324799 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.324826 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.324873 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.324903 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.324947 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.329423 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.330313 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.330574 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.330814 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.331077 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.331155 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.331458 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.331705 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.332506 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.332698 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.332977 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.333282 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.337381 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.359349 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7p4wk\" (UniqueName: \"kubernetes.io/projected/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-kube-api-access-7p4wk\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:30 crc kubenswrapper[4700]: I0227 17:35:30.463261 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:35:31 crc kubenswrapper[4700]: I0227 17:35:31.075225 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz"] Feb 27 17:35:31 crc kubenswrapper[4700]: I0227 17:35:31.920110 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" event={"ID":"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45","Type":"ContainerStarted","Data":"996000a33600164c46ef13be82fb87f23082025978799daeab4646a3a5e59180"} Feb 27 17:35:31 crc kubenswrapper[4700]: I0227 17:35:31.921096 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" event={"ID":"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45","Type":"ContainerStarted","Data":"4c2c6348d3d8a90ead072e5ea9c3bdcfd26b14cf238a9c84f0f6ed07aaf78589"} Feb 27 17:35:31 crc kubenswrapper[4700]: I0227 17:35:31.960250 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" podStartSLOduration=1.541006054 podStartE2EDuration="1.960220216s" podCreationTimestamp="2026-02-27 17:35:30 +0000 UTC" firstStartedPulling="2026-02-27 17:35:31.084862564 +0000 UTC m=+2091.070175311" lastFinishedPulling="2026-02-27 17:35:31.504076726 +0000 UTC m=+2091.489389473" observedRunningTime="2026-02-27 17:35:31.942972654 +0000 UTC m=+2091.928285401" watchObservedRunningTime="2026-02-27 17:35:31.960220216 +0000 UTC m=+2091.945533003" Feb 27 17:35:36 crc kubenswrapper[4700]: I0227 17:35:36.410664 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:35:36 crc kubenswrapper[4700]: I0227 17:35:36.411339 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:35:43 crc kubenswrapper[4700]: I0227 17:35:43.822316 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-flp9p"] Feb 27 17:35:43 crc kubenswrapper[4700]: I0227 17:35:43.825757 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-flp9p" Feb 27 17:35:43 crc kubenswrapper[4700]: I0227 17:35:43.850869 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-flp9p"] Feb 27 17:35:43 crc kubenswrapper[4700]: I0227 17:35:43.951176 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/788f0db2-fb92-40e2-8338-eff3dff28051-utilities\") pod \"community-operators-flp9p\" (UID: \"788f0db2-fb92-40e2-8338-eff3dff28051\") " pod="openshift-marketplace/community-operators-flp9p" Feb 27 17:35:43 crc kubenswrapper[4700]: I0227 17:35:43.951262 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b48tz\" (UniqueName: \"kubernetes.io/projected/788f0db2-fb92-40e2-8338-eff3dff28051-kube-api-access-b48tz\") pod \"community-operators-flp9p\" (UID: \"788f0db2-fb92-40e2-8338-eff3dff28051\") " pod="openshift-marketplace/community-operators-flp9p" Feb 27 17:35:43 crc kubenswrapper[4700]: I0227 17:35:43.951365 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/788f0db2-fb92-40e2-8338-eff3dff28051-catalog-content\") pod \"community-operators-flp9p\" (UID: \"788f0db2-fb92-40e2-8338-eff3dff28051\") " pod="openshift-marketplace/community-operators-flp9p" Feb 27 17:35:44 crc kubenswrapper[4700]: I0227 17:35:44.052834 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/788f0db2-fb92-40e2-8338-eff3dff28051-catalog-content\") pod \"community-operators-flp9p\" (UID: \"788f0db2-fb92-40e2-8338-eff3dff28051\") " pod="openshift-marketplace/community-operators-flp9p" Feb 27 17:35:44 crc kubenswrapper[4700]: I0227 17:35:44.052909 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/788f0db2-fb92-40e2-8338-eff3dff28051-utilities\") pod \"community-operators-flp9p\" (UID: \"788f0db2-fb92-40e2-8338-eff3dff28051\") " pod="openshift-marketplace/community-operators-flp9p" Feb 27 17:35:44 crc kubenswrapper[4700]: I0227 17:35:44.052992 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b48tz\" (UniqueName: \"kubernetes.io/projected/788f0db2-fb92-40e2-8338-eff3dff28051-kube-api-access-b48tz\") pod \"community-operators-flp9p\" (UID: \"788f0db2-fb92-40e2-8338-eff3dff28051\") " pod="openshift-marketplace/community-operators-flp9p" Feb 27 17:35:44 crc kubenswrapper[4700]: I0227 17:35:44.054077 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/788f0db2-fb92-40e2-8338-eff3dff28051-utilities\") pod \"community-operators-flp9p\" (UID: \"788f0db2-fb92-40e2-8338-eff3dff28051\") " pod="openshift-marketplace/community-operators-flp9p" Feb 27 17:35:44 crc kubenswrapper[4700]: I0227 17:35:44.054451 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/788f0db2-fb92-40e2-8338-eff3dff28051-catalog-content\") pod \"community-operators-flp9p\" (UID: \"788f0db2-fb92-40e2-8338-eff3dff28051\") " pod="openshift-marketplace/community-operators-flp9p" Feb 27 17:35:44 crc kubenswrapper[4700]: I0227 17:35:44.076523 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b48tz\" (UniqueName: \"kubernetes.io/projected/788f0db2-fb92-40e2-8338-eff3dff28051-kube-api-access-b48tz\") pod \"community-operators-flp9p\" (UID: \"788f0db2-fb92-40e2-8338-eff3dff28051\") " pod="openshift-marketplace/community-operators-flp9p" Feb 27 17:35:44 crc kubenswrapper[4700]: I0227 17:35:44.148685 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-flp9p" Feb 27 17:35:44 crc kubenswrapper[4700]: I0227 17:35:44.702239 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-flp9p"] Feb 27 17:35:45 crc kubenswrapper[4700]: I0227 17:35:45.056635 4700 generic.go:334] "Generic (PLEG): container finished" podID="788f0db2-fb92-40e2-8338-eff3dff28051" containerID="046ae5af55c6d4e24405cbeaef81098eb9b1efea2f33760303328e9f065edb6e" exitCode=0 Feb 27 17:35:45 crc kubenswrapper[4700]: I0227 17:35:45.056673 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-flp9p" event={"ID":"788f0db2-fb92-40e2-8338-eff3dff28051","Type":"ContainerDied","Data":"046ae5af55c6d4e24405cbeaef81098eb9b1efea2f33760303328e9f065edb6e"} Feb 27 17:35:45 crc kubenswrapper[4700]: I0227 17:35:45.056695 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-flp9p" event={"ID":"788f0db2-fb92-40e2-8338-eff3dff28051","Type":"ContainerStarted","Data":"262337045d3609b3cac85b8b800c9098838847fc68bf7e4a602fa5c4165faf21"} Feb 27 17:35:46 crc kubenswrapper[4700]: I0227 17:35:46.071559 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-flp9p" event={"ID":"788f0db2-fb92-40e2-8338-eff3dff28051","Type":"ContainerStarted","Data":"0e3a7b2799945dc1580b1c28f8c62ef9c30797ee05165d5ee9f18e53c2475e51"} Feb 27 17:35:47 crc kubenswrapper[4700]: I0227 17:35:47.089007 4700 generic.go:334] "Generic (PLEG): container finished" podID="788f0db2-fb92-40e2-8338-eff3dff28051" containerID="0e3a7b2799945dc1580b1c28f8c62ef9c30797ee05165d5ee9f18e53c2475e51" exitCode=0 Feb 27 17:35:47 crc kubenswrapper[4700]: I0227 17:35:47.089112 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-flp9p" event={"ID":"788f0db2-fb92-40e2-8338-eff3dff28051","Type":"ContainerDied","Data":"0e3a7b2799945dc1580b1c28f8c62ef9c30797ee05165d5ee9f18e53c2475e51"} Feb 27 17:35:48 crc kubenswrapper[4700]: I0227 17:35:48.110605 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-flp9p" event={"ID":"788f0db2-fb92-40e2-8338-eff3dff28051","Type":"ContainerStarted","Data":"79409960b087e99bf93243fa4fad5435856857da540a9afbae90ab9a35bf9f2a"} Feb 27 17:35:48 crc kubenswrapper[4700]: I0227 17:35:48.142997 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-flp9p" podStartSLOduration=2.705490718 podStartE2EDuration="5.142972461s" podCreationTimestamp="2026-02-27 17:35:43 +0000 UTC" firstStartedPulling="2026-02-27 17:35:45.058840223 +0000 UTC m=+2105.044152970" lastFinishedPulling="2026-02-27 17:35:47.496321966 +0000 UTC m=+2107.481634713" observedRunningTime="2026-02-27 17:35:48.134409067 +0000 UTC m=+2108.119721824" watchObservedRunningTime="2026-02-27 17:35:48.142972461 +0000 UTC m=+2108.128285218" Feb 27 17:35:51 crc kubenswrapper[4700]: I0227 17:35:51.776500 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nwwll"] Feb 27 17:35:51 crc kubenswrapper[4700]: I0227 17:35:51.779642 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nwwll" Feb 27 17:35:51 crc kubenswrapper[4700]: I0227 17:35:51.814876 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nwwll"] Feb 27 17:35:51 crc kubenswrapper[4700]: I0227 17:35:51.920631 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73b0e666-5679-4096-9db0-8540f586c138-utilities\") pod \"redhat-operators-nwwll\" (UID: \"73b0e666-5679-4096-9db0-8540f586c138\") " pod="openshift-marketplace/redhat-operators-nwwll" Feb 27 17:35:51 crc kubenswrapper[4700]: I0227 17:35:51.921537 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvxqz\" (UniqueName: \"kubernetes.io/projected/73b0e666-5679-4096-9db0-8540f586c138-kube-api-access-kvxqz\") pod \"redhat-operators-nwwll\" (UID: \"73b0e666-5679-4096-9db0-8540f586c138\") " pod="openshift-marketplace/redhat-operators-nwwll" Feb 27 17:35:51 crc kubenswrapper[4700]: I0227 17:35:51.921770 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73b0e666-5679-4096-9db0-8540f586c138-catalog-content\") pod \"redhat-operators-nwwll\" (UID: \"73b0e666-5679-4096-9db0-8540f586c138\") " pod="openshift-marketplace/redhat-operators-nwwll" Feb 27 17:35:52 crc kubenswrapper[4700]: I0227 17:35:52.023249 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73b0e666-5679-4096-9db0-8540f586c138-catalog-content\") pod \"redhat-operators-nwwll\" (UID: \"73b0e666-5679-4096-9db0-8540f586c138\") " pod="openshift-marketplace/redhat-operators-nwwll" Feb 27 17:35:52 crc kubenswrapper[4700]: I0227 17:35:52.023339 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73b0e666-5679-4096-9db0-8540f586c138-utilities\") pod \"redhat-operators-nwwll\" (UID: \"73b0e666-5679-4096-9db0-8540f586c138\") " pod="openshift-marketplace/redhat-operators-nwwll" Feb 27 17:35:52 crc kubenswrapper[4700]: I0227 17:35:52.023854 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvxqz\" (UniqueName: \"kubernetes.io/projected/73b0e666-5679-4096-9db0-8540f586c138-kube-api-access-kvxqz\") pod \"redhat-operators-nwwll\" (UID: \"73b0e666-5679-4096-9db0-8540f586c138\") " pod="openshift-marketplace/redhat-operators-nwwll" Feb 27 17:35:52 crc kubenswrapper[4700]: I0227 17:35:52.023957 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73b0e666-5679-4096-9db0-8540f586c138-catalog-content\") pod \"redhat-operators-nwwll\" (UID: \"73b0e666-5679-4096-9db0-8540f586c138\") " pod="openshift-marketplace/redhat-operators-nwwll" Feb 27 17:35:52 crc kubenswrapper[4700]: I0227 17:35:52.024067 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73b0e666-5679-4096-9db0-8540f586c138-utilities\") pod \"redhat-operators-nwwll\" (UID: \"73b0e666-5679-4096-9db0-8540f586c138\") " pod="openshift-marketplace/redhat-operators-nwwll" Feb 27 17:35:52 crc kubenswrapper[4700]: I0227 17:35:52.041559 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvxqz\" (UniqueName: \"kubernetes.io/projected/73b0e666-5679-4096-9db0-8540f586c138-kube-api-access-kvxqz\") pod \"redhat-operators-nwwll\" (UID: \"73b0e666-5679-4096-9db0-8540f586c138\") " pod="openshift-marketplace/redhat-operators-nwwll" Feb 27 17:35:52 crc kubenswrapper[4700]: I0227 17:35:52.111417 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nwwll" Feb 27 17:35:52 crc kubenswrapper[4700]: I0227 17:35:52.592554 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nwwll"] Feb 27 17:35:53 crc kubenswrapper[4700]: I0227 17:35:53.169668 4700 generic.go:334] "Generic (PLEG): container finished" podID="73b0e666-5679-4096-9db0-8540f586c138" containerID="21f1a8f04f83670571c07c5055286bcd037091408d7630b666c69577913749ca" exitCode=0 Feb 27 17:35:53 crc kubenswrapper[4700]: I0227 17:35:53.169792 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nwwll" event={"ID":"73b0e666-5679-4096-9db0-8540f586c138","Type":"ContainerDied","Data":"21f1a8f04f83670571c07c5055286bcd037091408d7630b666c69577913749ca"} Feb 27 17:35:53 crc kubenswrapper[4700]: I0227 17:35:53.170005 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nwwll" event={"ID":"73b0e666-5679-4096-9db0-8540f586c138","Type":"ContainerStarted","Data":"ba210275508b8a2f927143a5be4d77b852a80915688940469fabf058ab7844c7"} Feb 27 17:35:54 crc kubenswrapper[4700]: I0227 17:35:54.149824 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-flp9p" Feb 27 17:35:54 crc kubenswrapper[4700]: I0227 17:35:54.150245 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-flp9p" Feb 27 17:35:54 crc kubenswrapper[4700]: I0227 17:35:54.212282 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-flp9p" Feb 27 17:35:54 crc kubenswrapper[4700]: I0227 17:35:54.285989 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-flp9p" Feb 27 17:35:55 crc kubenswrapper[4700]: I0227 17:35:55.196234 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nwwll" event={"ID":"73b0e666-5679-4096-9db0-8540f586c138","Type":"ContainerStarted","Data":"465ad60ee57d0f3937fb31c53ec097b31736164817721ab2a93a911b962ab961"} Feb 27 17:35:56 crc kubenswrapper[4700]: I0227 17:35:56.588904 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-flp9p"] Feb 27 17:35:56 crc kubenswrapper[4700]: I0227 17:35:56.589911 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-flp9p" podUID="788f0db2-fb92-40e2-8338-eff3dff28051" containerName="registry-server" containerID="cri-o://79409960b087e99bf93243fa4fad5435856857da540a9afbae90ab9a35bf9f2a" gracePeriod=2 Feb 27 17:35:57 crc kubenswrapper[4700]: I0227 17:35:57.149711 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-flp9p" Feb 27 17:35:57 crc kubenswrapper[4700]: I0227 17:35:57.215652 4700 generic.go:334] "Generic (PLEG): container finished" podID="788f0db2-fb92-40e2-8338-eff3dff28051" containerID="79409960b087e99bf93243fa4fad5435856857da540a9afbae90ab9a35bf9f2a" exitCode=0 Feb 27 17:35:57 crc kubenswrapper[4700]: I0227 17:35:57.215699 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-flp9p" event={"ID":"788f0db2-fb92-40e2-8338-eff3dff28051","Type":"ContainerDied","Data":"79409960b087e99bf93243fa4fad5435856857da540a9afbae90ab9a35bf9f2a"} Feb 27 17:35:57 crc kubenswrapper[4700]: I0227 17:35:57.215726 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-flp9p" event={"ID":"788f0db2-fb92-40e2-8338-eff3dff28051","Type":"ContainerDied","Data":"262337045d3609b3cac85b8b800c9098838847fc68bf7e4a602fa5c4165faf21"} Feb 27 17:35:57 crc kubenswrapper[4700]: I0227 17:35:57.215746 4700 scope.go:117] "RemoveContainer" containerID="79409960b087e99bf93243fa4fad5435856857da540a9afbae90ab9a35bf9f2a" Feb 27 17:35:57 crc kubenswrapper[4700]: I0227 17:35:57.215771 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-flp9p" Feb 27 17:35:57 crc kubenswrapper[4700]: I0227 17:35:57.237954 4700 scope.go:117] "RemoveContainer" containerID="0e3a7b2799945dc1580b1c28f8c62ef9c30797ee05165d5ee9f18e53c2475e51" Feb 27 17:35:57 crc kubenswrapper[4700]: I0227 17:35:57.250376 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/788f0db2-fb92-40e2-8338-eff3dff28051-utilities\") pod \"788f0db2-fb92-40e2-8338-eff3dff28051\" (UID: \"788f0db2-fb92-40e2-8338-eff3dff28051\") " Feb 27 17:35:57 crc kubenswrapper[4700]: I0227 17:35:57.250737 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b48tz\" (UniqueName: \"kubernetes.io/projected/788f0db2-fb92-40e2-8338-eff3dff28051-kube-api-access-b48tz\") pod \"788f0db2-fb92-40e2-8338-eff3dff28051\" (UID: \"788f0db2-fb92-40e2-8338-eff3dff28051\") " Feb 27 17:35:57 crc kubenswrapper[4700]: I0227 17:35:57.250900 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/788f0db2-fb92-40e2-8338-eff3dff28051-catalog-content\") pod \"788f0db2-fb92-40e2-8338-eff3dff28051\" (UID: \"788f0db2-fb92-40e2-8338-eff3dff28051\") " Feb 27 17:35:57 crc kubenswrapper[4700]: I0227 17:35:57.251370 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/788f0db2-fb92-40e2-8338-eff3dff28051-utilities" (OuterVolumeSpecName: "utilities") pod "788f0db2-fb92-40e2-8338-eff3dff28051" (UID: "788f0db2-fb92-40e2-8338-eff3dff28051"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:35:57 crc kubenswrapper[4700]: I0227 17:35:57.251814 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/788f0db2-fb92-40e2-8338-eff3dff28051-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:35:57 crc kubenswrapper[4700]: I0227 17:35:57.258744 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/788f0db2-fb92-40e2-8338-eff3dff28051-kube-api-access-b48tz" (OuterVolumeSpecName: "kube-api-access-b48tz") pod "788f0db2-fb92-40e2-8338-eff3dff28051" (UID: "788f0db2-fb92-40e2-8338-eff3dff28051"). InnerVolumeSpecName "kube-api-access-b48tz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:35:57 crc kubenswrapper[4700]: I0227 17:35:57.265177 4700 scope.go:117] "RemoveContainer" containerID="046ae5af55c6d4e24405cbeaef81098eb9b1efea2f33760303328e9f065edb6e" Feb 27 17:35:57 crc kubenswrapper[4700]: I0227 17:35:57.308822 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/788f0db2-fb92-40e2-8338-eff3dff28051-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "788f0db2-fb92-40e2-8338-eff3dff28051" (UID: "788f0db2-fb92-40e2-8338-eff3dff28051"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:35:57 crc kubenswrapper[4700]: I0227 17:35:57.346481 4700 scope.go:117] "RemoveContainer" containerID="79409960b087e99bf93243fa4fad5435856857da540a9afbae90ab9a35bf9f2a" Feb 27 17:35:57 crc kubenswrapper[4700]: E0227 17:35:57.347244 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79409960b087e99bf93243fa4fad5435856857da540a9afbae90ab9a35bf9f2a\": container with ID starting with 79409960b087e99bf93243fa4fad5435856857da540a9afbae90ab9a35bf9f2a not found: ID does not exist" containerID="79409960b087e99bf93243fa4fad5435856857da540a9afbae90ab9a35bf9f2a" Feb 27 17:35:57 crc kubenswrapper[4700]: I0227 17:35:57.347283 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79409960b087e99bf93243fa4fad5435856857da540a9afbae90ab9a35bf9f2a"} err="failed to get container status \"79409960b087e99bf93243fa4fad5435856857da540a9afbae90ab9a35bf9f2a\": rpc error: code = NotFound desc = could not find container \"79409960b087e99bf93243fa4fad5435856857da540a9afbae90ab9a35bf9f2a\": container with ID starting with 79409960b087e99bf93243fa4fad5435856857da540a9afbae90ab9a35bf9f2a not found: ID does not exist" Feb 27 17:35:57 crc kubenswrapper[4700]: I0227 17:35:57.347337 4700 scope.go:117] "RemoveContainer" containerID="0e3a7b2799945dc1580b1c28f8c62ef9c30797ee05165d5ee9f18e53c2475e51" Feb 27 17:35:57 crc kubenswrapper[4700]: E0227 17:35:57.347904 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e3a7b2799945dc1580b1c28f8c62ef9c30797ee05165d5ee9f18e53c2475e51\": container with ID starting with 0e3a7b2799945dc1580b1c28f8c62ef9c30797ee05165d5ee9f18e53c2475e51 not found: ID does not exist" containerID="0e3a7b2799945dc1580b1c28f8c62ef9c30797ee05165d5ee9f18e53c2475e51" Feb 27 17:35:57 crc kubenswrapper[4700]: I0227 17:35:57.348155 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e3a7b2799945dc1580b1c28f8c62ef9c30797ee05165d5ee9f18e53c2475e51"} err="failed to get container status \"0e3a7b2799945dc1580b1c28f8c62ef9c30797ee05165d5ee9f18e53c2475e51\": rpc error: code = NotFound desc = could not find container \"0e3a7b2799945dc1580b1c28f8c62ef9c30797ee05165d5ee9f18e53c2475e51\": container with ID starting with 0e3a7b2799945dc1580b1c28f8c62ef9c30797ee05165d5ee9f18e53c2475e51 not found: ID does not exist" Feb 27 17:35:57 crc kubenswrapper[4700]: I0227 17:35:57.348309 4700 scope.go:117] "RemoveContainer" containerID="046ae5af55c6d4e24405cbeaef81098eb9b1efea2f33760303328e9f065edb6e" Feb 27 17:35:57 crc kubenswrapper[4700]: E0227 17:35:57.348961 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"046ae5af55c6d4e24405cbeaef81098eb9b1efea2f33760303328e9f065edb6e\": container with ID starting with 046ae5af55c6d4e24405cbeaef81098eb9b1efea2f33760303328e9f065edb6e not found: ID does not exist" containerID="046ae5af55c6d4e24405cbeaef81098eb9b1efea2f33760303328e9f065edb6e" Feb 27 17:35:57 crc kubenswrapper[4700]: I0227 17:35:57.349011 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"046ae5af55c6d4e24405cbeaef81098eb9b1efea2f33760303328e9f065edb6e"} err="failed to get container status \"046ae5af55c6d4e24405cbeaef81098eb9b1efea2f33760303328e9f065edb6e\": rpc error: code = NotFound desc = could not find container \"046ae5af55c6d4e24405cbeaef81098eb9b1efea2f33760303328e9f065edb6e\": container with ID starting with 046ae5af55c6d4e24405cbeaef81098eb9b1efea2f33760303328e9f065edb6e not found: ID does not exist" Feb 27 17:35:57 crc kubenswrapper[4700]: I0227 17:35:57.354863 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b48tz\" (UniqueName: \"kubernetes.io/projected/788f0db2-fb92-40e2-8338-eff3dff28051-kube-api-access-b48tz\") on node \"crc\" DevicePath \"\"" Feb 27 17:35:57 crc kubenswrapper[4700]: I0227 17:35:57.354911 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/788f0db2-fb92-40e2-8338-eff3dff28051-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:35:57 crc kubenswrapper[4700]: I0227 17:35:57.567578 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-flp9p"] Feb 27 17:35:57 crc kubenswrapper[4700]: I0227 17:35:57.580532 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-flp9p"] Feb 27 17:35:58 crc kubenswrapper[4700]: I0227 17:35:58.229699 4700 generic.go:334] "Generic (PLEG): container finished" podID="73b0e666-5679-4096-9db0-8540f586c138" containerID="465ad60ee57d0f3937fb31c53ec097b31736164817721ab2a93a911b962ab961" exitCode=0 Feb 27 17:35:58 crc kubenswrapper[4700]: I0227 17:35:58.229799 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nwwll" event={"ID":"73b0e666-5679-4096-9db0-8540f586c138","Type":"ContainerDied","Data":"465ad60ee57d0f3937fb31c53ec097b31736164817721ab2a93a911b962ab961"} Feb 27 17:35:59 crc kubenswrapper[4700]: I0227 17:35:59.003485 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="788f0db2-fb92-40e2-8338-eff3dff28051" path="/var/lib/kubelet/pods/788f0db2-fb92-40e2-8338-eff3dff28051/volumes" Feb 27 17:35:59 crc kubenswrapper[4700]: I0227 17:35:59.248508 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nwwll" event={"ID":"73b0e666-5679-4096-9db0-8540f586c138","Type":"ContainerStarted","Data":"52a20123e4cb827703769db1696083204b6d01576f5b1e7cd86ee4e75ba27bcd"} Feb 27 17:35:59 crc kubenswrapper[4700]: I0227 17:35:59.275857 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nwwll" podStartSLOduration=2.513334687 podStartE2EDuration="8.275835044s" podCreationTimestamp="2026-02-27 17:35:51 +0000 UTC" firstStartedPulling="2026-02-27 17:35:53.172677963 +0000 UTC m=+2113.157990711" lastFinishedPulling="2026-02-27 17:35:58.935178311 +0000 UTC m=+2118.920491068" observedRunningTime="2026-02-27 17:35:59.26805072 +0000 UTC m=+2119.253363497" watchObservedRunningTime="2026-02-27 17:35:59.275835044 +0000 UTC m=+2119.261147791" Feb 27 17:36:00 crc kubenswrapper[4700]: I0227 17:36:00.167164 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536896-zjbb8"] Feb 27 17:36:00 crc kubenswrapper[4700]: E0227 17:36:00.167829 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="788f0db2-fb92-40e2-8338-eff3dff28051" containerName="extract-utilities" Feb 27 17:36:00 crc kubenswrapper[4700]: I0227 17:36:00.167859 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="788f0db2-fb92-40e2-8338-eff3dff28051" containerName="extract-utilities" Feb 27 17:36:00 crc kubenswrapper[4700]: E0227 17:36:00.167898 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="788f0db2-fb92-40e2-8338-eff3dff28051" containerName="extract-content" Feb 27 17:36:00 crc kubenswrapper[4700]: I0227 17:36:00.167913 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="788f0db2-fb92-40e2-8338-eff3dff28051" containerName="extract-content" Feb 27 17:36:00 crc kubenswrapper[4700]: E0227 17:36:00.167952 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="788f0db2-fb92-40e2-8338-eff3dff28051" containerName="registry-server" Feb 27 17:36:00 crc kubenswrapper[4700]: I0227 17:36:00.167965 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="788f0db2-fb92-40e2-8338-eff3dff28051" containerName="registry-server" Feb 27 17:36:00 crc kubenswrapper[4700]: I0227 17:36:00.168283 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="788f0db2-fb92-40e2-8338-eff3dff28051" containerName="registry-server" Feb 27 17:36:00 crc kubenswrapper[4700]: I0227 17:36:00.169372 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536896-zjbb8" Feb 27 17:36:00 crc kubenswrapper[4700]: I0227 17:36:00.172504 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d62kc" Feb 27 17:36:00 crc kubenswrapper[4700]: I0227 17:36:00.172777 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:36:00 crc kubenswrapper[4700]: I0227 17:36:00.172971 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:36:00 crc kubenswrapper[4700]: I0227 17:36:00.194314 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536896-zjbb8"] Feb 27 17:36:00 crc kubenswrapper[4700]: I0227 17:36:00.320552 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jwxm\" (UniqueName: \"kubernetes.io/projected/c8daf403-3c61-4f52-9aa0-09b3e86aacd8-kube-api-access-2jwxm\") pod \"auto-csr-approver-29536896-zjbb8\" (UID: \"c8daf403-3c61-4f52-9aa0-09b3e86aacd8\") " pod="openshift-infra/auto-csr-approver-29536896-zjbb8" Feb 27 17:36:00 crc kubenswrapper[4700]: I0227 17:36:00.422240 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jwxm\" (UniqueName: \"kubernetes.io/projected/c8daf403-3c61-4f52-9aa0-09b3e86aacd8-kube-api-access-2jwxm\") pod \"auto-csr-approver-29536896-zjbb8\" (UID: \"c8daf403-3c61-4f52-9aa0-09b3e86aacd8\") " pod="openshift-infra/auto-csr-approver-29536896-zjbb8" Feb 27 17:36:00 crc kubenswrapper[4700]: I0227 17:36:00.445708 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jwxm\" (UniqueName: \"kubernetes.io/projected/c8daf403-3c61-4f52-9aa0-09b3e86aacd8-kube-api-access-2jwxm\") pod \"auto-csr-approver-29536896-zjbb8\" (UID: \"c8daf403-3c61-4f52-9aa0-09b3e86aacd8\") " pod="openshift-infra/auto-csr-approver-29536896-zjbb8" Feb 27 17:36:00 crc kubenswrapper[4700]: I0227 17:36:00.494886 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536896-zjbb8" Feb 27 17:36:00 crc kubenswrapper[4700]: I0227 17:36:00.992750 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536896-zjbb8"] Feb 27 17:36:00 crc kubenswrapper[4700]: W0227 17:36:00.993557 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc8daf403_3c61_4f52_9aa0_09b3e86aacd8.slice/crio-9431db6bed6a13504062de8e3d0eb80370d8babb822646f5146ffd1a6b633161 WatchSource:0}: Error finding container 9431db6bed6a13504062de8e3d0eb80370d8babb822646f5146ffd1a6b633161: Status 404 returned error can't find the container with id 9431db6bed6a13504062de8e3d0eb80370d8babb822646f5146ffd1a6b633161 Feb 27 17:36:01 crc kubenswrapper[4700]: I0227 17:36:01.285295 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536896-zjbb8" event={"ID":"c8daf403-3c61-4f52-9aa0-09b3e86aacd8","Type":"ContainerStarted","Data":"9431db6bed6a13504062de8e3d0eb80370d8babb822646f5146ffd1a6b633161"} Feb 27 17:36:02 crc kubenswrapper[4700]: I0227 17:36:02.111808 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nwwll" Feb 27 17:36:02 crc kubenswrapper[4700]: I0227 17:36:02.113291 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nwwll" Feb 27 17:36:02 crc kubenswrapper[4700]: I0227 17:36:02.297590 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536896-zjbb8" event={"ID":"c8daf403-3c61-4f52-9aa0-09b3e86aacd8","Type":"ContainerStarted","Data":"ae8c6a0f43dbb98d2cdda2193f245b84f860b855e30384c2778b3a181bb538ce"} Feb 27 17:36:02 crc kubenswrapper[4700]: I0227 17:36:02.316647 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536896-zjbb8" podStartSLOduration=1.490694629 podStartE2EDuration="2.316631526s" podCreationTimestamp="2026-02-27 17:36:00 +0000 UTC" firstStartedPulling="2026-02-27 17:36:00.995780212 +0000 UTC m=+2120.981092959" lastFinishedPulling="2026-02-27 17:36:01.821717109 +0000 UTC m=+2121.807029856" observedRunningTime="2026-02-27 17:36:02.314452019 +0000 UTC m=+2122.299764776" watchObservedRunningTime="2026-02-27 17:36:02.316631526 +0000 UTC m=+2122.301944273" Feb 27 17:36:03 crc kubenswrapper[4700]: I0227 17:36:03.168557 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nwwll" podUID="73b0e666-5679-4096-9db0-8540f586c138" containerName="registry-server" probeResult="failure" output=< Feb 27 17:36:03 crc kubenswrapper[4700]: timeout: failed to connect service ":50051" within 1s Feb 27 17:36:03 crc kubenswrapper[4700]: > Feb 27 17:36:03 crc kubenswrapper[4700]: I0227 17:36:03.316472 4700 generic.go:334] "Generic (PLEG): container finished" podID="c8daf403-3c61-4f52-9aa0-09b3e86aacd8" containerID="ae8c6a0f43dbb98d2cdda2193f245b84f860b855e30384c2778b3a181bb538ce" exitCode=0 Feb 27 17:36:03 crc kubenswrapper[4700]: I0227 17:36:03.316512 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536896-zjbb8" event={"ID":"c8daf403-3c61-4f52-9aa0-09b3e86aacd8","Type":"ContainerDied","Data":"ae8c6a0f43dbb98d2cdda2193f245b84f860b855e30384c2778b3a181bb538ce"} Feb 27 17:36:04 crc kubenswrapper[4700]: I0227 17:36:04.712906 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536896-zjbb8" Feb 27 17:36:04 crc kubenswrapper[4700]: I0227 17:36:04.810486 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jwxm\" (UniqueName: \"kubernetes.io/projected/c8daf403-3c61-4f52-9aa0-09b3e86aacd8-kube-api-access-2jwxm\") pod \"c8daf403-3c61-4f52-9aa0-09b3e86aacd8\" (UID: \"c8daf403-3c61-4f52-9aa0-09b3e86aacd8\") " Feb 27 17:36:04 crc kubenswrapper[4700]: I0227 17:36:04.821906 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8daf403-3c61-4f52-9aa0-09b3e86aacd8-kube-api-access-2jwxm" (OuterVolumeSpecName: "kube-api-access-2jwxm") pod "c8daf403-3c61-4f52-9aa0-09b3e86aacd8" (UID: "c8daf403-3c61-4f52-9aa0-09b3e86aacd8"). InnerVolumeSpecName "kube-api-access-2jwxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:36:04 crc kubenswrapper[4700]: I0227 17:36:04.912927 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jwxm\" (UniqueName: \"kubernetes.io/projected/c8daf403-3c61-4f52-9aa0-09b3e86aacd8-kube-api-access-2jwxm\") on node \"crc\" DevicePath \"\"" Feb 27 17:36:05 crc kubenswrapper[4700]: I0227 17:36:05.347726 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536896-zjbb8" event={"ID":"c8daf403-3c61-4f52-9aa0-09b3e86aacd8","Type":"ContainerDied","Data":"9431db6bed6a13504062de8e3d0eb80370d8babb822646f5146ffd1a6b633161"} Feb 27 17:36:05 crc kubenswrapper[4700]: I0227 17:36:05.347766 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9431db6bed6a13504062de8e3d0eb80370d8babb822646f5146ffd1a6b633161" Feb 27 17:36:05 crc kubenswrapper[4700]: I0227 17:36:05.347819 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536896-zjbb8" Feb 27 17:36:05 crc kubenswrapper[4700]: I0227 17:36:05.428146 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536890-p9c6t"] Feb 27 17:36:05 crc kubenswrapper[4700]: I0227 17:36:05.435807 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536890-p9c6t"] Feb 27 17:36:06 crc kubenswrapper[4700]: I0227 17:36:06.410418 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:36:06 crc kubenswrapper[4700]: I0227 17:36:06.410486 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:36:06 crc kubenswrapper[4700]: I0227 17:36:06.999217 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b6d9b14-d4a9-4633-bba4-197eac2766a4" path="/var/lib/kubelet/pods/1b6d9b14-d4a9-4633-bba4-197eac2766a4/volumes" Feb 27 17:36:09 crc kubenswrapper[4700]: I0227 17:36:09.226648 4700 scope.go:117] "RemoveContainer" containerID="2bb6222ce1bf85664559dcaf0292653e231e4d9e5541158eb0296b51394c0ed4" Feb 27 17:36:12 crc kubenswrapper[4700]: I0227 17:36:12.166817 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nwwll" Feb 27 17:36:12 crc kubenswrapper[4700]: I0227 17:36:12.236129 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nwwll" Feb 27 17:36:12 crc kubenswrapper[4700]: I0227 17:36:12.402598 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nwwll"] Feb 27 17:36:13 crc kubenswrapper[4700]: I0227 17:36:13.449454 4700 generic.go:334] "Generic (PLEG): container finished" podID="8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45" containerID="996000a33600164c46ef13be82fb87f23082025978799daeab4646a3a5e59180" exitCode=0 Feb 27 17:36:13 crc kubenswrapper[4700]: I0227 17:36:13.449524 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" event={"ID":"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45","Type":"ContainerDied","Data":"996000a33600164c46ef13be82fb87f23082025978799daeab4646a3a5e59180"} Feb 27 17:36:13 crc kubenswrapper[4700]: I0227 17:36:13.450600 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nwwll" podUID="73b0e666-5679-4096-9db0-8540f586c138" containerName="registry-server" containerID="cri-o://52a20123e4cb827703769db1696083204b6d01576f5b1e7cd86ee4e75ba27bcd" gracePeriod=2 Feb 27 17:36:14 crc kubenswrapper[4700]: I0227 17:36:14.133134 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nwwll" Feb 27 17:36:14 crc kubenswrapper[4700]: I0227 17:36:14.228484 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73b0e666-5679-4096-9db0-8540f586c138-utilities\") pod \"73b0e666-5679-4096-9db0-8540f586c138\" (UID: \"73b0e666-5679-4096-9db0-8540f586c138\") " Feb 27 17:36:14 crc kubenswrapper[4700]: I0227 17:36:14.228613 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73b0e666-5679-4096-9db0-8540f586c138-catalog-content\") pod \"73b0e666-5679-4096-9db0-8540f586c138\" (UID: \"73b0e666-5679-4096-9db0-8540f586c138\") " Feb 27 17:36:14 crc kubenswrapper[4700]: I0227 17:36:14.228707 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvxqz\" (UniqueName: \"kubernetes.io/projected/73b0e666-5679-4096-9db0-8540f586c138-kube-api-access-kvxqz\") pod \"73b0e666-5679-4096-9db0-8540f586c138\" (UID: \"73b0e666-5679-4096-9db0-8540f586c138\") " Feb 27 17:36:14 crc kubenswrapper[4700]: I0227 17:36:14.229625 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73b0e666-5679-4096-9db0-8540f586c138-utilities" (OuterVolumeSpecName: "utilities") pod "73b0e666-5679-4096-9db0-8540f586c138" (UID: "73b0e666-5679-4096-9db0-8540f586c138"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:36:14 crc kubenswrapper[4700]: I0227 17:36:14.238913 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73b0e666-5679-4096-9db0-8540f586c138-kube-api-access-kvxqz" (OuterVolumeSpecName: "kube-api-access-kvxqz") pod "73b0e666-5679-4096-9db0-8540f586c138" (UID: "73b0e666-5679-4096-9db0-8540f586c138"). InnerVolumeSpecName "kube-api-access-kvxqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:36:14 crc kubenswrapper[4700]: I0227 17:36:14.331027 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73b0e666-5679-4096-9db0-8540f586c138-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:36:14 crc kubenswrapper[4700]: I0227 17:36:14.331069 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvxqz\" (UniqueName: \"kubernetes.io/projected/73b0e666-5679-4096-9db0-8540f586c138-kube-api-access-kvxqz\") on node \"crc\" DevicePath \"\"" Feb 27 17:36:14 crc kubenswrapper[4700]: I0227 17:36:14.352687 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73b0e666-5679-4096-9db0-8540f586c138-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "73b0e666-5679-4096-9db0-8540f586c138" (UID: "73b0e666-5679-4096-9db0-8540f586c138"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:36:14 crc kubenswrapper[4700]: I0227 17:36:14.433183 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73b0e666-5679-4096-9db0-8540f586c138-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:36:14 crc kubenswrapper[4700]: I0227 17:36:14.468362 4700 generic.go:334] "Generic (PLEG): container finished" podID="73b0e666-5679-4096-9db0-8540f586c138" containerID="52a20123e4cb827703769db1696083204b6d01576f5b1e7cd86ee4e75ba27bcd" exitCode=0 Feb 27 17:36:14 crc kubenswrapper[4700]: I0227 17:36:14.468436 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nwwll" Feb 27 17:36:14 crc kubenswrapper[4700]: I0227 17:36:14.468828 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nwwll" event={"ID":"73b0e666-5679-4096-9db0-8540f586c138","Type":"ContainerDied","Data":"52a20123e4cb827703769db1696083204b6d01576f5b1e7cd86ee4e75ba27bcd"} Feb 27 17:36:14 crc kubenswrapper[4700]: I0227 17:36:14.468887 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nwwll" event={"ID":"73b0e666-5679-4096-9db0-8540f586c138","Type":"ContainerDied","Data":"ba210275508b8a2f927143a5be4d77b852a80915688940469fabf058ab7844c7"} Feb 27 17:36:14 crc kubenswrapper[4700]: I0227 17:36:14.468923 4700 scope.go:117] "RemoveContainer" containerID="52a20123e4cb827703769db1696083204b6d01576f5b1e7cd86ee4e75ba27bcd" Feb 27 17:36:14 crc kubenswrapper[4700]: I0227 17:36:14.527440 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nwwll"] Feb 27 17:36:14 crc kubenswrapper[4700]: I0227 17:36:14.532986 4700 scope.go:117] "RemoveContainer" containerID="465ad60ee57d0f3937fb31c53ec097b31736164817721ab2a93a911b962ab961" Feb 27 17:36:14 crc kubenswrapper[4700]: I0227 17:36:14.538402 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nwwll"] Feb 27 17:36:14 crc kubenswrapper[4700]: I0227 17:36:14.582262 4700 scope.go:117] "RemoveContainer" containerID="21f1a8f04f83670571c07c5055286bcd037091408d7630b666c69577913749ca" Feb 27 17:36:14 crc kubenswrapper[4700]: I0227 17:36:14.631052 4700 scope.go:117] "RemoveContainer" containerID="52a20123e4cb827703769db1696083204b6d01576f5b1e7cd86ee4e75ba27bcd" Feb 27 17:36:14 crc kubenswrapper[4700]: E0227 17:36:14.631441 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52a20123e4cb827703769db1696083204b6d01576f5b1e7cd86ee4e75ba27bcd\": container with ID starting with 52a20123e4cb827703769db1696083204b6d01576f5b1e7cd86ee4e75ba27bcd not found: ID does not exist" containerID="52a20123e4cb827703769db1696083204b6d01576f5b1e7cd86ee4e75ba27bcd" Feb 27 17:36:14 crc kubenswrapper[4700]: I0227 17:36:14.631494 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52a20123e4cb827703769db1696083204b6d01576f5b1e7cd86ee4e75ba27bcd"} err="failed to get container status \"52a20123e4cb827703769db1696083204b6d01576f5b1e7cd86ee4e75ba27bcd\": rpc error: code = NotFound desc = could not find container \"52a20123e4cb827703769db1696083204b6d01576f5b1e7cd86ee4e75ba27bcd\": container with ID starting with 52a20123e4cb827703769db1696083204b6d01576f5b1e7cd86ee4e75ba27bcd not found: ID does not exist" Feb 27 17:36:14 crc kubenswrapper[4700]: I0227 17:36:14.631521 4700 scope.go:117] "RemoveContainer" containerID="465ad60ee57d0f3937fb31c53ec097b31736164817721ab2a93a911b962ab961" Feb 27 17:36:14 crc kubenswrapper[4700]: E0227 17:36:14.631747 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"465ad60ee57d0f3937fb31c53ec097b31736164817721ab2a93a911b962ab961\": container with ID starting with 465ad60ee57d0f3937fb31c53ec097b31736164817721ab2a93a911b962ab961 not found: ID does not exist" containerID="465ad60ee57d0f3937fb31c53ec097b31736164817721ab2a93a911b962ab961" Feb 27 17:36:14 crc kubenswrapper[4700]: I0227 17:36:14.631770 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"465ad60ee57d0f3937fb31c53ec097b31736164817721ab2a93a911b962ab961"} err="failed to get container status \"465ad60ee57d0f3937fb31c53ec097b31736164817721ab2a93a911b962ab961\": rpc error: code = NotFound desc = could not find container \"465ad60ee57d0f3937fb31c53ec097b31736164817721ab2a93a911b962ab961\": container with ID starting with 465ad60ee57d0f3937fb31c53ec097b31736164817721ab2a93a911b962ab961 not found: ID does not exist" Feb 27 17:36:14 crc kubenswrapper[4700]: I0227 17:36:14.631786 4700 scope.go:117] "RemoveContainer" containerID="21f1a8f04f83670571c07c5055286bcd037091408d7630b666c69577913749ca" Feb 27 17:36:14 crc kubenswrapper[4700]: E0227 17:36:14.631991 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21f1a8f04f83670571c07c5055286bcd037091408d7630b666c69577913749ca\": container with ID starting with 21f1a8f04f83670571c07c5055286bcd037091408d7630b666c69577913749ca not found: ID does not exist" containerID="21f1a8f04f83670571c07c5055286bcd037091408d7630b666c69577913749ca" Feb 27 17:36:14 crc kubenswrapper[4700]: I0227 17:36:14.632012 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21f1a8f04f83670571c07c5055286bcd037091408d7630b666c69577913749ca"} err="failed to get container status \"21f1a8f04f83670571c07c5055286bcd037091408d7630b666c69577913749ca\": rpc error: code = NotFound desc = could not find container \"21f1a8f04f83670571c07c5055286bcd037091408d7630b666c69577913749ca\": container with ID starting with 21f1a8f04f83670571c07c5055286bcd037091408d7630b666c69577913749ca not found: ID does not exist" Feb 27 17:36:14 crc kubenswrapper[4700]: I0227 17:36:14.991518 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73b0e666-5679-4096-9db0-8540f586c138" path="/var/lib/kubelet/pods/73b0e666-5679-4096-9db0-8540f586c138/volumes" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.021811 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.148508 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-bootstrap-combined-ca-bundle\") pod \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.149054 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.149098 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-neutron-metadata-combined-ca-bundle\") pod \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.149119 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7p4wk\" (UniqueName: \"kubernetes.io/projected/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-kube-api-access-7p4wk\") pod \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.149150 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-repo-setup-combined-ca-bundle\") pod \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.149196 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-openstack-edpm-ipam-ovn-default-certs-0\") pod \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.149225 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.149254 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-nova-combined-ca-bundle\") pod \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.149316 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-ssh-key-openstack-edpm-ipam\") pod \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.149358 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-inventory\") pod \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.149383 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.149428 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-ovn-combined-ca-bundle\") pod \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.149487 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-telemetry-combined-ca-bundle\") pod \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.149570 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-libvirt-combined-ca-bundle\") pod \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\" (UID: \"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45\") " Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.157266 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-kube-api-access-7p4wk" (OuterVolumeSpecName: "kube-api-access-7p4wk") pod "8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45" (UID: "8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45"). InnerVolumeSpecName "kube-api-access-7p4wk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.157623 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45" (UID: "8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.158427 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45" (UID: "8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.158647 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45" (UID: "8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.160285 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45" (UID: "8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.162138 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45" (UID: "8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.162622 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45" (UID: "8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.162962 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45" (UID: "8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.164012 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45" (UID: "8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.164039 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45" (UID: "8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.165375 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45" (UID: "8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.167138 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45" (UID: "8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.204592 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45" (UID: "8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.206453 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-inventory" (OuterVolumeSpecName: "inventory") pod "8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45" (UID: "8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.252971 4700 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.253039 4700 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.253136 4700 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.253251 4700 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.253287 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7p4wk\" (UniqueName: \"kubernetes.io/projected/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-kube-api-access-7p4wk\") on node \"crc\" DevicePath \"\"" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.253314 4700 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.253342 4700 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.253372 4700 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.253402 4700 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.253429 4700 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.253498 4700 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.253530 4700 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.253561 4700 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.253590 4700 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.485999 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" event={"ID":"8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45","Type":"ContainerDied","Data":"4c2c6348d3d8a90ead072e5ea9c3bdcfd26b14cf238a9c84f0f6ed07aaf78589"} Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.486039 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c2c6348d3d8a90ead072e5ea9c3bdcfd26b14cf238a9c84f0f6ed07aaf78589" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.486103 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.632124 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-kf8kd"] Feb 27 17:36:15 crc kubenswrapper[4700]: E0227 17:36:15.632619 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73b0e666-5679-4096-9db0-8540f586c138" containerName="extract-content" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.632643 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="73b0e666-5679-4096-9db0-8540f586c138" containerName="extract-content" Feb 27 17:36:15 crc kubenswrapper[4700]: E0227 17:36:15.632675 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8daf403-3c61-4f52-9aa0-09b3e86aacd8" containerName="oc" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.632684 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8daf403-3c61-4f52-9aa0-09b3e86aacd8" containerName="oc" Feb 27 17:36:15 crc kubenswrapper[4700]: E0227 17:36:15.632705 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.632717 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 27 17:36:15 crc kubenswrapper[4700]: E0227 17:36:15.632737 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73b0e666-5679-4096-9db0-8540f586c138" containerName="registry-server" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.632744 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="73b0e666-5679-4096-9db0-8540f586c138" containerName="registry-server" Feb 27 17:36:15 crc kubenswrapper[4700]: E0227 17:36:15.632760 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73b0e666-5679-4096-9db0-8540f586c138" containerName="extract-utilities" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.632769 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="73b0e666-5679-4096-9db0-8540f586c138" containerName="extract-utilities" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.632994 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8daf403-3c61-4f52-9aa0-09b3e86aacd8" containerName="oc" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.633015 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="73b0e666-5679-4096-9db0-8540f586c138" containerName="registry-server" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.633026 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.647044 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-kf8kd"] Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.647200 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kf8kd" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.666525 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.667223 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.667361 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.667544 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dtdm4" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.667663 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.770810 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e50903c5-4b11-4c01-b14f-7f0e51be5319-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kf8kd\" (UID: \"e50903c5-4b11-4c01-b14f-7f0e51be5319\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kf8kd" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.770934 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d76jq\" (UniqueName: \"kubernetes.io/projected/e50903c5-4b11-4c01-b14f-7f0e51be5319-kube-api-access-d76jq\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kf8kd\" (UID: \"e50903c5-4b11-4c01-b14f-7f0e51be5319\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kf8kd" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.771330 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e50903c5-4b11-4c01-b14f-7f0e51be5319-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kf8kd\" (UID: \"e50903c5-4b11-4c01-b14f-7f0e51be5319\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kf8kd" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.771407 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e50903c5-4b11-4c01-b14f-7f0e51be5319-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kf8kd\" (UID: \"e50903c5-4b11-4c01-b14f-7f0e51be5319\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kf8kd" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.771766 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e50903c5-4b11-4c01-b14f-7f0e51be5319-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kf8kd\" (UID: \"e50903c5-4b11-4c01-b14f-7f0e51be5319\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kf8kd" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.874261 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d76jq\" (UniqueName: \"kubernetes.io/projected/e50903c5-4b11-4c01-b14f-7f0e51be5319-kube-api-access-d76jq\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kf8kd\" (UID: \"e50903c5-4b11-4c01-b14f-7f0e51be5319\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kf8kd" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.874481 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e50903c5-4b11-4c01-b14f-7f0e51be5319-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kf8kd\" (UID: \"e50903c5-4b11-4c01-b14f-7f0e51be5319\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kf8kd" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.874516 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e50903c5-4b11-4c01-b14f-7f0e51be5319-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kf8kd\" (UID: \"e50903c5-4b11-4c01-b14f-7f0e51be5319\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kf8kd" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.874587 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e50903c5-4b11-4c01-b14f-7f0e51be5319-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kf8kd\" (UID: \"e50903c5-4b11-4c01-b14f-7f0e51be5319\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kf8kd" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.874639 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e50903c5-4b11-4c01-b14f-7f0e51be5319-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kf8kd\" (UID: \"e50903c5-4b11-4c01-b14f-7f0e51be5319\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kf8kd" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.876179 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e50903c5-4b11-4c01-b14f-7f0e51be5319-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kf8kd\" (UID: \"e50903c5-4b11-4c01-b14f-7f0e51be5319\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kf8kd" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.879002 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e50903c5-4b11-4c01-b14f-7f0e51be5319-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kf8kd\" (UID: \"e50903c5-4b11-4c01-b14f-7f0e51be5319\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kf8kd" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.879880 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e50903c5-4b11-4c01-b14f-7f0e51be5319-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kf8kd\" (UID: \"e50903c5-4b11-4c01-b14f-7f0e51be5319\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kf8kd" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.880771 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e50903c5-4b11-4c01-b14f-7f0e51be5319-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kf8kd\" (UID: \"e50903c5-4b11-4c01-b14f-7f0e51be5319\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kf8kd" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.905403 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d76jq\" (UniqueName: \"kubernetes.io/projected/e50903c5-4b11-4c01-b14f-7f0e51be5319-kube-api-access-d76jq\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kf8kd\" (UID: \"e50903c5-4b11-4c01-b14f-7f0e51be5319\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kf8kd" Feb 27 17:36:15 crc kubenswrapper[4700]: I0227 17:36:15.976967 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kf8kd" Feb 27 17:36:16 crc kubenswrapper[4700]: I0227 17:36:16.556866 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-kf8kd"] Feb 27 17:36:17 crc kubenswrapper[4700]: I0227 17:36:17.526986 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kf8kd" event={"ID":"e50903c5-4b11-4c01-b14f-7f0e51be5319","Type":"ContainerStarted","Data":"71f08a4f6d54dd94c7595ce865ec0f4c94cf700a807bebfaa8fac16fe39e21a8"} Feb 27 17:36:17 crc kubenswrapper[4700]: I0227 17:36:17.527677 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kf8kd" event={"ID":"e50903c5-4b11-4c01-b14f-7f0e51be5319","Type":"ContainerStarted","Data":"5ad40fcc4feaeef8cf9274b650002597fef1738eb1266eeffd1cd2f9b04a4f06"} Feb 27 17:36:17 crc kubenswrapper[4700]: I0227 17:36:17.555008 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kf8kd" podStartSLOduration=2.060094931 podStartE2EDuration="2.554967367s" podCreationTimestamp="2026-02-27 17:36:15 +0000 UTC" firstStartedPulling="2026-02-27 17:36:16.557456981 +0000 UTC m=+2136.542769758" lastFinishedPulling="2026-02-27 17:36:17.052329427 +0000 UTC m=+2137.037642194" observedRunningTime="2026-02-27 17:36:17.547244585 +0000 UTC m=+2137.532557362" watchObservedRunningTime="2026-02-27 17:36:17.554967367 +0000 UTC m=+2137.540280144" Feb 27 17:36:36 crc kubenswrapper[4700]: I0227 17:36:36.410244 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:36:36 crc kubenswrapper[4700]: I0227 17:36:36.410892 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:36:36 crc kubenswrapper[4700]: I0227 17:36:36.410956 4700 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" Feb 27 17:36:36 crc kubenswrapper[4700]: I0227 17:36:36.412042 4700 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d14a07f9a7a8a0cecf47fa738d13cde159084a95d4b48612e05d84dc5f14de89"} pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 17:36:36 crc kubenswrapper[4700]: I0227 17:36:36.412141 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" containerID="cri-o://d14a07f9a7a8a0cecf47fa738d13cde159084a95d4b48612e05d84dc5f14de89" gracePeriod=600 Feb 27 17:36:36 crc kubenswrapper[4700]: I0227 17:36:36.743739 4700 generic.go:334] "Generic (PLEG): container finished" podID="84b45511-b94c-479f-98db-fd2c4eceec46" containerID="d14a07f9a7a8a0cecf47fa738d13cde159084a95d4b48612e05d84dc5f14de89" exitCode=0 Feb 27 17:36:36 crc kubenswrapper[4700]: I0227 17:36:36.743765 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerDied","Data":"d14a07f9a7a8a0cecf47fa738d13cde159084a95d4b48612e05d84dc5f14de89"} Feb 27 17:36:36 crc kubenswrapper[4700]: I0227 17:36:36.744125 4700 scope.go:117] "RemoveContainer" containerID="88adee85741f363c1411884023846cd0aab91e2b1c4dca4a79262d1b15c4192f" Feb 27 17:36:37 crc kubenswrapper[4700]: I0227 17:36:37.757686 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerStarted","Data":"6843fae59a5a033363e409d439168388b8322ceaaef013a2a8f29582e161bd02"} Feb 27 17:36:48 crc kubenswrapper[4700]: I0227 17:36:48.037838 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ldt5k"] Feb 27 17:36:48 crc kubenswrapper[4700]: I0227 17:36:48.043528 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ldt5k" Feb 27 17:36:48 crc kubenswrapper[4700]: I0227 17:36:48.049064 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ldt5k"] Feb 27 17:36:48 crc kubenswrapper[4700]: I0227 17:36:48.058064 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a33b54d1-a07b-42cf-96a1-47d55aaeab3f-utilities\") pod \"certified-operators-ldt5k\" (UID: \"a33b54d1-a07b-42cf-96a1-47d55aaeab3f\") " pod="openshift-marketplace/certified-operators-ldt5k" Feb 27 17:36:48 crc kubenswrapper[4700]: I0227 17:36:48.058193 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a33b54d1-a07b-42cf-96a1-47d55aaeab3f-catalog-content\") pod \"certified-operators-ldt5k\" (UID: \"a33b54d1-a07b-42cf-96a1-47d55aaeab3f\") " pod="openshift-marketplace/certified-operators-ldt5k" Feb 27 17:36:48 crc kubenswrapper[4700]: I0227 17:36:48.058354 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbkfz\" (UniqueName: \"kubernetes.io/projected/a33b54d1-a07b-42cf-96a1-47d55aaeab3f-kube-api-access-zbkfz\") pod \"certified-operators-ldt5k\" (UID: \"a33b54d1-a07b-42cf-96a1-47d55aaeab3f\") " pod="openshift-marketplace/certified-operators-ldt5k" Feb 27 17:36:48 crc kubenswrapper[4700]: I0227 17:36:48.159592 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbkfz\" (UniqueName: \"kubernetes.io/projected/a33b54d1-a07b-42cf-96a1-47d55aaeab3f-kube-api-access-zbkfz\") pod \"certified-operators-ldt5k\" (UID: \"a33b54d1-a07b-42cf-96a1-47d55aaeab3f\") " pod="openshift-marketplace/certified-operators-ldt5k" Feb 27 17:36:48 crc kubenswrapper[4700]: I0227 17:36:48.160123 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a33b54d1-a07b-42cf-96a1-47d55aaeab3f-utilities\") pod \"certified-operators-ldt5k\" (UID: \"a33b54d1-a07b-42cf-96a1-47d55aaeab3f\") " pod="openshift-marketplace/certified-operators-ldt5k" Feb 27 17:36:48 crc kubenswrapper[4700]: I0227 17:36:48.160211 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a33b54d1-a07b-42cf-96a1-47d55aaeab3f-catalog-content\") pod \"certified-operators-ldt5k\" (UID: \"a33b54d1-a07b-42cf-96a1-47d55aaeab3f\") " pod="openshift-marketplace/certified-operators-ldt5k" Feb 27 17:36:48 crc kubenswrapper[4700]: I0227 17:36:48.160607 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a33b54d1-a07b-42cf-96a1-47d55aaeab3f-utilities\") pod \"certified-operators-ldt5k\" (UID: \"a33b54d1-a07b-42cf-96a1-47d55aaeab3f\") " pod="openshift-marketplace/certified-operators-ldt5k" Feb 27 17:36:48 crc kubenswrapper[4700]: I0227 17:36:48.160672 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a33b54d1-a07b-42cf-96a1-47d55aaeab3f-catalog-content\") pod \"certified-operators-ldt5k\" (UID: \"a33b54d1-a07b-42cf-96a1-47d55aaeab3f\") " pod="openshift-marketplace/certified-operators-ldt5k" Feb 27 17:36:48 crc kubenswrapper[4700]: I0227 17:36:48.185805 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbkfz\" (UniqueName: \"kubernetes.io/projected/a33b54d1-a07b-42cf-96a1-47d55aaeab3f-kube-api-access-zbkfz\") pod \"certified-operators-ldt5k\" (UID: \"a33b54d1-a07b-42cf-96a1-47d55aaeab3f\") " pod="openshift-marketplace/certified-operators-ldt5k" Feb 27 17:36:48 crc kubenswrapper[4700]: I0227 17:36:48.368815 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ldt5k" Feb 27 17:36:48 crc kubenswrapper[4700]: I0227 17:36:48.918262 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ldt5k"] Feb 27 17:36:49 crc kubenswrapper[4700]: I0227 17:36:49.884956 4700 generic.go:334] "Generic (PLEG): container finished" podID="a33b54d1-a07b-42cf-96a1-47d55aaeab3f" containerID="1051ba9d83f809d6b2d69f161e6f8d628f06e97edb8052bc96ff3e373985accf" exitCode=0 Feb 27 17:36:49 crc kubenswrapper[4700]: I0227 17:36:49.885213 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ldt5k" event={"ID":"a33b54d1-a07b-42cf-96a1-47d55aaeab3f","Type":"ContainerDied","Data":"1051ba9d83f809d6b2d69f161e6f8d628f06e97edb8052bc96ff3e373985accf"} Feb 27 17:36:49 crc kubenswrapper[4700]: I0227 17:36:49.885653 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ldt5k" event={"ID":"a33b54d1-a07b-42cf-96a1-47d55aaeab3f","Type":"ContainerStarted","Data":"89fd4673693161218c025091647b88abb1f6ea6706618eeb31e06490e5ae1fe8"} Feb 27 17:36:49 crc kubenswrapper[4700]: I0227 17:36:49.889724 4700 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 17:36:54 crc kubenswrapper[4700]: E0227 17:36:54.704670 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 27 17:36:54 crc kubenswrapper[4700]: E0227 17:36:54.705291 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zbkfz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-ldt5k_openshift-marketplace(a33b54d1-a07b-42cf-96a1-47d55aaeab3f): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 17:36:54 crc kubenswrapper[4700]: E0227 17:36:54.706481 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/certified-operators-ldt5k" podUID="a33b54d1-a07b-42cf-96a1-47d55aaeab3f" Feb 27 17:36:54 crc kubenswrapper[4700]: E0227 17:36:54.967960 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-ldt5k" podUID="a33b54d1-a07b-42cf-96a1-47d55aaeab3f" Feb 27 17:37:08 crc kubenswrapper[4700]: E0227 17:37:08.096733 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 27 17:37:08 crc kubenswrapper[4700]: E0227 17:37:08.097729 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zbkfz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-ldt5k_openshift-marketplace(a33b54d1-a07b-42cf-96a1-47d55aaeab3f): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 17:37:08 crc kubenswrapper[4700]: E0227 17:37:08.098914 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/certified-operators-ldt5k" podUID="a33b54d1-a07b-42cf-96a1-47d55aaeab3f" Feb 27 17:37:22 crc kubenswrapper[4700]: E0227 17:37:22.985343 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-ldt5k" podUID="a33b54d1-a07b-42cf-96a1-47d55aaeab3f" Feb 27 17:37:26 crc kubenswrapper[4700]: I0227 17:37:26.371073 4700 generic.go:334] "Generic (PLEG): container finished" podID="e50903c5-4b11-4c01-b14f-7f0e51be5319" containerID="71f08a4f6d54dd94c7595ce865ec0f4c94cf700a807bebfaa8fac16fe39e21a8" exitCode=0 Feb 27 17:37:26 crc kubenswrapper[4700]: I0227 17:37:26.371202 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kf8kd" event={"ID":"e50903c5-4b11-4c01-b14f-7f0e51be5319","Type":"ContainerDied","Data":"71f08a4f6d54dd94c7595ce865ec0f4c94cf700a807bebfaa8fac16fe39e21a8"} Feb 27 17:37:27 crc kubenswrapper[4700]: I0227 17:37:27.852582 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kf8kd" Feb 27 17:37:27 crc kubenswrapper[4700]: I0227 17:37:27.967395 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e50903c5-4b11-4c01-b14f-7f0e51be5319-inventory\") pod \"e50903c5-4b11-4c01-b14f-7f0e51be5319\" (UID: \"e50903c5-4b11-4c01-b14f-7f0e51be5319\") " Feb 27 17:37:27 crc kubenswrapper[4700]: I0227 17:37:27.967476 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e50903c5-4b11-4c01-b14f-7f0e51be5319-ovn-combined-ca-bundle\") pod \"e50903c5-4b11-4c01-b14f-7f0e51be5319\" (UID: \"e50903c5-4b11-4c01-b14f-7f0e51be5319\") " Feb 27 17:37:27 crc kubenswrapper[4700]: I0227 17:37:27.967520 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d76jq\" (UniqueName: \"kubernetes.io/projected/e50903c5-4b11-4c01-b14f-7f0e51be5319-kube-api-access-d76jq\") pod \"e50903c5-4b11-4c01-b14f-7f0e51be5319\" (UID: \"e50903c5-4b11-4c01-b14f-7f0e51be5319\") " Feb 27 17:37:27 crc kubenswrapper[4700]: I0227 17:37:27.967662 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e50903c5-4b11-4c01-b14f-7f0e51be5319-ssh-key-openstack-edpm-ipam\") pod \"e50903c5-4b11-4c01-b14f-7f0e51be5319\" (UID: \"e50903c5-4b11-4c01-b14f-7f0e51be5319\") " Feb 27 17:37:27 crc kubenswrapper[4700]: I0227 17:37:27.967727 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e50903c5-4b11-4c01-b14f-7f0e51be5319-ovncontroller-config-0\") pod \"e50903c5-4b11-4c01-b14f-7f0e51be5319\" (UID: \"e50903c5-4b11-4c01-b14f-7f0e51be5319\") " Feb 27 17:37:27 crc kubenswrapper[4700]: I0227 17:37:27.980589 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e50903c5-4b11-4c01-b14f-7f0e51be5319-kube-api-access-d76jq" (OuterVolumeSpecName: "kube-api-access-d76jq") pod "e50903c5-4b11-4c01-b14f-7f0e51be5319" (UID: "e50903c5-4b11-4c01-b14f-7f0e51be5319"). InnerVolumeSpecName "kube-api-access-d76jq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:37:27 crc kubenswrapper[4700]: I0227 17:37:27.989623 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e50903c5-4b11-4c01-b14f-7f0e51be5319-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "e50903c5-4b11-4c01-b14f-7f0e51be5319" (UID: "e50903c5-4b11-4c01-b14f-7f0e51be5319"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.002723 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e50903c5-4b11-4c01-b14f-7f0e51be5319-inventory" (OuterVolumeSpecName: "inventory") pod "e50903c5-4b11-4c01-b14f-7f0e51be5319" (UID: "e50903c5-4b11-4c01-b14f-7f0e51be5319"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.006311 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e50903c5-4b11-4c01-b14f-7f0e51be5319-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "e50903c5-4b11-4c01-b14f-7f0e51be5319" (UID: "e50903c5-4b11-4c01-b14f-7f0e51be5319"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.020938 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e50903c5-4b11-4c01-b14f-7f0e51be5319-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "e50903c5-4b11-4c01-b14f-7f0e51be5319" (UID: "e50903c5-4b11-4c01-b14f-7f0e51be5319"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.071215 4700 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e50903c5-4b11-4c01-b14f-7f0e51be5319-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.071269 4700 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e50903c5-4b11-4c01-b14f-7f0e51be5319-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.071292 4700 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e50903c5-4b11-4c01-b14f-7f0e51be5319-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.071310 4700 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e50903c5-4b11-4c01-b14f-7f0e51be5319-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.071327 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d76jq\" (UniqueName: \"kubernetes.io/projected/e50903c5-4b11-4c01-b14f-7f0e51be5319-kube-api-access-d76jq\") on node \"crc\" DevicePath \"\"" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.395403 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kf8kd" event={"ID":"e50903c5-4b11-4c01-b14f-7f0e51be5319","Type":"ContainerDied","Data":"5ad40fcc4feaeef8cf9274b650002597fef1738eb1266eeffd1cd2f9b04a4f06"} Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.395678 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ad40fcc4feaeef8cf9274b650002597fef1738eb1266eeffd1cd2f9b04a4f06" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.395502 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kf8kd" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.618929 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg"] Feb 27 17:37:28 crc kubenswrapper[4700]: E0227 17:37:28.619483 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e50903c5-4b11-4c01-b14f-7f0e51be5319" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.619503 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="e50903c5-4b11-4c01-b14f-7f0e51be5319" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.619770 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="e50903c5-4b11-4c01-b14f-7f0e51be5319" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.620668 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.624562 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.624584 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.624667 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dtdm4" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.626443 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.626454 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.629122 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.631986 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg"] Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.785904 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a7c20871-f705-46a0-9856-e747aed1874e-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg\" (UID: \"a7c20871-f705-46a0-9856-e747aed1874e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.785981 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a7c20871-f705-46a0-9856-e747aed1874e-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg\" (UID: \"a7c20871-f705-46a0-9856-e747aed1874e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.786106 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7c20871-f705-46a0-9856-e747aed1874e-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg\" (UID: \"a7c20871-f705-46a0-9856-e747aed1874e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.786306 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7c20871-f705-46a0-9856-e747aed1874e-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg\" (UID: \"a7c20871-f705-46a0-9856-e747aed1874e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.786377 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a7c20871-f705-46a0-9856-e747aed1874e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg\" (UID: \"a7c20871-f705-46a0-9856-e747aed1874e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.786571 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wm9wk\" (UniqueName: \"kubernetes.io/projected/a7c20871-f705-46a0-9856-e747aed1874e-kube-api-access-wm9wk\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg\" (UID: \"a7c20871-f705-46a0-9856-e747aed1874e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.888887 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7c20871-f705-46a0-9856-e747aed1874e-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg\" (UID: \"a7c20871-f705-46a0-9856-e747aed1874e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.889056 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7c20871-f705-46a0-9856-e747aed1874e-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg\" (UID: \"a7c20871-f705-46a0-9856-e747aed1874e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.889147 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a7c20871-f705-46a0-9856-e747aed1874e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg\" (UID: \"a7c20871-f705-46a0-9856-e747aed1874e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.889279 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wm9wk\" (UniqueName: \"kubernetes.io/projected/a7c20871-f705-46a0-9856-e747aed1874e-kube-api-access-wm9wk\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg\" (UID: \"a7c20871-f705-46a0-9856-e747aed1874e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.889392 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a7c20871-f705-46a0-9856-e747aed1874e-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg\" (UID: \"a7c20871-f705-46a0-9856-e747aed1874e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.889428 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a7c20871-f705-46a0-9856-e747aed1874e-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg\" (UID: \"a7c20871-f705-46a0-9856-e747aed1874e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.895800 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a7c20871-f705-46a0-9856-e747aed1874e-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg\" (UID: \"a7c20871-f705-46a0-9856-e747aed1874e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.895858 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7c20871-f705-46a0-9856-e747aed1874e-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg\" (UID: \"a7c20871-f705-46a0-9856-e747aed1874e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.897375 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a7c20871-f705-46a0-9856-e747aed1874e-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg\" (UID: \"a7c20871-f705-46a0-9856-e747aed1874e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.898205 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a7c20871-f705-46a0-9856-e747aed1874e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg\" (UID: \"a7c20871-f705-46a0-9856-e747aed1874e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.899146 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7c20871-f705-46a0-9856-e747aed1874e-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg\" (UID: \"a7c20871-f705-46a0-9856-e747aed1874e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.913872 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wm9wk\" (UniqueName: \"kubernetes.io/projected/a7c20871-f705-46a0-9856-e747aed1874e-kube-api-access-wm9wk\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg\" (UID: \"a7c20871-f705-46a0-9856-e747aed1874e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg" Feb 27 17:37:28 crc kubenswrapper[4700]: I0227 17:37:28.942647 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg" Feb 27 17:37:29 crc kubenswrapper[4700]: I0227 17:37:29.537849 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg"] Feb 27 17:37:30 crc kubenswrapper[4700]: I0227 17:37:30.414976 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg" event={"ID":"a7c20871-f705-46a0-9856-e747aed1874e","Type":"ContainerStarted","Data":"3c0c4df09edd1e8d4f2f7db8532acd7a6ee85cd299dfe6eddb0f3572557c1c36"} Feb 27 17:37:31 crc kubenswrapper[4700]: I0227 17:37:31.428553 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg" event={"ID":"a7c20871-f705-46a0-9856-e747aed1874e","Type":"ContainerStarted","Data":"f10d128113577afd0337b2c39d9a373d401e2029d79790a28347ad3b4fb2bb24"} Feb 27 17:37:31 crc kubenswrapper[4700]: I0227 17:37:31.457222 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg" podStartSLOduration=2.8033189 podStartE2EDuration="3.457201723s" podCreationTimestamp="2026-02-27 17:37:28 +0000 UTC" firstStartedPulling="2026-02-27 17:37:29.539671249 +0000 UTC m=+2209.524984006" lastFinishedPulling="2026-02-27 17:37:30.193554042 +0000 UTC m=+2210.178866829" observedRunningTime="2026-02-27 17:37:31.45171142 +0000 UTC m=+2211.437024167" watchObservedRunningTime="2026-02-27 17:37:31.457201723 +0000 UTC m=+2211.442514470" Feb 27 17:37:39 crc kubenswrapper[4700]: I0227 17:37:39.531658 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ldt5k" event={"ID":"a33b54d1-a07b-42cf-96a1-47d55aaeab3f","Type":"ContainerStarted","Data":"6bf2e96841cc95eb17e24558dc9b173d7cdc427ea69a72b42cb7905c7003499e"} Feb 27 17:37:42 crc kubenswrapper[4700]: I0227 17:37:42.580977 4700 generic.go:334] "Generic (PLEG): container finished" podID="a33b54d1-a07b-42cf-96a1-47d55aaeab3f" containerID="6bf2e96841cc95eb17e24558dc9b173d7cdc427ea69a72b42cb7905c7003499e" exitCode=0 Feb 27 17:37:42 crc kubenswrapper[4700]: I0227 17:37:42.582010 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ldt5k" event={"ID":"a33b54d1-a07b-42cf-96a1-47d55aaeab3f","Type":"ContainerDied","Data":"6bf2e96841cc95eb17e24558dc9b173d7cdc427ea69a72b42cb7905c7003499e"} Feb 27 17:37:45 crc kubenswrapper[4700]: I0227 17:37:45.621313 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ldt5k" event={"ID":"a33b54d1-a07b-42cf-96a1-47d55aaeab3f","Type":"ContainerStarted","Data":"f4c0dc984002982f69111010f7eee414c97519e95ca24033ead6c86e1c17b9e2"} Feb 27 17:37:45 crc kubenswrapper[4700]: I0227 17:37:45.656607 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ldt5k" podStartSLOduration=3.107480979 podStartE2EDuration="57.656583808s" podCreationTimestamp="2026-02-27 17:36:48 +0000 UTC" firstStartedPulling="2026-02-27 17:36:49.889424142 +0000 UTC m=+2169.874736899" lastFinishedPulling="2026-02-27 17:37:44.438526971 +0000 UTC m=+2224.423839728" observedRunningTime="2026-02-27 17:37:45.642747806 +0000 UTC m=+2225.628060543" watchObservedRunningTime="2026-02-27 17:37:45.656583808 +0000 UTC m=+2225.641896555" Feb 27 17:37:48 crc kubenswrapper[4700]: I0227 17:37:48.369922 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ldt5k" Feb 27 17:37:48 crc kubenswrapper[4700]: I0227 17:37:48.370929 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ldt5k" Feb 27 17:37:48 crc kubenswrapper[4700]: I0227 17:37:48.446867 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ldt5k" Feb 27 17:37:58 crc kubenswrapper[4700]: I0227 17:37:58.467348 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ldt5k" Feb 27 17:37:58 crc kubenswrapper[4700]: I0227 17:37:58.530511 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ldt5k"] Feb 27 17:37:58 crc kubenswrapper[4700]: I0227 17:37:58.789573 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ldt5k" podUID="a33b54d1-a07b-42cf-96a1-47d55aaeab3f" containerName="registry-server" containerID="cri-o://f4c0dc984002982f69111010f7eee414c97519e95ca24033ead6c86e1c17b9e2" gracePeriod=2 Feb 27 17:37:59 crc kubenswrapper[4700]: I0227 17:37:59.369287 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ldt5k" Feb 27 17:37:59 crc kubenswrapper[4700]: I0227 17:37:59.426515 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a33b54d1-a07b-42cf-96a1-47d55aaeab3f-catalog-content\") pod \"a33b54d1-a07b-42cf-96a1-47d55aaeab3f\" (UID: \"a33b54d1-a07b-42cf-96a1-47d55aaeab3f\") " Feb 27 17:37:59 crc kubenswrapper[4700]: I0227 17:37:59.426780 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a33b54d1-a07b-42cf-96a1-47d55aaeab3f-utilities\") pod \"a33b54d1-a07b-42cf-96a1-47d55aaeab3f\" (UID: \"a33b54d1-a07b-42cf-96a1-47d55aaeab3f\") " Feb 27 17:37:59 crc kubenswrapper[4700]: I0227 17:37:59.426849 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbkfz\" (UniqueName: \"kubernetes.io/projected/a33b54d1-a07b-42cf-96a1-47d55aaeab3f-kube-api-access-zbkfz\") pod \"a33b54d1-a07b-42cf-96a1-47d55aaeab3f\" (UID: \"a33b54d1-a07b-42cf-96a1-47d55aaeab3f\") " Feb 27 17:37:59 crc kubenswrapper[4700]: I0227 17:37:59.428563 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a33b54d1-a07b-42cf-96a1-47d55aaeab3f-utilities" (OuterVolumeSpecName: "utilities") pod "a33b54d1-a07b-42cf-96a1-47d55aaeab3f" (UID: "a33b54d1-a07b-42cf-96a1-47d55aaeab3f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:37:59 crc kubenswrapper[4700]: I0227 17:37:59.440500 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a33b54d1-a07b-42cf-96a1-47d55aaeab3f-kube-api-access-zbkfz" (OuterVolumeSpecName: "kube-api-access-zbkfz") pod "a33b54d1-a07b-42cf-96a1-47d55aaeab3f" (UID: "a33b54d1-a07b-42cf-96a1-47d55aaeab3f"). InnerVolumeSpecName "kube-api-access-zbkfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:37:59 crc kubenswrapper[4700]: I0227 17:37:59.485890 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a33b54d1-a07b-42cf-96a1-47d55aaeab3f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a33b54d1-a07b-42cf-96a1-47d55aaeab3f" (UID: "a33b54d1-a07b-42cf-96a1-47d55aaeab3f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:37:59 crc kubenswrapper[4700]: I0227 17:37:59.529980 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a33b54d1-a07b-42cf-96a1-47d55aaeab3f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:37:59 crc kubenswrapper[4700]: I0227 17:37:59.530006 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a33b54d1-a07b-42cf-96a1-47d55aaeab3f-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:37:59 crc kubenswrapper[4700]: I0227 17:37:59.530017 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zbkfz\" (UniqueName: \"kubernetes.io/projected/a33b54d1-a07b-42cf-96a1-47d55aaeab3f-kube-api-access-zbkfz\") on node \"crc\" DevicePath \"\"" Feb 27 17:37:59 crc kubenswrapper[4700]: I0227 17:37:59.803721 4700 generic.go:334] "Generic (PLEG): container finished" podID="a33b54d1-a07b-42cf-96a1-47d55aaeab3f" containerID="f4c0dc984002982f69111010f7eee414c97519e95ca24033ead6c86e1c17b9e2" exitCode=0 Feb 27 17:37:59 crc kubenswrapper[4700]: I0227 17:37:59.803765 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ldt5k" event={"ID":"a33b54d1-a07b-42cf-96a1-47d55aaeab3f","Type":"ContainerDied","Data":"f4c0dc984002982f69111010f7eee414c97519e95ca24033ead6c86e1c17b9e2"} Feb 27 17:37:59 crc kubenswrapper[4700]: I0227 17:37:59.803794 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ldt5k" event={"ID":"a33b54d1-a07b-42cf-96a1-47d55aaeab3f","Type":"ContainerDied","Data":"89fd4673693161218c025091647b88abb1f6ea6706618eeb31e06490e5ae1fe8"} Feb 27 17:37:59 crc kubenswrapper[4700]: I0227 17:37:59.803813 4700 scope.go:117] "RemoveContainer" containerID="f4c0dc984002982f69111010f7eee414c97519e95ca24033ead6c86e1c17b9e2" Feb 27 17:37:59 crc kubenswrapper[4700]: I0227 17:37:59.803833 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ldt5k" Feb 27 17:37:59 crc kubenswrapper[4700]: I0227 17:37:59.826110 4700 scope.go:117] "RemoveContainer" containerID="6bf2e96841cc95eb17e24558dc9b173d7cdc427ea69a72b42cb7905c7003499e" Feb 27 17:37:59 crc kubenswrapper[4700]: I0227 17:37:59.863872 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ldt5k"] Feb 27 17:37:59 crc kubenswrapper[4700]: I0227 17:37:59.864104 4700 scope.go:117] "RemoveContainer" containerID="1051ba9d83f809d6b2d69f161e6f8d628f06e97edb8052bc96ff3e373985accf" Feb 27 17:37:59 crc kubenswrapper[4700]: I0227 17:37:59.873676 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ldt5k"] Feb 27 17:37:59 crc kubenswrapper[4700]: I0227 17:37:59.929726 4700 scope.go:117] "RemoveContainer" containerID="f4c0dc984002982f69111010f7eee414c97519e95ca24033ead6c86e1c17b9e2" Feb 27 17:37:59 crc kubenswrapper[4700]: E0227 17:37:59.930837 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4c0dc984002982f69111010f7eee414c97519e95ca24033ead6c86e1c17b9e2\": container with ID starting with f4c0dc984002982f69111010f7eee414c97519e95ca24033ead6c86e1c17b9e2 not found: ID does not exist" containerID="f4c0dc984002982f69111010f7eee414c97519e95ca24033ead6c86e1c17b9e2" Feb 27 17:37:59 crc kubenswrapper[4700]: I0227 17:37:59.930916 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4c0dc984002982f69111010f7eee414c97519e95ca24033ead6c86e1c17b9e2"} err="failed to get container status \"f4c0dc984002982f69111010f7eee414c97519e95ca24033ead6c86e1c17b9e2\": rpc error: code = NotFound desc = could not find container \"f4c0dc984002982f69111010f7eee414c97519e95ca24033ead6c86e1c17b9e2\": container with ID starting with f4c0dc984002982f69111010f7eee414c97519e95ca24033ead6c86e1c17b9e2 not found: ID does not exist" Feb 27 17:37:59 crc kubenswrapper[4700]: I0227 17:37:59.931004 4700 scope.go:117] "RemoveContainer" containerID="6bf2e96841cc95eb17e24558dc9b173d7cdc427ea69a72b42cb7905c7003499e" Feb 27 17:37:59 crc kubenswrapper[4700]: E0227 17:37:59.931704 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bf2e96841cc95eb17e24558dc9b173d7cdc427ea69a72b42cb7905c7003499e\": container with ID starting with 6bf2e96841cc95eb17e24558dc9b173d7cdc427ea69a72b42cb7905c7003499e not found: ID does not exist" containerID="6bf2e96841cc95eb17e24558dc9b173d7cdc427ea69a72b42cb7905c7003499e" Feb 27 17:37:59 crc kubenswrapper[4700]: I0227 17:37:59.931758 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bf2e96841cc95eb17e24558dc9b173d7cdc427ea69a72b42cb7905c7003499e"} err="failed to get container status \"6bf2e96841cc95eb17e24558dc9b173d7cdc427ea69a72b42cb7905c7003499e\": rpc error: code = NotFound desc = could not find container \"6bf2e96841cc95eb17e24558dc9b173d7cdc427ea69a72b42cb7905c7003499e\": container with ID starting with 6bf2e96841cc95eb17e24558dc9b173d7cdc427ea69a72b42cb7905c7003499e not found: ID does not exist" Feb 27 17:37:59 crc kubenswrapper[4700]: I0227 17:37:59.931794 4700 scope.go:117] "RemoveContainer" containerID="1051ba9d83f809d6b2d69f161e6f8d628f06e97edb8052bc96ff3e373985accf" Feb 27 17:37:59 crc kubenswrapper[4700]: E0227 17:37:59.932196 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1051ba9d83f809d6b2d69f161e6f8d628f06e97edb8052bc96ff3e373985accf\": container with ID starting with 1051ba9d83f809d6b2d69f161e6f8d628f06e97edb8052bc96ff3e373985accf not found: ID does not exist" containerID="1051ba9d83f809d6b2d69f161e6f8d628f06e97edb8052bc96ff3e373985accf" Feb 27 17:37:59 crc kubenswrapper[4700]: I0227 17:37:59.932249 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1051ba9d83f809d6b2d69f161e6f8d628f06e97edb8052bc96ff3e373985accf"} err="failed to get container status \"1051ba9d83f809d6b2d69f161e6f8d628f06e97edb8052bc96ff3e373985accf\": rpc error: code = NotFound desc = could not find container \"1051ba9d83f809d6b2d69f161e6f8d628f06e97edb8052bc96ff3e373985accf\": container with ID starting with 1051ba9d83f809d6b2d69f161e6f8d628f06e97edb8052bc96ff3e373985accf not found: ID does not exist" Feb 27 17:38:00 crc kubenswrapper[4700]: I0227 17:38:00.172861 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536898-4zjfl"] Feb 27 17:38:00 crc kubenswrapper[4700]: E0227 17:38:00.173539 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a33b54d1-a07b-42cf-96a1-47d55aaeab3f" containerName="extract-content" Feb 27 17:38:00 crc kubenswrapper[4700]: I0227 17:38:00.173570 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="a33b54d1-a07b-42cf-96a1-47d55aaeab3f" containerName="extract-content" Feb 27 17:38:00 crc kubenswrapper[4700]: E0227 17:38:00.173615 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a33b54d1-a07b-42cf-96a1-47d55aaeab3f" containerName="extract-utilities" Feb 27 17:38:00 crc kubenswrapper[4700]: I0227 17:38:00.173630 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="a33b54d1-a07b-42cf-96a1-47d55aaeab3f" containerName="extract-utilities" Feb 27 17:38:00 crc kubenswrapper[4700]: E0227 17:38:00.173658 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a33b54d1-a07b-42cf-96a1-47d55aaeab3f" containerName="registry-server" Feb 27 17:38:00 crc kubenswrapper[4700]: I0227 17:38:00.173670 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="a33b54d1-a07b-42cf-96a1-47d55aaeab3f" containerName="registry-server" Feb 27 17:38:00 crc kubenswrapper[4700]: I0227 17:38:00.174021 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="a33b54d1-a07b-42cf-96a1-47d55aaeab3f" containerName="registry-server" Feb 27 17:38:00 crc kubenswrapper[4700]: I0227 17:38:00.175141 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536898-4zjfl" Feb 27 17:38:00 crc kubenswrapper[4700]: I0227 17:38:00.177660 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d62kc" Feb 27 17:38:00 crc kubenswrapper[4700]: I0227 17:38:00.177885 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:38:00 crc kubenswrapper[4700]: I0227 17:38:00.180128 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:38:00 crc kubenswrapper[4700]: I0227 17:38:00.195709 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536898-4zjfl"] Feb 27 17:38:00 crc kubenswrapper[4700]: I0227 17:38:00.244997 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fjcr\" (UniqueName: \"kubernetes.io/projected/6c095a55-8ba9-44a9-b4c0-656cf3de7880-kube-api-access-5fjcr\") pod \"auto-csr-approver-29536898-4zjfl\" (UID: \"6c095a55-8ba9-44a9-b4c0-656cf3de7880\") " pod="openshift-infra/auto-csr-approver-29536898-4zjfl" Feb 27 17:38:00 crc kubenswrapper[4700]: I0227 17:38:00.347529 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fjcr\" (UniqueName: \"kubernetes.io/projected/6c095a55-8ba9-44a9-b4c0-656cf3de7880-kube-api-access-5fjcr\") pod \"auto-csr-approver-29536898-4zjfl\" (UID: \"6c095a55-8ba9-44a9-b4c0-656cf3de7880\") " pod="openshift-infra/auto-csr-approver-29536898-4zjfl" Feb 27 17:38:00 crc kubenswrapper[4700]: I0227 17:38:00.371187 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fjcr\" (UniqueName: \"kubernetes.io/projected/6c095a55-8ba9-44a9-b4c0-656cf3de7880-kube-api-access-5fjcr\") pod \"auto-csr-approver-29536898-4zjfl\" (UID: \"6c095a55-8ba9-44a9-b4c0-656cf3de7880\") " pod="openshift-infra/auto-csr-approver-29536898-4zjfl" Feb 27 17:38:00 crc kubenswrapper[4700]: I0227 17:38:00.494759 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536898-4zjfl" Feb 27 17:38:01 crc kubenswrapper[4700]: I0227 17:38:01.015790 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a33b54d1-a07b-42cf-96a1-47d55aaeab3f" path="/var/lib/kubelet/pods/a33b54d1-a07b-42cf-96a1-47d55aaeab3f/volumes" Feb 27 17:38:01 crc kubenswrapper[4700]: I0227 17:38:01.017477 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536898-4zjfl"] Feb 27 17:38:01 crc kubenswrapper[4700]: I0227 17:38:01.825636 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536898-4zjfl" event={"ID":"6c095a55-8ba9-44a9-b4c0-656cf3de7880","Type":"ContainerStarted","Data":"1ac9aa97e4610ebb53f8ddcaca87ecedf0f5d7ecf89e380e1ff8516bdcdd52f1"} Feb 27 17:38:01 crc kubenswrapper[4700]: E0227 17:38:01.935358 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 17:38:01 crc kubenswrapper[4700]: E0227 17:38:01.935566 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 17:38:01 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 17:38:01 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5fjcr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536898-4zjfl_openshift-infra(6c095a55-8ba9-44a9-b4c0-656cf3de7880): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 17:38:01 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 17:38:01 crc kubenswrapper[4700]: E0227 17:38:01.936751 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536898-4zjfl" podUID="6c095a55-8ba9-44a9-b4c0-656cf3de7880" Feb 27 17:38:02 crc kubenswrapper[4700]: E0227 17:38:02.841332 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536898-4zjfl" podUID="6c095a55-8ba9-44a9-b4c0-656cf3de7880" Feb 27 17:38:22 crc kubenswrapper[4700]: I0227 17:38:22.066801 4700 generic.go:334] "Generic (PLEG): container finished" podID="a7c20871-f705-46a0-9856-e747aed1874e" containerID="f10d128113577afd0337b2c39d9a373d401e2029d79790a28347ad3b4fb2bb24" exitCode=0 Feb 27 17:38:22 crc kubenswrapper[4700]: I0227 17:38:22.066894 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg" event={"ID":"a7c20871-f705-46a0-9856-e747aed1874e","Type":"ContainerDied","Data":"f10d128113577afd0337b2c39d9a373d401e2029d79790a28347ad3b4fb2bb24"} Feb 27 17:38:23 crc kubenswrapper[4700]: I0227 17:38:23.570124 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg" Feb 27 17:38:23 crc kubenswrapper[4700]: I0227 17:38:23.656405 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7c20871-f705-46a0-9856-e747aed1874e-inventory\") pod \"a7c20871-f705-46a0-9856-e747aed1874e\" (UID: \"a7c20871-f705-46a0-9856-e747aed1874e\") " Feb 27 17:38:23 crc kubenswrapper[4700]: I0227 17:38:23.656494 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wm9wk\" (UniqueName: \"kubernetes.io/projected/a7c20871-f705-46a0-9856-e747aed1874e-kube-api-access-wm9wk\") pod \"a7c20871-f705-46a0-9856-e747aed1874e\" (UID: \"a7c20871-f705-46a0-9856-e747aed1874e\") " Feb 27 17:38:23 crc kubenswrapper[4700]: I0227 17:38:23.656571 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a7c20871-f705-46a0-9856-e747aed1874e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"a7c20871-f705-46a0-9856-e747aed1874e\" (UID: \"a7c20871-f705-46a0-9856-e747aed1874e\") " Feb 27 17:38:23 crc kubenswrapper[4700]: I0227 17:38:23.656720 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a7c20871-f705-46a0-9856-e747aed1874e-ssh-key-openstack-edpm-ipam\") pod \"a7c20871-f705-46a0-9856-e747aed1874e\" (UID: \"a7c20871-f705-46a0-9856-e747aed1874e\") " Feb 27 17:38:23 crc kubenswrapper[4700]: I0227 17:38:23.656772 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a7c20871-f705-46a0-9856-e747aed1874e-nova-metadata-neutron-config-0\") pod \"a7c20871-f705-46a0-9856-e747aed1874e\" (UID: \"a7c20871-f705-46a0-9856-e747aed1874e\") " Feb 27 17:38:23 crc kubenswrapper[4700]: I0227 17:38:23.656844 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7c20871-f705-46a0-9856-e747aed1874e-neutron-metadata-combined-ca-bundle\") pod \"a7c20871-f705-46a0-9856-e747aed1874e\" (UID: \"a7c20871-f705-46a0-9856-e747aed1874e\") " Feb 27 17:38:23 crc kubenswrapper[4700]: I0227 17:38:23.668730 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7c20871-f705-46a0-9856-e747aed1874e-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "a7c20871-f705-46a0-9856-e747aed1874e" (UID: "a7c20871-f705-46a0-9856-e747aed1874e"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:38:23 crc kubenswrapper[4700]: I0227 17:38:23.669013 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7c20871-f705-46a0-9856-e747aed1874e-kube-api-access-wm9wk" (OuterVolumeSpecName: "kube-api-access-wm9wk") pod "a7c20871-f705-46a0-9856-e747aed1874e" (UID: "a7c20871-f705-46a0-9856-e747aed1874e"). InnerVolumeSpecName "kube-api-access-wm9wk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:38:23 crc kubenswrapper[4700]: I0227 17:38:23.696238 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7c20871-f705-46a0-9856-e747aed1874e-inventory" (OuterVolumeSpecName: "inventory") pod "a7c20871-f705-46a0-9856-e747aed1874e" (UID: "a7c20871-f705-46a0-9856-e747aed1874e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:38:23 crc kubenswrapper[4700]: I0227 17:38:23.706796 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7c20871-f705-46a0-9856-e747aed1874e-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "a7c20871-f705-46a0-9856-e747aed1874e" (UID: "a7c20871-f705-46a0-9856-e747aed1874e"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:38:23 crc kubenswrapper[4700]: I0227 17:38:23.735034 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7c20871-f705-46a0-9856-e747aed1874e-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "a7c20871-f705-46a0-9856-e747aed1874e" (UID: "a7c20871-f705-46a0-9856-e747aed1874e"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:38:23 crc kubenswrapper[4700]: I0227 17:38:23.744409 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7c20871-f705-46a0-9856-e747aed1874e-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "a7c20871-f705-46a0-9856-e747aed1874e" (UID: "a7c20871-f705-46a0-9856-e747aed1874e"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:38:23 crc kubenswrapper[4700]: I0227 17:38:23.758978 4700 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7c20871-f705-46a0-9856-e747aed1874e-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 17:38:23 crc kubenswrapper[4700]: I0227 17:38:23.759031 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wm9wk\" (UniqueName: \"kubernetes.io/projected/a7c20871-f705-46a0-9856-e747aed1874e-kube-api-access-wm9wk\") on node \"crc\" DevicePath \"\"" Feb 27 17:38:23 crc kubenswrapper[4700]: I0227 17:38:23.759047 4700 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a7c20871-f705-46a0-9856-e747aed1874e-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 27 17:38:23 crc kubenswrapper[4700]: I0227 17:38:23.759061 4700 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a7c20871-f705-46a0-9856-e747aed1874e-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 17:38:23 crc kubenswrapper[4700]: I0227 17:38:23.759076 4700 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a7c20871-f705-46a0-9856-e747aed1874e-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 27 17:38:23 crc kubenswrapper[4700]: I0227 17:38:23.759087 4700 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7c20871-f705-46a0-9856-e747aed1874e-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:38:24 crc kubenswrapper[4700]: I0227 17:38:24.092940 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg" event={"ID":"a7c20871-f705-46a0-9856-e747aed1874e","Type":"ContainerDied","Data":"3c0c4df09edd1e8d4f2f7db8532acd7a6ee85cd299dfe6eddb0f3572557c1c36"} Feb 27 17:38:24 crc kubenswrapper[4700]: I0227 17:38:24.093641 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c0c4df09edd1e8d4f2f7db8532acd7a6ee85cd299dfe6eddb0f3572557c1c36" Feb 27 17:38:24 crc kubenswrapper[4700]: I0227 17:38:24.093053 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg" Feb 27 17:38:24 crc kubenswrapper[4700]: I0227 17:38:24.202173 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p"] Feb 27 17:38:24 crc kubenswrapper[4700]: E0227 17:38:24.202801 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7c20871-f705-46a0-9856-e747aed1874e" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 27 17:38:24 crc kubenswrapper[4700]: I0227 17:38:24.202822 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7c20871-f705-46a0-9856-e747aed1874e" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 27 17:38:24 crc kubenswrapper[4700]: I0227 17:38:24.203007 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7c20871-f705-46a0-9856-e747aed1874e" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 27 17:38:24 crc kubenswrapper[4700]: I0227 17:38:24.203948 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p" Feb 27 17:38:24 crc kubenswrapper[4700]: I0227 17:38:24.206699 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 17:38:24 crc kubenswrapper[4700]: I0227 17:38:24.206998 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Feb 27 17:38:24 crc kubenswrapper[4700]: I0227 17:38:24.208949 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 17:38:24 crc kubenswrapper[4700]: I0227 17:38:24.208962 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 17:38:24 crc kubenswrapper[4700]: I0227 17:38:24.209571 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dtdm4" Feb 27 17:38:24 crc kubenswrapper[4700]: I0227 17:38:24.216678 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p"] Feb 27 17:38:24 crc kubenswrapper[4700]: I0227 17:38:24.275685 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/abd22381-3367-4aef-a05d-dd895c6de545-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p\" (UID: \"abd22381-3367-4aef-a05d-dd895c6de545\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p" Feb 27 17:38:24 crc kubenswrapper[4700]: I0227 17:38:24.276060 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abd22381-3367-4aef-a05d-dd895c6de545-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p\" (UID: \"abd22381-3367-4aef-a05d-dd895c6de545\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p" Feb 27 17:38:24 crc kubenswrapper[4700]: I0227 17:38:24.276241 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abd22381-3367-4aef-a05d-dd895c6de545-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p\" (UID: \"abd22381-3367-4aef-a05d-dd895c6de545\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p" Feb 27 17:38:24 crc kubenswrapper[4700]: I0227 17:38:24.276390 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/abd22381-3367-4aef-a05d-dd895c6de545-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p\" (UID: \"abd22381-3367-4aef-a05d-dd895c6de545\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p" Feb 27 17:38:24 crc kubenswrapper[4700]: I0227 17:38:24.276661 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kqhk\" (UniqueName: \"kubernetes.io/projected/abd22381-3367-4aef-a05d-dd895c6de545-kube-api-access-7kqhk\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p\" (UID: \"abd22381-3367-4aef-a05d-dd895c6de545\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p" Feb 27 17:38:24 crc kubenswrapper[4700]: I0227 17:38:24.380359 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kqhk\" (UniqueName: \"kubernetes.io/projected/abd22381-3367-4aef-a05d-dd895c6de545-kube-api-access-7kqhk\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p\" (UID: \"abd22381-3367-4aef-a05d-dd895c6de545\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p" Feb 27 17:38:24 crc kubenswrapper[4700]: I0227 17:38:24.380494 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/abd22381-3367-4aef-a05d-dd895c6de545-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p\" (UID: \"abd22381-3367-4aef-a05d-dd895c6de545\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p" Feb 27 17:38:24 crc kubenswrapper[4700]: I0227 17:38:24.380571 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abd22381-3367-4aef-a05d-dd895c6de545-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p\" (UID: \"abd22381-3367-4aef-a05d-dd895c6de545\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p" Feb 27 17:38:24 crc kubenswrapper[4700]: I0227 17:38:24.380694 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abd22381-3367-4aef-a05d-dd895c6de545-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p\" (UID: \"abd22381-3367-4aef-a05d-dd895c6de545\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p" Feb 27 17:38:24 crc kubenswrapper[4700]: I0227 17:38:24.380741 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/abd22381-3367-4aef-a05d-dd895c6de545-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p\" (UID: \"abd22381-3367-4aef-a05d-dd895c6de545\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p" Feb 27 17:38:24 crc kubenswrapper[4700]: I0227 17:38:24.386797 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abd22381-3367-4aef-a05d-dd895c6de545-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p\" (UID: \"abd22381-3367-4aef-a05d-dd895c6de545\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p" Feb 27 17:38:24 crc kubenswrapper[4700]: I0227 17:38:24.387784 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abd22381-3367-4aef-a05d-dd895c6de545-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p\" (UID: \"abd22381-3367-4aef-a05d-dd895c6de545\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p" Feb 27 17:38:24 crc kubenswrapper[4700]: I0227 17:38:24.388223 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/abd22381-3367-4aef-a05d-dd895c6de545-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p\" (UID: \"abd22381-3367-4aef-a05d-dd895c6de545\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p" Feb 27 17:38:24 crc kubenswrapper[4700]: I0227 17:38:24.388712 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/abd22381-3367-4aef-a05d-dd895c6de545-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p\" (UID: \"abd22381-3367-4aef-a05d-dd895c6de545\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p" Feb 27 17:38:24 crc kubenswrapper[4700]: I0227 17:38:24.401966 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kqhk\" (UniqueName: \"kubernetes.io/projected/abd22381-3367-4aef-a05d-dd895c6de545-kube-api-access-7kqhk\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p\" (UID: \"abd22381-3367-4aef-a05d-dd895c6de545\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p" Feb 27 17:38:24 crc kubenswrapper[4700]: I0227 17:38:24.535138 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p" Feb 27 17:38:25 crc kubenswrapper[4700]: I0227 17:38:25.203258 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p"] Feb 27 17:38:26 crc kubenswrapper[4700]: I0227 17:38:26.139619 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p" event={"ID":"abd22381-3367-4aef-a05d-dd895c6de545","Type":"ContainerStarted","Data":"53d22f1a20a387a5e781985709b6115c369b61add719a04d9602464cb9782a0e"} Feb 27 17:38:26 crc kubenswrapper[4700]: I0227 17:38:26.140391 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p" event={"ID":"abd22381-3367-4aef-a05d-dd895c6de545","Type":"ContainerStarted","Data":"d6896d08355f4cead3b7747aba2863e435604a7920ad7e65859dc49789ed315a"} Feb 27 17:38:26 crc kubenswrapper[4700]: I0227 17:38:26.176565 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p" podStartSLOduration=1.697640322 podStartE2EDuration="2.176539393s" podCreationTimestamp="2026-02-27 17:38:24 +0000 UTC" firstStartedPulling="2026-02-27 17:38:25.203350228 +0000 UTC m=+2265.188663005" lastFinishedPulling="2026-02-27 17:38:25.682249329 +0000 UTC m=+2265.667562076" observedRunningTime="2026-02-27 17:38:26.163114291 +0000 UTC m=+2266.148427038" watchObservedRunningTime="2026-02-27 17:38:26.176539393 +0000 UTC m=+2266.161852150" Feb 27 17:38:36 crc kubenswrapper[4700]: I0227 17:38:36.410731 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:38:36 crc kubenswrapper[4700]: I0227 17:38:36.411345 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:39:06 crc kubenswrapper[4700]: I0227 17:39:06.411355 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:39:06 crc kubenswrapper[4700]: I0227 17:39:06.412221 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:39:18 crc kubenswrapper[4700]: E0227 17:39:18.823537 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 17:39:18 crc kubenswrapper[4700]: E0227 17:39:18.825959 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 17:39:18 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 17:39:18 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5fjcr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536898-4zjfl_openshift-infra(6c095a55-8ba9-44a9-b4c0-656cf3de7880): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 17:39:18 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 17:39:18 crc kubenswrapper[4700]: E0227 17:39:18.828225 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536898-4zjfl" podUID="6c095a55-8ba9-44a9-b4c0-656cf3de7880" Feb 27 17:39:33 crc kubenswrapper[4700]: E0227 17:39:33.985128 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536898-4zjfl" podUID="6c095a55-8ba9-44a9-b4c0-656cf3de7880" Feb 27 17:39:36 crc kubenswrapper[4700]: I0227 17:39:36.411097 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:39:36 crc kubenswrapper[4700]: I0227 17:39:36.411817 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:39:36 crc kubenswrapper[4700]: I0227 17:39:36.411891 4700 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" Feb 27 17:39:36 crc kubenswrapper[4700]: I0227 17:39:36.413113 4700 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6843fae59a5a033363e409d439168388b8322ceaaef013a2a8f29582e161bd02"} pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 17:39:36 crc kubenswrapper[4700]: I0227 17:39:36.413226 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" containerID="cri-o://6843fae59a5a033363e409d439168388b8322ceaaef013a2a8f29582e161bd02" gracePeriod=600 Feb 27 17:39:36 crc kubenswrapper[4700]: E0227 17:39:36.540450 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:39:37 crc kubenswrapper[4700]: I0227 17:39:37.095410 4700 generic.go:334] "Generic (PLEG): container finished" podID="84b45511-b94c-479f-98db-fd2c4eceec46" containerID="6843fae59a5a033363e409d439168388b8322ceaaef013a2a8f29582e161bd02" exitCode=0 Feb 27 17:39:37 crc kubenswrapper[4700]: I0227 17:39:37.095476 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerDied","Data":"6843fae59a5a033363e409d439168388b8322ceaaef013a2a8f29582e161bd02"} Feb 27 17:39:37 crc kubenswrapper[4700]: I0227 17:39:37.095516 4700 scope.go:117] "RemoveContainer" containerID="d14a07f9a7a8a0cecf47fa738d13cde159084a95d4b48612e05d84dc5f14de89" Feb 27 17:39:37 crc kubenswrapper[4700]: I0227 17:39:37.096330 4700 scope.go:117] "RemoveContainer" containerID="6843fae59a5a033363e409d439168388b8322ceaaef013a2a8f29582e161bd02" Feb 27 17:39:37 crc kubenswrapper[4700]: E0227 17:39:37.096782 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:39:48 crc kubenswrapper[4700]: E0227 17:39:48.809234 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 17:39:48 crc kubenswrapper[4700]: E0227 17:39:48.809964 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 17:39:48 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 17:39:48 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5fjcr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536898-4zjfl_openshift-infra(6c095a55-8ba9-44a9-b4c0-656cf3de7880): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 17:39:48 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 17:39:48 crc kubenswrapper[4700]: E0227 17:39:48.811113 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536898-4zjfl" podUID="6c095a55-8ba9-44a9-b4c0-656cf3de7880" Feb 27 17:39:50 crc kubenswrapper[4700]: I0227 17:39:50.993220 4700 scope.go:117] "RemoveContainer" containerID="6843fae59a5a033363e409d439168388b8322ceaaef013a2a8f29582e161bd02" Feb 27 17:39:50 crc kubenswrapper[4700]: E0227 17:39:50.993684 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:40:00 crc kubenswrapper[4700]: I0227 17:40:00.180570 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536900-xp5cs"] Feb 27 17:40:00 crc kubenswrapper[4700]: I0227 17:40:00.183828 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536900-xp5cs" Feb 27 17:40:00 crc kubenswrapper[4700]: I0227 17:40:00.197874 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536900-xp5cs"] Feb 27 17:40:00 crc kubenswrapper[4700]: I0227 17:40:00.367666 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29ql8\" (UniqueName: \"kubernetes.io/projected/28371b07-fe9e-41c2-bb2d-d2a714510732-kube-api-access-29ql8\") pod \"auto-csr-approver-29536900-xp5cs\" (UID: \"28371b07-fe9e-41c2-bb2d-d2a714510732\") " pod="openshift-infra/auto-csr-approver-29536900-xp5cs" Feb 27 17:40:00 crc kubenswrapper[4700]: I0227 17:40:00.470110 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29ql8\" (UniqueName: \"kubernetes.io/projected/28371b07-fe9e-41c2-bb2d-d2a714510732-kube-api-access-29ql8\") pod \"auto-csr-approver-29536900-xp5cs\" (UID: \"28371b07-fe9e-41c2-bb2d-d2a714510732\") " pod="openshift-infra/auto-csr-approver-29536900-xp5cs" Feb 27 17:40:00 crc kubenswrapper[4700]: I0227 17:40:00.489429 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29ql8\" (UniqueName: \"kubernetes.io/projected/28371b07-fe9e-41c2-bb2d-d2a714510732-kube-api-access-29ql8\") pod \"auto-csr-approver-29536900-xp5cs\" (UID: \"28371b07-fe9e-41c2-bb2d-d2a714510732\") " pod="openshift-infra/auto-csr-approver-29536900-xp5cs" Feb 27 17:40:00 crc kubenswrapper[4700]: I0227 17:40:00.517128 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536900-xp5cs" Feb 27 17:40:01 crc kubenswrapper[4700]: I0227 17:40:01.088256 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536900-xp5cs"] Feb 27 17:40:01 crc kubenswrapper[4700]: I0227 17:40:01.359792 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536900-xp5cs" event={"ID":"28371b07-fe9e-41c2-bb2d-d2a714510732","Type":"ContainerStarted","Data":"3f87737e665b670094b9ce74003208e2c425d0d6ac3a5c0258e0c1cd62fec1a7"} Feb 27 17:40:02 crc kubenswrapper[4700]: E0227 17:40:02.253263 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 17:40:02 crc kubenswrapper[4700]: E0227 17:40:02.253809 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 17:40:02 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 17:40:02 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-29ql8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536900-xp5cs_openshift-infra(28371b07-fe9e-41c2-bb2d-d2a714510732): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 17:40:02 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 17:40:02 crc kubenswrapper[4700]: E0227 17:40:02.255174 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536900-xp5cs" podUID="28371b07-fe9e-41c2-bb2d-d2a714510732" Feb 27 17:40:02 crc kubenswrapper[4700]: E0227 17:40:02.377432 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536900-xp5cs" podUID="28371b07-fe9e-41c2-bb2d-d2a714510732" Feb 27 17:40:02 crc kubenswrapper[4700]: I0227 17:40:02.983103 4700 scope.go:117] "RemoveContainer" containerID="6843fae59a5a033363e409d439168388b8322ceaaef013a2a8f29582e161bd02" Feb 27 17:40:02 crc kubenswrapper[4700]: E0227 17:40:02.983604 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:40:02 crc kubenswrapper[4700]: E0227 17:40:02.984224 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536898-4zjfl" podUID="6c095a55-8ba9-44a9-b4c0-656cf3de7880" Feb 27 17:40:04 crc kubenswrapper[4700]: I0227 17:40:04.132099 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zppm5"] Feb 27 17:40:04 crc kubenswrapper[4700]: I0227 17:40:04.134259 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zppm5" Feb 27 17:40:04 crc kubenswrapper[4700]: I0227 17:40:04.149295 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zppm5"] Feb 27 17:40:04 crc kubenswrapper[4700]: I0227 17:40:04.258331 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46a9c254-bda9-4b7e-a082-7ab137ef99b9-utilities\") pod \"redhat-marketplace-zppm5\" (UID: \"46a9c254-bda9-4b7e-a082-7ab137ef99b9\") " pod="openshift-marketplace/redhat-marketplace-zppm5" Feb 27 17:40:04 crc kubenswrapper[4700]: I0227 17:40:04.258401 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46a9c254-bda9-4b7e-a082-7ab137ef99b9-catalog-content\") pod \"redhat-marketplace-zppm5\" (UID: \"46a9c254-bda9-4b7e-a082-7ab137ef99b9\") " pod="openshift-marketplace/redhat-marketplace-zppm5" Feb 27 17:40:04 crc kubenswrapper[4700]: I0227 17:40:04.259073 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsdjj\" (UniqueName: \"kubernetes.io/projected/46a9c254-bda9-4b7e-a082-7ab137ef99b9-kube-api-access-tsdjj\") pod \"redhat-marketplace-zppm5\" (UID: \"46a9c254-bda9-4b7e-a082-7ab137ef99b9\") " pod="openshift-marketplace/redhat-marketplace-zppm5" Feb 27 17:40:04 crc kubenswrapper[4700]: I0227 17:40:04.361067 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsdjj\" (UniqueName: \"kubernetes.io/projected/46a9c254-bda9-4b7e-a082-7ab137ef99b9-kube-api-access-tsdjj\") pod \"redhat-marketplace-zppm5\" (UID: \"46a9c254-bda9-4b7e-a082-7ab137ef99b9\") " pod="openshift-marketplace/redhat-marketplace-zppm5" Feb 27 17:40:04 crc kubenswrapper[4700]: I0227 17:40:04.361124 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46a9c254-bda9-4b7e-a082-7ab137ef99b9-utilities\") pod \"redhat-marketplace-zppm5\" (UID: \"46a9c254-bda9-4b7e-a082-7ab137ef99b9\") " pod="openshift-marketplace/redhat-marketplace-zppm5" Feb 27 17:40:04 crc kubenswrapper[4700]: I0227 17:40:04.361160 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46a9c254-bda9-4b7e-a082-7ab137ef99b9-catalog-content\") pod \"redhat-marketplace-zppm5\" (UID: \"46a9c254-bda9-4b7e-a082-7ab137ef99b9\") " pod="openshift-marketplace/redhat-marketplace-zppm5" Feb 27 17:40:04 crc kubenswrapper[4700]: I0227 17:40:04.361690 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46a9c254-bda9-4b7e-a082-7ab137ef99b9-catalog-content\") pod \"redhat-marketplace-zppm5\" (UID: \"46a9c254-bda9-4b7e-a082-7ab137ef99b9\") " pod="openshift-marketplace/redhat-marketplace-zppm5" Feb 27 17:40:04 crc kubenswrapper[4700]: I0227 17:40:04.361824 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46a9c254-bda9-4b7e-a082-7ab137ef99b9-utilities\") pod \"redhat-marketplace-zppm5\" (UID: \"46a9c254-bda9-4b7e-a082-7ab137ef99b9\") " pod="openshift-marketplace/redhat-marketplace-zppm5" Feb 27 17:40:04 crc kubenswrapper[4700]: I0227 17:40:04.393732 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsdjj\" (UniqueName: \"kubernetes.io/projected/46a9c254-bda9-4b7e-a082-7ab137ef99b9-kube-api-access-tsdjj\") pod \"redhat-marketplace-zppm5\" (UID: \"46a9c254-bda9-4b7e-a082-7ab137ef99b9\") " pod="openshift-marketplace/redhat-marketplace-zppm5" Feb 27 17:40:04 crc kubenswrapper[4700]: I0227 17:40:04.463519 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zppm5" Feb 27 17:40:05 crc kubenswrapper[4700]: I0227 17:40:05.034604 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zppm5"] Feb 27 17:40:05 crc kubenswrapper[4700]: I0227 17:40:05.410000 4700 generic.go:334] "Generic (PLEG): container finished" podID="46a9c254-bda9-4b7e-a082-7ab137ef99b9" containerID="372198118beebb754dd568de514e36ac9f4a963c6a0d8cff5bf807cb0631671d" exitCode=0 Feb 27 17:40:05 crc kubenswrapper[4700]: I0227 17:40:05.410059 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zppm5" event={"ID":"46a9c254-bda9-4b7e-a082-7ab137ef99b9","Type":"ContainerDied","Data":"372198118beebb754dd568de514e36ac9f4a963c6a0d8cff5bf807cb0631671d"} Feb 27 17:40:05 crc kubenswrapper[4700]: I0227 17:40:05.410335 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zppm5" event={"ID":"46a9c254-bda9-4b7e-a082-7ab137ef99b9","Type":"ContainerStarted","Data":"3fc5307928ff5142eee1c66b765236d71212ce7d56049b8a9c0fef01990a2fcb"} Feb 27 17:40:05 crc kubenswrapper[4700]: E0227 17:40:05.411178 4700 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46a9c254_bda9_4b7e_a082_7ab137ef99b9.slice/crio-conmon-372198118beebb754dd568de514e36ac9f4a963c6a0d8cff5bf807cb0631671d.scope\": RecentStats: unable to find data in memory cache]" Feb 27 17:40:06 crc kubenswrapper[4700]: E0227 17:40:06.080182 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 17:40:06 crc kubenswrapper[4700]: E0227 17:40:06.080675 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tsdjj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-zppm5_openshift-marketplace(46a9c254-bda9-4b7e-a082-7ab137ef99b9): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 17:40:06 crc kubenswrapper[4700]: E0227 17:40:06.081869 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-marketplace-zppm5" podUID="46a9c254-bda9-4b7e-a082-7ab137ef99b9" Feb 27 17:40:06 crc kubenswrapper[4700]: E0227 17:40:06.430295 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-zppm5" podUID="46a9c254-bda9-4b7e-a082-7ab137ef99b9" Feb 27 17:40:12 crc kubenswrapper[4700]: I0227 17:40:12.907015 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-65457d8799-dclxm" podUID="e43c5d5e-63ef-45b9-af4e-627a035b376e" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 502" Feb 27 17:40:14 crc kubenswrapper[4700]: E0227 17:40:14.858539 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 17:40:14 crc kubenswrapper[4700]: E0227 17:40:14.859169 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 17:40:14 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 17:40:14 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-29ql8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536900-xp5cs_openshift-infra(28371b07-fe9e-41c2-bb2d-d2a714510732): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 17:40:14 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 17:40:14 crc kubenswrapper[4700]: E0227 17:40:14.860334 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536900-xp5cs" podUID="28371b07-fe9e-41c2-bb2d-d2a714510732" Feb 27 17:40:17 crc kubenswrapper[4700]: I0227 17:40:17.001911 4700 scope.go:117] "RemoveContainer" containerID="6843fae59a5a033363e409d439168388b8322ceaaef013a2a8f29582e161bd02" Feb 27 17:40:17 crc kubenswrapper[4700]: E0227 17:40:17.003420 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536898-4zjfl" podUID="6c095a55-8ba9-44a9-b4c0-656cf3de7880" Feb 27 17:40:17 crc kubenswrapper[4700]: E0227 17:40:17.003902 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:40:21 crc kubenswrapper[4700]: E0227 17:40:21.619434 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 17:40:21 crc kubenswrapper[4700]: E0227 17:40:21.620088 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tsdjj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-zppm5_openshift-marketplace(46a9c254-bda9-4b7e-a082-7ab137ef99b9): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 17:40:21 crc kubenswrapper[4700]: E0227 17:40:21.621334 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-marketplace-zppm5" podUID="46a9c254-bda9-4b7e-a082-7ab137ef99b9" Feb 27 17:40:28 crc kubenswrapper[4700]: E0227 17:40:27.984455 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536898-4zjfl" podUID="6c095a55-8ba9-44a9-b4c0-656cf3de7880" Feb 27 17:40:28 crc kubenswrapper[4700]: E0227 17:40:28.983806 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536900-xp5cs" podUID="28371b07-fe9e-41c2-bb2d-d2a714510732" Feb 27 17:40:29 crc kubenswrapper[4700]: I0227 17:40:29.981892 4700 scope.go:117] "RemoveContainer" containerID="6843fae59a5a033363e409d439168388b8322ceaaef013a2a8f29582e161bd02" Feb 27 17:40:29 crc kubenswrapper[4700]: E0227 17:40:29.982700 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:40:34 crc kubenswrapper[4700]: E0227 17:40:34.984742 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-zppm5" podUID="46a9c254-bda9-4b7e-a082-7ab137ef99b9" Feb 27 17:40:40 crc kubenswrapper[4700]: I0227 17:40:40.846527 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536898-4zjfl" event={"ID":"6c095a55-8ba9-44a9-b4c0-656cf3de7880","Type":"ContainerStarted","Data":"15995fc6325bfa84bcf6c7aed97aea03c4466ba91003e4703e6648dd4081aa78"} Feb 27 17:40:40 crc kubenswrapper[4700]: I0227 17:40:40.876098 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536898-4zjfl" podStartSLOduration=1.6403639810000001 podStartE2EDuration="2m40.876068825s" podCreationTimestamp="2026-02-27 17:38:00 +0000 UTC" firstStartedPulling="2026-02-27 17:38:01.025984622 +0000 UTC m=+2241.011297369" lastFinishedPulling="2026-02-27 17:40:40.261689436 +0000 UTC m=+2400.247002213" observedRunningTime="2026-02-27 17:40:40.864907383 +0000 UTC m=+2400.850220160" watchObservedRunningTime="2026-02-27 17:40:40.876068825 +0000 UTC m=+2400.861381602" Feb 27 17:40:41 crc kubenswrapper[4700]: I0227 17:40:41.856048 4700 generic.go:334] "Generic (PLEG): container finished" podID="6c095a55-8ba9-44a9-b4c0-656cf3de7880" containerID="15995fc6325bfa84bcf6c7aed97aea03c4466ba91003e4703e6648dd4081aa78" exitCode=0 Feb 27 17:40:41 crc kubenswrapper[4700]: I0227 17:40:41.856141 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536898-4zjfl" event={"ID":"6c095a55-8ba9-44a9-b4c0-656cf3de7880","Type":"ContainerDied","Data":"15995fc6325bfa84bcf6c7aed97aea03c4466ba91003e4703e6648dd4081aa78"} Feb 27 17:40:43 crc kubenswrapper[4700]: I0227 17:40:43.357078 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536898-4zjfl" Feb 27 17:40:43 crc kubenswrapper[4700]: I0227 17:40:43.518725 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5fjcr\" (UniqueName: \"kubernetes.io/projected/6c095a55-8ba9-44a9-b4c0-656cf3de7880-kube-api-access-5fjcr\") pod \"6c095a55-8ba9-44a9-b4c0-656cf3de7880\" (UID: \"6c095a55-8ba9-44a9-b4c0-656cf3de7880\") " Feb 27 17:40:43 crc kubenswrapper[4700]: I0227 17:40:43.525191 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c095a55-8ba9-44a9-b4c0-656cf3de7880-kube-api-access-5fjcr" (OuterVolumeSpecName: "kube-api-access-5fjcr") pod "6c095a55-8ba9-44a9-b4c0-656cf3de7880" (UID: "6c095a55-8ba9-44a9-b4c0-656cf3de7880"). InnerVolumeSpecName "kube-api-access-5fjcr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:40:43 crc kubenswrapper[4700]: I0227 17:40:43.622082 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5fjcr\" (UniqueName: \"kubernetes.io/projected/6c095a55-8ba9-44a9-b4c0-656cf3de7880-kube-api-access-5fjcr\") on node \"crc\" DevicePath \"\"" Feb 27 17:40:43 crc kubenswrapper[4700]: I0227 17:40:43.981956 4700 scope.go:117] "RemoveContainer" containerID="6843fae59a5a033363e409d439168388b8322ceaaef013a2a8f29582e161bd02" Feb 27 17:40:43 crc kubenswrapper[4700]: E0227 17:40:43.982859 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:40:44 crc kubenswrapper[4700]: I0227 17:40:44.040831 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536898-4zjfl" event={"ID":"6c095a55-8ba9-44a9-b4c0-656cf3de7880","Type":"ContainerDied","Data":"1ac9aa97e4610ebb53f8ddcaca87ecedf0f5d7ecf89e380e1ff8516bdcdd52f1"} Feb 27 17:40:44 crc kubenswrapper[4700]: I0227 17:40:44.040879 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ac9aa97e4610ebb53f8ddcaca87ecedf0f5d7ecf89e380e1ff8516bdcdd52f1" Feb 27 17:40:44 crc kubenswrapper[4700]: I0227 17:40:44.040885 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536898-4zjfl" Feb 27 17:40:44 crc kubenswrapper[4700]: I0227 17:40:44.089056 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536892-ndtg4"] Feb 27 17:40:44 crc kubenswrapper[4700]: I0227 17:40:44.107585 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536892-ndtg4"] Feb 27 17:40:45 crc kubenswrapper[4700]: I0227 17:40:45.000168 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a6c1496-be0e-4733-8176-19c62bbc329e" path="/var/lib/kubelet/pods/5a6c1496-be0e-4733-8176-19c62bbc329e/volumes" Feb 27 17:40:46 crc kubenswrapper[4700]: I0227 17:40:46.060782 4700 generic.go:334] "Generic (PLEG): container finished" podID="28371b07-fe9e-41c2-bb2d-d2a714510732" containerID="bf1f073c4ee8e6cebb4d8470d6277d6ce6aebb31fd4dd9396283535b9ee76e1c" exitCode=0 Feb 27 17:40:46 crc kubenswrapper[4700]: I0227 17:40:46.060809 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536900-xp5cs" event={"ID":"28371b07-fe9e-41c2-bb2d-d2a714510732","Type":"ContainerDied","Data":"bf1f073c4ee8e6cebb4d8470d6277d6ce6aebb31fd4dd9396283535b9ee76e1c"} Feb 27 17:40:46 crc kubenswrapper[4700]: E0227 17:40:46.740375 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 17:40:46 crc kubenswrapper[4700]: E0227 17:40:46.740833 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tsdjj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-zppm5_openshift-marketplace(46a9c254-bda9-4b7e-a082-7ab137ef99b9): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 17:40:46 crc kubenswrapper[4700]: E0227 17:40:46.742081 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-marketplace-zppm5" podUID="46a9c254-bda9-4b7e-a082-7ab137ef99b9" Feb 27 17:40:47 crc kubenswrapper[4700]: I0227 17:40:47.496295 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536900-xp5cs" Feb 27 17:40:47 crc kubenswrapper[4700]: I0227 17:40:47.516046 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29ql8\" (UniqueName: \"kubernetes.io/projected/28371b07-fe9e-41c2-bb2d-d2a714510732-kube-api-access-29ql8\") pod \"28371b07-fe9e-41c2-bb2d-d2a714510732\" (UID: \"28371b07-fe9e-41c2-bb2d-d2a714510732\") " Feb 27 17:40:47 crc kubenswrapper[4700]: I0227 17:40:47.524341 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28371b07-fe9e-41c2-bb2d-d2a714510732-kube-api-access-29ql8" (OuterVolumeSpecName: "kube-api-access-29ql8") pod "28371b07-fe9e-41c2-bb2d-d2a714510732" (UID: "28371b07-fe9e-41c2-bb2d-d2a714510732"). InnerVolumeSpecName "kube-api-access-29ql8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:40:47 crc kubenswrapper[4700]: I0227 17:40:47.617825 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29ql8\" (UniqueName: \"kubernetes.io/projected/28371b07-fe9e-41c2-bb2d-d2a714510732-kube-api-access-29ql8\") on node \"crc\" DevicePath \"\"" Feb 27 17:40:48 crc kubenswrapper[4700]: I0227 17:40:48.086956 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536900-xp5cs" event={"ID":"28371b07-fe9e-41c2-bb2d-d2a714510732","Type":"ContainerDied","Data":"3f87737e665b670094b9ce74003208e2c425d0d6ac3a5c0258e0c1cd62fec1a7"} Feb 27 17:40:48 crc kubenswrapper[4700]: I0227 17:40:48.087704 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f87737e665b670094b9ce74003208e2c425d0d6ac3a5c0258e0c1cd62fec1a7" Feb 27 17:40:48 crc kubenswrapper[4700]: I0227 17:40:48.087064 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536900-xp5cs" Feb 27 17:40:48 crc kubenswrapper[4700]: I0227 17:40:48.583476 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536894-59kkv"] Feb 27 17:40:48 crc kubenswrapper[4700]: I0227 17:40:48.597582 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536894-59kkv"] Feb 27 17:40:48 crc kubenswrapper[4700]: I0227 17:40:48.993992 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e3527f7-bc5c-4321-bc5c-0fe5c8284b4b" path="/var/lib/kubelet/pods/3e3527f7-bc5c-4321-bc5c-0fe5c8284b4b/volumes" Feb 27 17:40:54 crc kubenswrapper[4700]: I0227 17:40:54.982415 4700 scope.go:117] "RemoveContainer" containerID="6843fae59a5a033363e409d439168388b8322ceaaef013a2a8f29582e161bd02" Feb 27 17:40:54 crc kubenswrapper[4700]: E0227 17:40:54.983521 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:41:00 crc kubenswrapper[4700]: E0227 17:41:00.998215 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-zppm5" podUID="46a9c254-bda9-4b7e-a082-7ab137ef99b9" Feb 27 17:41:09 crc kubenswrapper[4700]: I0227 17:41:09.545662 4700 scope.go:117] "RemoveContainer" containerID="c681a8e61f5a02078b4b5673c622a07ce6ededd6ccc76440bac388e04d9d87d1" Feb 27 17:41:09 crc kubenswrapper[4700]: I0227 17:41:09.633582 4700 scope.go:117] "RemoveContainer" containerID="3ed6861afca6d101aa226387a12b0eeb3f4daadb50454b739b39d920a10f0cf2" Feb 27 17:41:09 crc kubenswrapper[4700]: I0227 17:41:09.981569 4700 scope.go:117] "RemoveContainer" containerID="6843fae59a5a033363e409d439168388b8322ceaaef013a2a8f29582e161bd02" Feb 27 17:41:09 crc kubenswrapper[4700]: E0227 17:41:09.982103 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:41:13 crc kubenswrapper[4700]: E0227 17:41:13.984043 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-zppm5" podUID="46a9c254-bda9-4b7e-a082-7ab137ef99b9" Feb 27 17:41:24 crc kubenswrapper[4700]: I0227 17:41:24.981159 4700 scope.go:117] "RemoveContainer" containerID="6843fae59a5a033363e409d439168388b8322ceaaef013a2a8f29582e161bd02" Feb 27 17:41:24 crc kubenswrapper[4700]: E0227 17:41:24.983222 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:41:24 crc kubenswrapper[4700]: E0227 17:41:24.983686 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-zppm5" podUID="46a9c254-bda9-4b7e-a082-7ab137ef99b9" Feb 27 17:41:35 crc kubenswrapper[4700]: I0227 17:41:35.982443 4700 scope.go:117] "RemoveContainer" containerID="6843fae59a5a033363e409d439168388b8322ceaaef013a2a8f29582e161bd02" Feb 27 17:41:35 crc kubenswrapper[4700]: E0227 17:41:35.984907 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:41:39 crc kubenswrapper[4700]: I0227 17:41:39.728208 4700 generic.go:334] "Generic (PLEG): container finished" podID="46a9c254-bda9-4b7e-a082-7ab137ef99b9" containerID="cdf7f2d8bde48e743eb34b6df7e0873415da2c16a4020cbfbd9e93a23a9c9a47" exitCode=0 Feb 27 17:41:39 crc kubenswrapper[4700]: I0227 17:41:39.728345 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zppm5" event={"ID":"46a9c254-bda9-4b7e-a082-7ab137ef99b9","Type":"ContainerDied","Data":"cdf7f2d8bde48e743eb34b6df7e0873415da2c16a4020cbfbd9e93a23a9c9a47"} Feb 27 17:41:40 crc kubenswrapper[4700]: I0227 17:41:40.744164 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zppm5" event={"ID":"46a9c254-bda9-4b7e-a082-7ab137ef99b9","Type":"ContainerStarted","Data":"49b0fde47b7cb027c1910a44f91ae8b636f7c84e05dfbd0bd01e5b8aeb41204f"} Feb 27 17:41:40 crc kubenswrapper[4700]: I0227 17:41:40.791491 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zppm5" podStartSLOduration=1.94243274 podStartE2EDuration="1m36.791441802s" podCreationTimestamp="2026-02-27 17:40:04 +0000 UTC" firstStartedPulling="2026-02-27 17:40:05.413725995 +0000 UTC m=+2365.399038742" lastFinishedPulling="2026-02-27 17:41:40.262735017 +0000 UTC m=+2460.248047804" observedRunningTime="2026-02-27 17:41:40.772141727 +0000 UTC m=+2460.757454504" watchObservedRunningTime="2026-02-27 17:41:40.791441802 +0000 UTC m=+2460.776754579" Feb 27 17:41:44 crc kubenswrapper[4700]: I0227 17:41:44.464710 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zppm5" Feb 27 17:41:44 crc kubenswrapper[4700]: I0227 17:41:44.465586 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zppm5" Feb 27 17:41:44 crc kubenswrapper[4700]: I0227 17:41:44.522691 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zppm5" Feb 27 17:41:45 crc kubenswrapper[4700]: I0227 17:41:45.136286 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zppm5" Feb 27 17:41:45 crc kubenswrapper[4700]: I0227 17:41:45.205311 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zppm5"] Feb 27 17:41:47 crc kubenswrapper[4700]: I0227 17:41:47.091771 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zppm5" podUID="46a9c254-bda9-4b7e-a082-7ab137ef99b9" containerName="registry-server" containerID="cri-o://49b0fde47b7cb027c1910a44f91ae8b636f7c84e05dfbd0bd01e5b8aeb41204f" gracePeriod=2 Feb 27 17:41:47 crc kubenswrapper[4700]: I0227 17:41:47.590110 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zppm5" Feb 27 17:41:47 crc kubenswrapper[4700]: I0227 17:41:47.722669 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46a9c254-bda9-4b7e-a082-7ab137ef99b9-catalog-content\") pod \"46a9c254-bda9-4b7e-a082-7ab137ef99b9\" (UID: \"46a9c254-bda9-4b7e-a082-7ab137ef99b9\") " Feb 27 17:41:47 crc kubenswrapper[4700]: I0227 17:41:47.723194 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46a9c254-bda9-4b7e-a082-7ab137ef99b9-utilities\") pod \"46a9c254-bda9-4b7e-a082-7ab137ef99b9\" (UID: \"46a9c254-bda9-4b7e-a082-7ab137ef99b9\") " Feb 27 17:41:47 crc kubenswrapper[4700]: I0227 17:41:47.724099 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46a9c254-bda9-4b7e-a082-7ab137ef99b9-utilities" (OuterVolumeSpecName: "utilities") pod "46a9c254-bda9-4b7e-a082-7ab137ef99b9" (UID: "46a9c254-bda9-4b7e-a082-7ab137ef99b9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:41:47 crc kubenswrapper[4700]: I0227 17:41:47.724198 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tsdjj\" (UniqueName: \"kubernetes.io/projected/46a9c254-bda9-4b7e-a082-7ab137ef99b9-kube-api-access-tsdjj\") pod \"46a9c254-bda9-4b7e-a082-7ab137ef99b9\" (UID: \"46a9c254-bda9-4b7e-a082-7ab137ef99b9\") " Feb 27 17:41:47 crc kubenswrapper[4700]: I0227 17:41:47.725243 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46a9c254-bda9-4b7e-a082-7ab137ef99b9-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:41:47 crc kubenswrapper[4700]: I0227 17:41:47.730193 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46a9c254-bda9-4b7e-a082-7ab137ef99b9-kube-api-access-tsdjj" (OuterVolumeSpecName: "kube-api-access-tsdjj") pod "46a9c254-bda9-4b7e-a082-7ab137ef99b9" (UID: "46a9c254-bda9-4b7e-a082-7ab137ef99b9"). InnerVolumeSpecName "kube-api-access-tsdjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:41:47 crc kubenswrapper[4700]: I0227 17:41:47.745906 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46a9c254-bda9-4b7e-a082-7ab137ef99b9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "46a9c254-bda9-4b7e-a082-7ab137ef99b9" (UID: "46a9c254-bda9-4b7e-a082-7ab137ef99b9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:41:47 crc kubenswrapper[4700]: I0227 17:41:47.826837 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46a9c254-bda9-4b7e-a082-7ab137ef99b9-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:41:47 crc kubenswrapper[4700]: I0227 17:41:47.826875 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tsdjj\" (UniqueName: \"kubernetes.io/projected/46a9c254-bda9-4b7e-a082-7ab137ef99b9-kube-api-access-tsdjj\") on node \"crc\" DevicePath \"\"" Feb 27 17:41:47 crc kubenswrapper[4700]: I0227 17:41:47.981325 4700 scope.go:117] "RemoveContainer" containerID="6843fae59a5a033363e409d439168388b8322ceaaef013a2a8f29582e161bd02" Feb 27 17:41:47 crc kubenswrapper[4700]: E0227 17:41:47.981842 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:41:48 crc kubenswrapper[4700]: I0227 17:41:48.109174 4700 generic.go:334] "Generic (PLEG): container finished" podID="46a9c254-bda9-4b7e-a082-7ab137ef99b9" containerID="49b0fde47b7cb027c1910a44f91ae8b636f7c84e05dfbd0bd01e5b8aeb41204f" exitCode=0 Feb 27 17:41:48 crc kubenswrapper[4700]: I0227 17:41:48.109330 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zppm5" Feb 27 17:41:48 crc kubenswrapper[4700]: I0227 17:41:48.109359 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zppm5" event={"ID":"46a9c254-bda9-4b7e-a082-7ab137ef99b9","Type":"ContainerDied","Data":"49b0fde47b7cb027c1910a44f91ae8b636f7c84e05dfbd0bd01e5b8aeb41204f"} Feb 27 17:41:48 crc kubenswrapper[4700]: I0227 17:41:48.110986 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zppm5" event={"ID":"46a9c254-bda9-4b7e-a082-7ab137ef99b9","Type":"ContainerDied","Data":"3fc5307928ff5142eee1c66b765236d71212ce7d56049b8a9c0fef01990a2fcb"} Feb 27 17:41:48 crc kubenswrapper[4700]: I0227 17:41:48.111056 4700 scope.go:117] "RemoveContainer" containerID="49b0fde47b7cb027c1910a44f91ae8b636f7c84e05dfbd0bd01e5b8aeb41204f" Feb 27 17:41:48 crc kubenswrapper[4700]: I0227 17:41:48.153564 4700 scope.go:117] "RemoveContainer" containerID="cdf7f2d8bde48e743eb34b6df7e0873415da2c16a4020cbfbd9e93a23a9c9a47" Feb 27 17:41:48 crc kubenswrapper[4700]: I0227 17:41:48.175047 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zppm5"] Feb 27 17:41:48 crc kubenswrapper[4700]: I0227 17:41:48.185373 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zppm5"] Feb 27 17:41:48 crc kubenswrapper[4700]: I0227 17:41:48.197299 4700 scope.go:117] "RemoveContainer" containerID="372198118beebb754dd568de514e36ac9f4a963c6a0d8cff5bf807cb0631671d" Feb 27 17:41:48 crc kubenswrapper[4700]: I0227 17:41:48.254863 4700 scope.go:117] "RemoveContainer" containerID="49b0fde47b7cb027c1910a44f91ae8b636f7c84e05dfbd0bd01e5b8aeb41204f" Feb 27 17:41:48 crc kubenswrapper[4700]: E0227 17:41:48.255247 4700 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46a9c254_bda9_4b7e_a082_7ab137ef99b9.slice\": RecentStats: unable to find data in memory cache]" Feb 27 17:41:48 crc kubenswrapper[4700]: E0227 17:41:48.255307 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49b0fde47b7cb027c1910a44f91ae8b636f7c84e05dfbd0bd01e5b8aeb41204f\": container with ID starting with 49b0fde47b7cb027c1910a44f91ae8b636f7c84e05dfbd0bd01e5b8aeb41204f not found: ID does not exist" containerID="49b0fde47b7cb027c1910a44f91ae8b636f7c84e05dfbd0bd01e5b8aeb41204f" Feb 27 17:41:48 crc kubenswrapper[4700]: I0227 17:41:48.255359 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49b0fde47b7cb027c1910a44f91ae8b636f7c84e05dfbd0bd01e5b8aeb41204f"} err="failed to get container status \"49b0fde47b7cb027c1910a44f91ae8b636f7c84e05dfbd0bd01e5b8aeb41204f\": rpc error: code = NotFound desc = could not find container \"49b0fde47b7cb027c1910a44f91ae8b636f7c84e05dfbd0bd01e5b8aeb41204f\": container with ID starting with 49b0fde47b7cb027c1910a44f91ae8b636f7c84e05dfbd0bd01e5b8aeb41204f not found: ID does not exist" Feb 27 17:41:48 crc kubenswrapper[4700]: I0227 17:41:48.255393 4700 scope.go:117] "RemoveContainer" containerID="cdf7f2d8bde48e743eb34b6df7e0873415da2c16a4020cbfbd9e93a23a9c9a47" Feb 27 17:41:48 crc kubenswrapper[4700]: E0227 17:41:48.257454 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cdf7f2d8bde48e743eb34b6df7e0873415da2c16a4020cbfbd9e93a23a9c9a47\": container with ID starting with cdf7f2d8bde48e743eb34b6df7e0873415da2c16a4020cbfbd9e93a23a9c9a47 not found: ID does not exist" containerID="cdf7f2d8bde48e743eb34b6df7e0873415da2c16a4020cbfbd9e93a23a9c9a47" Feb 27 17:41:48 crc kubenswrapper[4700]: I0227 17:41:48.257625 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdf7f2d8bde48e743eb34b6df7e0873415da2c16a4020cbfbd9e93a23a9c9a47"} err="failed to get container status \"cdf7f2d8bde48e743eb34b6df7e0873415da2c16a4020cbfbd9e93a23a9c9a47\": rpc error: code = NotFound desc = could not find container \"cdf7f2d8bde48e743eb34b6df7e0873415da2c16a4020cbfbd9e93a23a9c9a47\": container with ID starting with cdf7f2d8bde48e743eb34b6df7e0873415da2c16a4020cbfbd9e93a23a9c9a47 not found: ID does not exist" Feb 27 17:41:48 crc kubenswrapper[4700]: I0227 17:41:48.257653 4700 scope.go:117] "RemoveContainer" containerID="372198118beebb754dd568de514e36ac9f4a963c6a0d8cff5bf807cb0631671d" Feb 27 17:41:48 crc kubenswrapper[4700]: E0227 17:41:48.257985 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"372198118beebb754dd568de514e36ac9f4a963c6a0d8cff5bf807cb0631671d\": container with ID starting with 372198118beebb754dd568de514e36ac9f4a963c6a0d8cff5bf807cb0631671d not found: ID does not exist" containerID="372198118beebb754dd568de514e36ac9f4a963c6a0d8cff5bf807cb0631671d" Feb 27 17:41:48 crc kubenswrapper[4700]: I0227 17:41:48.258034 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"372198118beebb754dd568de514e36ac9f4a963c6a0d8cff5bf807cb0631671d"} err="failed to get container status \"372198118beebb754dd568de514e36ac9f4a963c6a0d8cff5bf807cb0631671d\": rpc error: code = NotFound desc = could not find container \"372198118beebb754dd568de514e36ac9f4a963c6a0d8cff5bf807cb0631671d\": container with ID starting with 372198118beebb754dd568de514e36ac9f4a963c6a0d8cff5bf807cb0631671d not found: ID does not exist" Feb 27 17:41:49 crc kubenswrapper[4700]: I0227 17:41:49.002993 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46a9c254-bda9-4b7e-a082-7ab137ef99b9" path="/var/lib/kubelet/pods/46a9c254-bda9-4b7e-a082-7ab137ef99b9/volumes" Feb 27 17:42:00 crc kubenswrapper[4700]: I0227 17:42:00.183715 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536902-vfh4z"] Feb 27 17:42:00 crc kubenswrapper[4700]: E0227 17:42:00.184773 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c095a55-8ba9-44a9-b4c0-656cf3de7880" containerName="oc" Feb 27 17:42:00 crc kubenswrapper[4700]: I0227 17:42:00.184790 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c095a55-8ba9-44a9-b4c0-656cf3de7880" containerName="oc" Feb 27 17:42:00 crc kubenswrapper[4700]: E0227 17:42:00.184813 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46a9c254-bda9-4b7e-a082-7ab137ef99b9" containerName="extract-content" Feb 27 17:42:00 crc kubenswrapper[4700]: I0227 17:42:00.184823 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="46a9c254-bda9-4b7e-a082-7ab137ef99b9" containerName="extract-content" Feb 27 17:42:00 crc kubenswrapper[4700]: E0227 17:42:00.184846 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28371b07-fe9e-41c2-bb2d-d2a714510732" containerName="oc" Feb 27 17:42:00 crc kubenswrapper[4700]: I0227 17:42:00.184856 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="28371b07-fe9e-41c2-bb2d-d2a714510732" containerName="oc" Feb 27 17:42:00 crc kubenswrapper[4700]: E0227 17:42:00.184868 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46a9c254-bda9-4b7e-a082-7ab137ef99b9" containerName="registry-server" Feb 27 17:42:00 crc kubenswrapper[4700]: I0227 17:42:00.184876 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="46a9c254-bda9-4b7e-a082-7ab137ef99b9" containerName="registry-server" Feb 27 17:42:00 crc kubenswrapper[4700]: E0227 17:42:00.184896 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46a9c254-bda9-4b7e-a082-7ab137ef99b9" containerName="extract-utilities" Feb 27 17:42:00 crc kubenswrapper[4700]: I0227 17:42:00.184904 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="46a9c254-bda9-4b7e-a082-7ab137ef99b9" containerName="extract-utilities" Feb 27 17:42:00 crc kubenswrapper[4700]: I0227 17:42:00.185148 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="46a9c254-bda9-4b7e-a082-7ab137ef99b9" containerName="registry-server" Feb 27 17:42:00 crc kubenswrapper[4700]: I0227 17:42:00.185206 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="28371b07-fe9e-41c2-bb2d-d2a714510732" containerName="oc" Feb 27 17:42:00 crc kubenswrapper[4700]: I0227 17:42:00.185223 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c095a55-8ba9-44a9-b4c0-656cf3de7880" containerName="oc" Feb 27 17:42:00 crc kubenswrapper[4700]: I0227 17:42:00.186100 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536902-vfh4z" Feb 27 17:42:00 crc kubenswrapper[4700]: I0227 17:42:00.190238 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:42:00 crc kubenswrapper[4700]: I0227 17:42:00.190615 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:42:00 crc kubenswrapper[4700]: I0227 17:42:00.191538 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d62kc" Feb 27 17:42:00 crc kubenswrapper[4700]: I0227 17:42:00.192613 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536902-vfh4z"] Feb 27 17:42:00 crc kubenswrapper[4700]: I0227 17:42:00.323957 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmwct\" (UniqueName: \"kubernetes.io/projected/82ea0ed5-7fde-447d-8fb4-780ad1fa6215-kube-api-access-vmwct\") pod \"auto-csr-approver-29536902-vfh4z\" (UID: \"82ea0ed5-7fde-447d-8fb4-780ad1fa6215\") " pod="openshift-infra/auto-csr-approver-29536902-vfh4z" Feb 27 17:42:00 crc kubenswrapper[4700]: I0227 17:42:00.426314 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmwct\" (UniqueName: \"kubernetes.io/projected/82ea0ed5-7fde-447d-8fb4-780ad1fa6215-kube-api-access-vmwct\") pod \"auto-csr-approver-29536902-vfh4z\" (UID: \"82ea0ed5-7fde-447d-8fb4-780ad1fa6215\") " pod="openshift-infra/auto-csr-approver-29536902-vfh4z" Feb 27 17:42:00 crc kubenswrapper[4700]: I0227 17:42:00.449895 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmwct\" (UniqueName: \"kubernetes.io/projected/82ea0ed5-7fde-447d-8fb4-780ad1fa6215-kube-api-access-vmwct\") pod \"auto-csr-approver-29536902-vfh4z\" (UID: \"82ea0ed5-7fde-447d-8fb4-780ad1fa6215\") " pod="openshift-infra/auto-csr-approver-29536902-vfh4z" Feb 27 17:42:00 crc kubenswrapper[4700]: I0227 17:42:00.526071 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536902-vfh4z" Feb 27 17:42:01 crc kubenswrapper[4700]: I0227 17:42:01.004756 4700 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 17:42:01 crc kubenswrapper[4700]: I0227 17:42:01.017564 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536902-vfh4z"] Feb 27 17:42:01 crc kubenswrapper[4700]: I0227 17:42:01.257675 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536902-vfh4z" event={"ID":"82ea0ed5-7fde-447d-8fb4-780ad1fa6215","Type":"ContainerStarted","Data":"00c1fd6ff3c849c12c56fde5d291b7257f41f8d3630ec96ec45bcce491ea389f"} Feb 27 17:42:02 crc kubenswrapper[4700]: E0227 17:42:02.365412 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 17:42:02 crc kubenswrapper[4700]: E0227 17:42:02.365938 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 17:42:02 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 17:42:02 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vmwct,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536902-vfh4z_openshift-infra(82ea0ed5-7fde-447d-8fb4-780ad1fa6215): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 17:42:02 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 17:42:02 crc kubenswrapper[4700]: E0227 17:42:02.367193 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536902-vfh4z" podUID="82ea0ed5-7fde-447d-8fb4-780ad1fa6215" Feb 27 17:42:02 crc kubenswrapper[4700]: I0227 17:42:02.982907 4700 scope.go:117] "RemoveContainer" containerID="6843fae59a5a033363e409d439168388b8322ceaaef013a2a8f29582e161bd02" Feb 27 17:42:02 crc kubenswrapper[4700]: E0227 17:42:02.983401 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:42:03 crc kubenswrapper[4700]: E0227 17:42:03.281690 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536902-vfh4z" podUID="82ea0ed5-7fde-447d-8fb4-780ad1fa6215" Feb 27 17:42:14 crc kubenswrapper[4700]: I0227 17:42:14.982095 4700 scope.go:117] "RemoveContainer" containerID="6843fae59a5a033363e409d439168388b8322ceaaef013a2a8f29582e161bd02" Feb 27 17:42:14 crc kubenswrapper[4700]: E0227 17:42:14.983111 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:42:17 crc kubenswrapper[4700]: E0227 17:42:17.060628 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 17:42:17 crc kubenswrapper[4700]: E0227 17:42:17.061615 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 17:42:17 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 17:42:17 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vmwct,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536902-vfh4z_openshift-infra(82ea0ed5-7fde-447d-8fb4-780ad1fa6215): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 17:42:17 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 17:42:17 crc kubenswrapper[4700]: E0227 17:42:17.062930 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536902-vfh4z" podUID="82ea0ed5-7fde-447d-8fb4-780ad1fa6215" Feb 27 17:42:27 crc kubenswrapper[4700]: E0227 17:42:27.985239 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536902-vfh4z" podUID="82ea0ed5-7fde-447d-8fb4-780ad1fa6215" Feb 27 17:42:28 crc kubenswrapper[4700]: I0227 17:42:28.981238 4700 scope.go:117] "RemoveContainer" containerID="6843fae59a5a033363e409d439168388b8322ceaaef013a2a8f29582e161bd02" Feb 27 17:42:28 crc kubenswrapper[4700]: E0227 17:42:28.981999 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:42:40 crc kubenswrapper[4700]: I0227 17:42:40.991120 4700 scope.go:117] "RemoveContainer" containerID="6843fae59a5a033363e409d439168388b8322ceaaef013a2a8f29582e161bd02" Feb 27 17:42:40 crc kubenswrapper[4700]: E0227 17:42:40.992187 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:42:41 crc kubenswrapper[4700]: I0227 17:42:41.748710 4700 generic.go:334] "Generic (PLEG): container finished" podID="82ea0ed5-7fde-447d-8fb4-780ad1fa6215" containerID="9f3419d394ef17ef080db92c79c7ae493996b3e66af50e73d3e42294f51bf616" exitCode=0 Feb 27 17:42:41 crc kubenswrapper[4700]: I0227 17:42:41.748836 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536902-vfh4z" event={"ID":"82ea0ed5-7fde-447d-8fb4-780ad1fa6215","Type":"ContainerDied","Data":"9f3419d394ef17ef080db92c79c7ae493996b3e66af50e73d3e42294f51bf616"} Feb 27 17:42:43 crc kubenswrapper[4700]: I0227 17:42:43.217937 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536902-vfh4z" Feb 27 17:42:43 crc kubenswrapper[4700]: I0227 17:42:43.362879 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmwct\" (UniqueName: \"kubernetes.io/projected/82ea0ed5-7fde-447d-8fb4-780ad1fa6215-kube-api-access-vmwct\") pod \"82ea0ed5-7fde-447d-8fb4-780ad1fa6215\" (UID: \"82ea0ed5-7fde-447d-8fb4-780ad1fa6215\") " Feb 27 17:42:43 crc kubenswrapper[4700]: I0227 17:42:43.373025 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82ea0ed5-7fde-447d-8fb4-780ad1fa6215-kube-api-access-vmwct" (OuterVolumeSpecName: "kube-api-access-vmwct") pod "82ea0ed5-7fde-447d-8fb4-780ad1fa6215" (UID: "82ea0ed5-7fde-447d-8fb4-780ad1fa6215"). InnerVolumeSpecName "kube-api-access-vmwct". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:42:43 crc kubenswrapper[4700]: I0227 17:42:43.465881 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmwct\" (UniqueName: \"kubernetes.io/projected/82ea0ed5-7fde-447d-8fb4-780ad1fa6215-kube-api-access-vmwct\") on node \"crc\" DevicePath \"\"" Feb 27 17:42:43 crc kubenswrapper[4700]: I0227 17:42:43.777596 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536902-vfh4z" event={"ID":"82ea0ed5-7fde-447d-8fb4-780ad1fa6215","Type":"ContainerDied","Data":"00c1fd6ff3c849c12c56fde5d291b7257f41f8d3630ec96ec45bcce491ea389f"} Feb 27 17:42:43 crc kubenswrapper[4700]: I0227 17:42:43.777655 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00c1fd6ff3c849c12c56fde5d291b7257f41f8d3630ec96ec45bcce491ea389f" Feb 27 17:42:43 crc kubenswrapper[4700]: I0227 17:42:43.778180 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536902-vfh4z" Feb 27 17:42:44 crc kubenswrapper[4700]: I0227 17:42:44.317319 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536896-zjbb8"] Feb 27 17:42:44 crc kubenswrapper[4700]: I0227 17:42:44.326485 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536896-zjbb8"] Feb 27 17:42:45 crc kubenswrapper[4700]: I0227 17:42:45.000481 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8daf403-3c61-4f52-9aa0-09b3e86aacd8" path="/var/lib/kubelet/pods/c8daf403-3c61-4f52-9aa0-09b3e86aacd8/volumes" Feb 27 17:42:49 crc kubenswrapper[4700]: I0227 17:42:49.850681 4700 generic.go:334] "Generic (PLEG): container finished" podID="abd22381-3367-4aef-a05d-dd895c6de545" containerID="53d22f1a20a387a5e781985709b6115c369b61add719a04d9602464cb9782a0e" exitCode=0 Feb 27 17:42:49 crc kubenswrapper[4700]: I0227 17:42:49.851681 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p" event={"ID":"abd22381-3367-4aef-a05d-dd895c6de545","Type":"ContainerDied","Data":"53d22f1a20a387a5e781985709b6115c369b61add719a04d9602464cb9782a0e"} Feb 27 17:42:51 crc kubenswrapper[4700]: I0227 17:42:51.345574 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p" Feb 27 17:42:51 crc kubenswrapper[4700]: I0227 17:42:51.461907 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kqhk\" (UniqueName: \"kubernetes.io/projected/abd22381-3367-4aef-a05d-dd895c6de545-kube-api-access-7kqhk\") pod \"abd22381-3367-4aef-a05d-dd895c6de545\" (UID: \"abd22381-3367-4aef-a05d-dd895c6de545\") " Feb 27 17:42:51 crc kubenswrapper[4700]: I0227 17:42:51.462223 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abd22381-3367-4aef-a05d-dd895c6de545-inventory\") pod \"abd22381-3367-4aef-a05d-dd895c6de545\" (UID: \"abd22381-3367-4aef-a05d-dd895c6de545\") " Feb 27 17:42:51 crc kubenswrapper[4700]: I0227 17:42:51.462571 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/abd22381-3367-4aef-a05d-dd895c6de545-libvirt-secret-0\") pod \"abd22381-3367-4aef-a05d-dd895c6de545\" (UID: \"abd22381-3367-4aef-a05d-dd895c6de545\") " Feb 27 17:42:51 crc kubenswrapper[4700]: I0227 17:42:51.462706 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abd22381-3367-4aef-a05d-dd895c6de545-libvirt-combined-ca-bundle\") pod \"abd22381-3367-4aef-a05d-dd895c6de545\" (UID: \"abd22381-3367-4aef-a05d-dd895c6de545\") " Feb 27 17:42:51 crc kubenswrapper[4700]: I0227 17:42:51.462748 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/abd22381-3367-4aef-a05d-dd895c6de545-ssh-key-openstack-edpm-ipam\") pod \"abd22381-3367-4aef-a05d-dd895c6de545\" (UID: \"abd22381-3367-4aef-a05d-dd895c6de545\") " Feb 27 17:42:51 crc kubenswrapper[4700]: I0227 17:42:51.477485 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abd22381-3367-4aef-a05d-dd895c6de545-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "abd22381-3367-4aef-a05d-dd895c6de545" (UID: "abd22381-3367-4aef-a05d-dd895c6de545"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:42:51 crc kubenswrapper[4700]: I0227 17:42:51.477682 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abd22381-3367-4aef-a05d-dd895c6de545-kube-api-access-7kqhk" (OuterVolumeSpecName: "kube-api-access-7kqhk") pod "abd22381-3367-4aef-a05d-dd895c6de545" (UID: "abd22381-3367-4aef-a05d-dd895c6de545"). InnerVolumeSpecName "kube-api-access-7kqhk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:42:51 crc kubenswrapper[4700]: I0227 17:42:51.496003 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abd22381-3367-4aef-a05d-dd895c6de545-inventory" (OuterVolumeSpecName: "inventory") pod "abd22381-3367-4aef-a05d-dd895c6de545" (UID: "abd22381-3367-4aef-a05d-dd895c6de545"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:42:51 crc kubenswrapper[4700]: I0227 17:42:51.512575 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abd22381-3367-4aef-a05d-dd895c6de545-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "abd22381-3367-4aef-a05d-dd895c6de545" (UID: "abd22381-3367-4aef-a05d-dd895c6de545"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:42:51 crc kubenswrapper[4700]: I0227 17:42:51.512609 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abd22381-3367-4aef-a05d-dd895c6de545-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "abd22381-3367-4aef-a05d-dd895c6de545" (UID: "abd22381-3367-4aef-a05d-dd895c6de545"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:42:51 crc kubenswrapper[4700]: I0227 17:42:51.565535 4700 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/abd22381-3367-4aef-a05d-dd895c6de545-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Feb 27 17:42:51 crc kubenswrapper[4700]: I0227 17:42:51.565573 4700 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abd22381-3367-4aef-a05d-dd895c6de545-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:42:51 crc kubenswrapper[4700]: I0227 17:42:51.565589 4700 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/abd22381-3367-4aef-a05d-dd895c6de545-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 17:42:51 crc kubenswrapper[4700]: I0227 17:42:51.565598 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kqhk\" (UniqueName: \"kubernetes.io/projected/abd22381-3367-4aef-a05d-dd895c6de545-kube-api-access-7kqhk\") on node \"crc\" DevicePath \"\"" Feb 27 17:42:51 crc kubenswrapper[4700]: I0227 17:42:51.565609 4700 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abd22381-3367-4aef-a05d-dd895c6de545-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 17:42:51 crc kubenswrapper[4700]: I0227 17:42:51.879037 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p" event={"ID":"abd22381-3367-4aef-a05d-dd895c6de545","Type":"ContainerDied","Data":"d6896d08355f4cead3b7747aba2863e435604a7920ad7e65859dc49789ed315a"} Feb 27 17:42:51 crc kubenswrapper[4700]: I0227 17:42:51.879102 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6896d08355f4cead3b7747aba2863e435604a7920ad7e65859dc49789ed315a" Feb 27 17:42:51 crc kubenswrapper[4700]: I0227 17:42:51.879191 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.043439 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq"] Feb 27 17:42:52 crc kubenswrapper[4700]: E0227 17:42:52.044337 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abd22381-3367-4aef-a05d-dd895c6de545" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.044379 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="abd22381-3367-4aef-a05d-dd895c6de545" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 27 17:42:52 crc kubenswrapper[4700]: E0227 17:42:52.044437 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82ea0ed5-7fde-447d-8fb4-780ad1fa6215" containerName="oc" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.044455 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="82ea0ed5-7fde-447d-8fb4-780ad1fa6215" containerName="oc" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.044989 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="abd22381-3367-4aef-a05d-dd895c6de545" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.045075 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="82ea0ed5-7fde-447d-8fb4-780ad1fa6215" containerName="oc" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.046668 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.050603 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.050807 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.051261 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dtdm4" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.051958 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.052008 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.052144 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.055669 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.059878 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq"] Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.078665 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7kqqq\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.078786 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7kqqq\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.078845 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7kqqq\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.078912 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7kqqq\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.079003 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7kqqq\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.079120 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7kqqq\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.079207 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7kqqq\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.079343 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvc6g\" (UniqueName: \"kubernetes.io/projected/fff27101-a84c-478b-9160-a7a5e6a189ea-kube-api-access-xvc6g\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7kqqq\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.079416 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7kqqq\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.079700 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7kqqq\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.079886 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7kqqq\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.181917 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvc6g\" (UniqueName: \"kubernetes.io/projected/fff27101-a84c-478b-9160-a7a5e6a189ea-kube-api-access-xvc6g\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7kqqq\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.181962 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7kqqq\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.182011 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7kqqq\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.182048 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7kqqq\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.182103 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7kqqq\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.182127 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7kqqq\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.182145 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7kqqq\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.182166 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7kqqq\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.182228 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7kqqq\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.182270 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7kqqq\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.182289 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7kqqq\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.188876 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7kqqq\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.189277 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7kqqq\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.189842 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7kqqq\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.190725 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7kqqq\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.190725 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7kqqq\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.191088 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7kqqq\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.192214 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7kqqq\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.193845 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7kqqq\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.194429 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7kqqq\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.197298 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7kqqq\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.210258 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvc6g\" (UniqueName: \"kubernetes.io/projected/fff27101-a84c-478b-9160-a7a5e6a189ea-kube-api-access-xvc6g\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7kqqq\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" Feb 27 17:42:52 crc kubenswrapper[4700]: I0227 17:42:52.367144 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" Feb 27 17:42:53 crc kubenswrapper[4700]: I0227 17:42:53.027996 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq"] Feb 27 17:42:53 crc kubenswrapper[4700]: I0227 17:42:53.908445 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" event={"ID":"fff27101-a84c-478b-9160-a7a5e6a189ea","Type":"ContainerStarted","Data":"c18ca5c6c09f76f2bbf92b9b097a1fb7ed278896d8b661a77d91513aaf37e28d"} Feb 27 17:42:53 crc kubenswrapper[4700]: I0227 17:42:53.908934 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" event={"ID":"fff27101-a84c-478b-9160-a7a5e6a189ea","Type":"ContainerStarted","Data":"b916560a60586e58a57ac074081d24d7ab46a99788adf885e3017427c97cfad7"} Feb 27 17:42:53 crc kubenswrapper[4700]: I0227 17:42:53.947501 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" podStartSLOduration=1.50195027 podStartE2EDuration="1.947443356s" podCreationTimestamp="2026-02-27 17:42:52 +0000 UTC" firstStartedPulling="2026-02-27 17:42:53.0139002 +0000 UTC m=+2532.999212957" lastFinishedPulling="2026-02-27 17:42:53.459393266 +0000 UTC m=+2533.444706043" observedRunningTime="2026-02-27 17:42:53.945632039 +0000 UTC m=+2533.930944826" watchObservedRunningTime="2026-02-27 17:42:53.947443356 +0000 UTC m=+2533.932756133" Feb 27 17:42:54 crc kubenswrapper[4700]: I0227 17:42:54.981617 4700 scope.go:117] "RemoveContainer" containerID="6843fae59a5a033363e409d439168388b8322ceaaef013a2a8f29582e161bd02" Feb 27 17:42:54 crc kubenswrapper[4700]: E0227 17:42:54.982513 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:43:07 crc kubenswrapper[4700]: I0227 17:43:07.981258 4700 scope.go:117] "RemoveContainer" containerID="6843fae59a5a033363e409d439168388b8322ceaaef013a2a8f29582e161bd02" Feb 27 17:43:07 crc kubenswrapper[4700]: E0227 17:43:07.982096 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:43:09 crc kubenswrapper[4700]: I0227 17:43:09.795648 4700 scope.go:117] "RemoveContainer" containerID="ae8c6a0f43dbb98d2cdda2193f245b84f860b855e30384c2778b3a181bb538ce" Feb 27 17:43:18 crc kubenswrapper[4700]: I0227 17:43:18.982039 4700 scope.go:117] "RemoveContainer" containerID="6843fae59a5a033363e409d439168388b8322ceaaef013a2a8f29582e161bd02" Feb 27 17:43:18 crc kubenswrapper[4700]: E0227 17:43:18.983005 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:43:32 crc kubenswrapper[4700]: I0227 17:43:32.982672 4700 scope.go:117] "RemoveContainer" containerID="6843fae59a5a033363e409d439168388b8322ceaaef013a2a8f29582e161bd02" Feb 27 17:43:32 crc kubenswrapper[4700]: E0227 17:43:32.983405 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:43:45 crc kubenswrapper[4700]: I0227 17:43:45.981390 4700 scope.go:117] "RemoveContainer" containerID="6843fae59a5a033363e409d439168388b8322ceaaef013a2a8f29582e161bd02" Feb 27 17:43:45 crc kubenswrapper[4700]: E0227 17:43:45.982062 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:43:59 crc kubenswrapper[4700]: I0227 17:43:59.980601 4700 scope.go:117] "RemoveContainer" containerID="6843fae59a5a033363e409d439168388b8322ceaaef013a2a8f29582e161bd02" Feb 27 17:43:59 crc kubenswrapper[4700]: E0227 17:43:59.981262 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:44:00 crc kubenswrapper[4700]: I0227 17:44:00.167859 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536904-fdk8p"] Feb 27 17:44:00 crc kubenswrapper[4700]: I0227 17:44:00.170163 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536904-fdk8p" Feb 27 17:44:00 crc kubenswrapper[4700]: I0227 17:44:00.176217 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d62kc" Feb 27 17:44:00 crc kubenswrapper[4700]: I0227 17:44:00.177409 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:44:00 crc kubenswrapper[4700]: I0227 17:44:00.177731 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:44:00 crc kubenswrapper[4700]: I0227 17:44:00.190791 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536904-fdk8p"] Feb 27 17:44:00 crc kubenswrapper[4700]: I0227 17:44:00.244717 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4b7pg\" (UniqueName: \"kubernetes.io/projected/cc474ff3-1a67-415b-9533-eefbe72a6822-kube-api-access-4b7pg\") pod \"auto-csr-approver-29536904-fdk8p\" (UID: \"cc474ff3-1a67-415b-9533-eefbe72a6822\") " pod="openshift-infra/auto-csr-approver-29536904-fdk8p" Feb 27 17:44:00 crc kubenswrapper[4700]: I0227 17:44:00.347311 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4b7pg\" (UniqueName: \"kubernetes.io/projected/cc474ff3-1a67-415b-9533-eefbe72a6822-kube-api-access-4b7pg\") pod \"auto-csr-approver-29536904-fdk8p\" (UID: \"cc474ff3-1a67-415b-9533-eefbe72a6822\") " pod="openshift-infra/auto-csr-approver-29536904-fdk8p" Feb 27 17:44:00 crc kubenswrapper[4700]: I0227 17:44:00.378397 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4b7pg\" (UniqueName: \"kubernetes.io/projected/cc474ff3-1a67-415b-9533-eefbe72a6822-kube-api-access-4b7pg\") pod \"auto-csr-approver-29536904-fdk8p\" (UID: \"cc474ff3-1a67-415b-9533-eefbe72a6822\") " pod="openshift-infra/auto-csr-approver-29536904-fdk8p" Feb 27 17:44:00 crc kubenswrapper[4700]: I0227 17:44:00.505259 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536904-fdk8p" Feb 27 17:44:01 crc kubenswrapper[4700]: I0227 17:44:01.006768 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536904-fdk8p"] Feb 27 17:44:01 crc kubenswrapper[4700]: W0227 17:44:01.012286 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc474ff3_1a67_415b_9533_eefbe72a6822.slice/crio-80d12144da0eb9e59c18dd64d8b2598353bf3a3ff06df2ac1b49c52f19171a15 WatchSource:0}: Error finding container 80d12144da0eb9e59c18dd64d8b2598353bf3a3ff06df2ac1b49c52f19171a15: Status 404 returned error can't find the container with id 80d12144da0eb9e59c18dd64d8b2598353bf3a3ff06df2ac1b49c52f19171a15 Feb 27 17:44:01 crc kubenswrapper[4700]: I0227 17:44:01.757668 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536904-fdk8p" event={"ID":"cc474ff3-1a67-415b-9533-eefbe72a6822","Type":"ContainerStarted","Data":"80d12144da0eb9e59c18dd64d8b2598353bf3a3ff06df2ac1b49c52f19171a15"} Feb 27 17:44:02 crc kubenswrapper[4700]: E0227 17:44:02.042979 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 17:44:02 crc kubenswrapper[4700]: E0227 17:44:02.043283 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 17:44:02 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 17:44:02 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4b7pg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536904-fdk8p_openshift-infra(cc474ff3-1a67-415b-9533-eefbe72a6822): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 17:44:02 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 17:44:02 crc kubenswrapper[4700]: E0227 17:44:02.044843 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536904-fdk8p" podUID="cc474ff3-1a67-415b-9533-eefbe72a6822" Feb 27 17:44:02 crc kubenswrapper[4700]: E0227 17:44:02.768663 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536904-fdk8p" podUID="cc474ff3-1a67-415b-9533-eefbe72a6822" Feb 27 17:44:13 crc kubenswrapper[4700]: I0227 17:44:13.981729 4700 scope.go:117] "RemoveContainer" containerID="6843fae59a5a033363e409d439168388b8322ceaaef013a2a8f29582e161bd02" Feb 27 17:44:13 crc kubenswrapper[4700]: E0227 17:44:13.982550 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:44:14 crc kubenswrapper[4700]: E0227 17:44:14.850947 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 17:44:14 crc kubenswrapper[4700]: E0227 17:44:14.851175 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 17:44:14 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 17:44:14 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4b7pg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536904-fdk8p_openshift-infra(cc474ff3-1a67-415b-9533-eefbe72a6822): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 17:44:14 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 17:44:14 crc kubenswrapper[4700]: E0227 17:44:14.852820 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536904-fdk8p" podUID="cc474ff3-1a67-415b-9533-eefbe72a6822" Feb 27 17:44:26 crc kubenswrapper[4700]: I0227 17:44:26.982349 4700 scope.go:117] "RemoveContainer" containerID="6843fae59a5a033363e409d439168388b8322ceaaef013a2a8f29582e161bd02" Feb 27 17:44:26 crc kubenswrapper[4700]: E0227 17:44:26.983425 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:44:28 crc kubenswrapper[4700]: E0227 17:44:28.985071 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536904-fdk8p" podUID="cc474ff3-1a67-415b-9533-eefbe72a6822" Feb 27 17:44:38 crc kubenswrapper[4700]: I0227 17:44:38.981440 4700 scope.go:117] "RemoveContainer" containerID="6843fae59a5a033363e409d439168388b8322ceaaef013a2a8f29582e161bd02" Feb 27 17:44:40 crc kubenswrapper[4700]: I0227 17:44:40.191779 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerStarted","Data":"9b478be1943d618189bcf0c56cb81fd36080f5aa6f5ea1159e5d5ecbda877263"} Feb 27 17:44:44 crc kubenswrapper[4700]: E0227 17:44:44.922287 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 17:44:44 crc kubenswrapper[4700]: E0227 17:44:44.923020 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 17:44:44 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 17:44:44 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4b7pg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536904-fdk8p_openshift-infra(cc474ff3-1a67-415b-9533-eefbe72a6822): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 17:44:44 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 17:44:44 crc kubenswrapper[4700]: E0227 17:44:44.924201 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536904-fdk8p" podUID="cc474ff3-1a67-415b-9533-eefbe72a6822" Feb 27 17:44:59 crc kubenswrapper[4700]: E0227 17:44:59.984118 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536904-fdk8p" podUID="cc474ff3-1a67-415b-9533-eefbe72a6822" Feb 27 17:45:00 crc kubenswrapper[4700]: I0227 17:45:00.187654 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536905-cchlk"] Feb 27 17:45:00 crc kubenswrapper[4700]: I0227 17:45:00.191425 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536905-cchlk" Feb 27 17:45:00 crc kubenswrapper[4700]: I0227 17:45:00.194752 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 27 17:45:00 crc kubenswrapper[4700]: I0227 17:45:00.197813 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 27 17:45:00 crc kubenswrapper[4700]: I0227 17:45:00.202226 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536905-cchlk"] Feb 27 17:45:00 crc kubenswrapper[4700]: I0227 17:45:00.287138 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rssvr\" (UniqueName: \"kubernetes.io/projected/86fc5416-1e17-405a-a34d-49965c235b33-kube-api-access-rssvr\") pod \"collect-profiles-29536905-cchlk\" (UID: \"86fc5416-1e17-405a-a34d-49965c235b33\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536905-cchlk" Feb 27 17:45:00 crc kubenswrapper[4700]: I0227 17:45:00.287207 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/86fc5416-1e17-405a-a34d-49965c235b33-config-volume\") pod \"collect-profiles-29536905-cchlk\" (UID: \"86fc5416-1e17-405a-a34d-49965c235b33\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536905-cchlk" Feb 27 17:45:00 crc kubenswrapper[4700]: I0227 17:45:00.287250 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/86fc5416-1e17-405a-a34d-49965c235b33-secret-volume\") pod \"collect-profiles-29536905-cchlk\" (UID: \"86fc5416-1e17-405a-a34d-49965c235b33\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536905-cchlk" Feb 27 17:45:00 crc kubenswrapper[4700]: I0227 17:45:00.421528 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rssvr\" (UniqueName: \"kubernetes.io/projected/86fc5416-1e17-405a-a34d-49965c235b33-kube-api-access-rssvr\") pod \"collect-profiles-29536905-cchlk\" (UID: \"86fc5416-1e17-405a-a34d-49965c235b33\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536905-cchlk" Feb 27 17:45:00 crc kubenswrapper[4700]: I0227 17:45:00.421700 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/86fc5416-1e17-405a-a34d-49965c235b33-config-volume\") pod \"collect-profiles-29536905-cchlk\" (UID: \"86fc5416-1e17-405a-a34d-49965c235b33\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536905-cchlk" Feb 27 17:45:00 crc kubenswrapper[4700]: I0227 17:45:00.421847 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/86fc5416-1e17-405a-a34d-49965c235b33-secret-volume\") pod \"collect-profiles-29536905-cchlk\" (UID: \"86fc5416-1e17-405a-a34d-49965c235b33\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536905-cchlk" Feb 27 17:45:00 crc kubenswrapper[4700]: I0227 17:45:00.422921 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/86fc5416-1e17-405a-a34d-49965c235b33-config-volume\") pod \"collect-profiles-29536905-cchlk\" (UID: \"86fc5416-1e17-405a-a34d-49965c235b33\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536905-cchlk" Feb 27 17:45:00 crc kubenswrapper[4700]: I0227 17:45:00.431827 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/86fc5416-1e17-405a-a34d-49965c235b33-secret-volume\") pod \"collect-profiles-29536905-cchlk\" (UID: \"86fc5416-1e17-405a-a34d-49965c235b33\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536905-cchlk" Feb 27 17:45:00 crc kubenswrapper[4700]: I0227 17:45:00.447838 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rssvr\" (UniqueName: \"kubernetes.io/projected/86fc5416-1e17-405a-a34d-49965c235b33-kube-api-access-rssvr\") pod \"collect-profiles-29536905-cchlk\" (UID: \"86fc5416-1e17-405a-a34d-49965c235b33\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536905-cchlk" Feb 27 17:45:00 crc kubenswrapper[4700]: I0227 17:45:00.521071 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536905-cchlk" Feb 27 17:45:01 crc kubenswrapper[4700]: I0227 17:45:00.999829 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536905-cchlk"] Feb 27 17:45:01 crc kubenswrapper[4700]: I0227 17:45:01.438135 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536905-cchlk" event={"ID":"86fc5416-1e17-405a-a34d-49965c235b33","Type":"ContainerStarted","Data":"371595c7840fd50eadd739ae642603d40a0efa929815ad8c6a687fc0d21b0dc7"} Feb 27 17:45:01 crc kubenswrapper[4700]: I0227 17:45:01.438174 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536905-cchlk" event={"ID":"86fc5416-1e17-405a-a34d-49965c235b33","Type":"ContainerStarted","Data":"c9736ccc2d11222a92b49602a8251329949045fa0576d80b8c03775d121ac671"} Feb 27 17:45:01 crc kubenswrapper[4700]: I0227 17:45:01.461515 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29536905-cchlk" podStartSLOduration=1.4614887730000001 podStartE2EDuration="1.461488773s" podCreationTimestamp="2026-02-27 17:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:45:01.455173668 +0000 UTC m=+2661.440486415" watchObservedRunningTime="2026-02-27 17:45:01.461488773 +0000 UTC m=+2661.446801520" Feb 27 17:45:02 crc kubenswrapper[4700]: I0227 17:45:02.452501 4700 generic.go:334] "Generic (PLEG): container finished" podID="86fc5416-1e17-405a-a34d-49965c235b33" containerID="371595c7840fd50eadd739ae642603d40a0efa929815ad8c6a687fc0d21b0dc7" exitCode=0 Feb 27 17:45:02 crc kubenswrapper[4700]: I0227 17:45:02.452554 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536905-cchlk" event={"ID":"86fc5416-1e17-405a-a34d-49965c235b33","Type":"ContainerDied","Data":"371595c7840fd50eadd739ae642603d40a0efa929815ad8c6a687fc0d21b0dc7"} Feb 27 17:45:04 crc kubenswrapper[4700]: I0227 17:45:03.928714 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536905-cchlk" Feb 27 17:45:04 crc kubenswrapper[4700]: I0227 17:45:04.013765 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rssvr\" (UniqueName: \"kubernetes.io/projected/86fc5416-1e17-405a-a34d-49965c235b33-kube-api-access-rssvr\") pod \"86fc5416-1e17-405a-a34d-49965c235b33\" (UID: \"86fc5416-1e17-405a-a34d-49965c235b33\") " Feb 27 17:45:04 crc kubenswrapper[4700]: I0227 17:45:04.013850 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/86fc5416-1e17-405a-a34d-49965c235b33-secret-volume\") pod \"86fc5416-1e17-405a-a34d-49965c235b33\" (UID: \"86fc5416-1e17-405a-a34d-49965c235b33\") " Feb 27 17:45:04 crc kubenswrapper[4700]: I0227 17:45:04.013927 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/86fc5416-1e17-405a-a34d-49965c235b33-config-volume\") pod \"86fc5416-1e17-405a-a34d-49965c235b33\" (UID: \"86fc5416-1e17-405a-a34d-49965c235b33\") " Feb 27 17:45:04 crc kubenswrapper[4700]: I0227 17:45:04.014908 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86fc5416-1e17-405a-a34d-49965c235b33-config-volume" (OuterVolumeSpecName: "config-volume") pod "86fc5416-1e17-405a-a34d-49965c235b33" (UID: "86fc5416-1e17-405a-a34d-49965c235b33"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:45:04 crc kubenswrapper[4700]: I0227 17:45:04.034675 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86fc5416-1e17-405a-a34d-49965c235b33-kube-api-access-rssvr" (OuterVolumeSpecName: "kube-api-access-rssvr") pod "86fc5416-1e17-405a-a34d-49965c235b33" (UID: "86fc5416-1e17-405a-a34d-49965c235b33"). InnerVolumeSpecName "kube-api-access-rssvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:45:04 crc kubenswrapper[4700]: I0227 17:45:04.035112 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86fc5416-1e17-405a-a34d-49965c235b33-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "86fc5416-1e17-405a-a34d-49965c235b33" (UID: "86fc5416-1e17-405a-a34d-49965c235b33"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:45:04 crc kubenswrapper[4700]: I0227 17:45:04.116185 4700 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/86fc5416-1e17-405a-a34d-49965c235b33-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 27 17:45:04 crc kubenswrapper[4700]: I0227 17:45:04.116214 4700 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/86fc5416-1e17-405a-a34d-49965c235b33-config-volume\") on node \"crc\" DevicePath \"\"" Feb 27 17:45:04 crc kubenswrapper[4700]: I0227 17:45:04.116224 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rssvr\" (UniqueName: \"kubernetes.io/projected/86fc5416-1e17-405a-a34d-49965c235b33-kube-api-access-rssvr\") on node \"crc\" DevicePath \"\"" Feb 27 17:45:04 crc kubenswrapper[4700]: I0227 17:45:04.477423 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536905-cchlk" event={"ID":"86fc5416-1e17-405a-a34d-49965c235b33","Type":"ContainerDied","Data":"c9736ccc2d11222a92b49602a8251329949045fa0576d80b8c03775d121ac671"} Feb 27 17:45:04 crc kubenswrapper[4700]: I0227 17:45:04.477508 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9736ccc2d11222a92b49602a8251329949045fa0576d80b8c03775d121ac671" Feb 27 17:45:04 crc kubenswrapper[4700]: I0227 17:45:04.477582 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536905-cchlk" Feb 27 17:45:04 crc kubenswrapper[4700]: I0227 17:45:04.544865 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536860-dw5fg"] Feb 27 17:45:04 crc kubenswrapper[4700]: I0227 17:45:04.556439 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536860-dw5fg"] Feb 27 17:45:05 crc kubenswrapper[4700]: I0227 17:45:05.003753 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66e8ed9b-99d5-4643-aa2a-4720ea4195b9" path="/var/lib/kubelet/pods/66e8ed9b-99d5-4643-aa2a-4720ea4195b9/volumes" Feb 27 17:45:09 crc kubenswrapper[4700]: I0227 17:45:09.912333 4700 scope.go:117] "RemoveContainer" containerID="9cb12c74550ad13f426db5679571715af91b50eaa633f56ce6a1c7a8c66ae2d3" Feb 27 17:45:14 crc kubenswrapper[4700]: E0227 17:45:14.984222 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536904-fdk8p" podUID="cc474ff3-1a67-415b-9533-eefbe72a6822" Feb 27 17:45:29 crc kubenswrapper[4700]: E0227 17:45:29.905082 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 17:45:29 crc kubenswrapper[4700]: E0227 17:45:29.905767 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 17:45:29 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 17:45:29 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4b7pg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536904-fdk8p_openshift-infra(cc474ff3-1a67-415b-9533-eefbe72a6822): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 17:45:29 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 17:45:29 crc kubenswrapper[4700]: E0227 17:45:29.907005 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536904-fdk8p" podUID="cc474ff3-1a67-415b-9533-eefbe72a6822" Feb 27 17:45:40 crc kubenswrapper[4700]: I0227 17:45:40.889151 4700 generic.go:334] "Generic (PLEG): container finished" podID="fff27101-a84c-478b-9160-a7a5e6a189ea" containerID="c18ca5c6c09f76f2bbf92b9b097a1fb7ed278896d8b661a77d91513aaf37e28d" exitCode=0 Feb 27 17:45:40 crc kubenswrapper[4700]: I0227 17:45:40.889250 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" event={"ID":"fff27101-a84c-478b-9160-a7a5e6a189ea","Type":"ContainerDied","Data":"c18ca5c6c09f76f2bbf92b9b097a1fb7ed278896d8b661a77d91513aaf37e28d"} Feb 27 17:45:41 crc kubenswrapper[4700]: E0227 17:45:41.006617 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536904-fdk8p" podUID="cc474ff3-1a67-415b-9533-eefbe72a6822" Feb 27 17:45:42 crc kubenswrapper[4700]: I0227 17:45:42.396315 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" Feb 27 17:45:42 crc kubenswrapper[4700]: I0227 17:45:42.495032 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-cell1-compute-config-1\") pod \"fff27101-a84c-478b-9160-a7a5e6a189ea\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " Feb 27 17:45:42 crc kubenswrapper[4700]: I0227 17:45:42.495196 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-extra-config-0\") pod \"fff27101-a84c-478b-9160-a7a5e6a189ea\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " Feb 27 17:45:42 crc kubenswrapper[4700]: I0227 17:45:42.495241 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-cell1-compute-config-2\") pod \"fff27101-a84c-478b-9160-a7a5e6a189ea\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " Feb 27 17:45:42 crc kubenswrapper[4700]: I0227 17:45:42.495372 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-combined-ca-bundle\") pod \"fff27101-a84c-478b-9160-a7a5e6a189ea\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " Feb 27 17:45:42 crc kubenswrapper[4700]: I0227 17:45:42.495495 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-migration-ssh-key-0\") pod \"fff27101-a84c-478b-9160-a7a5e6a189ea\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " Feb 27 17:45:42 crc kubenswrapper[4700]: I0227 17:45:42.495541 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-cell1-compute-config-3\") pod \"fff27101-a84c-478b-9160-a7a5e6a189ea\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " Feb 27 17:45:42 crc kubenswrapper[4700]: I0227 17:45:42.495676 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-ssh-key-openstack-edpm-ipam\") pod \"fff27101-a84c-478b-9160-a7a5e6a189ea\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " Feb 27 17:45:42 crc kubenswrapper[4700]: I0227 17:45:42.495897 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvc6g\" (UniqueName: \"kubernetes.io/projected/fff27101-a84c-478b-9160-a7a5e6a189ea-kube-api-access-xvc6g\") pod \"fff27101-a84c-478b-9160-a7a5e6a189ea\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " Feb 27 17:45:42 crc kubenswrapper[4700]: I0227 17:45:42.495995 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-migration-ssh-key-1\") pod \"fff27101-a84c-478b-9160-a7a5e6a189ea\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " Feb 27 17:45:42 crc kubenswrapper[4700]: I0227 17:45:42.496059 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-cell1-compute-config-0\") pod \"fff27101-a84c-478b-9160-a7a5e6a189ea\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " Feb 27 17:45:42 crc kubenswrapper[4700]: I0227 17:45:42.496158 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-inventory\") pod \"fff27101-a84c-478b-9160-a7a5e6a189ea\" (UID: \"fff27101-a84c-478b-9160-a7a5e6a189ea\") " Feb 27 17:45:42 crc kubenswrapper[4700]: I0227 17:45:42.523031 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fff27101-a84c-478b-9160-a7a5e6a189ea-kube-api-access-xvc6g" (OuterVolumeSpecName: "kube-api-access-xvc6g") pod "fff27101-a84c-478b-9160-a7a5e6a189ea" (UID: "fff27101-a84c-478b-9160-a7a5e6a189ea"). InnerVolumeSpecName "kube-api-access-xvc6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:45:42 crc kubenswrapper[4700]: I0227 17:45:42.523586 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "fff27101-a84c-478b-9160-a7a5e6a189ea" (UID: "fff27101-a84c-478b-9160-a7a5e6a189ea"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:45:42 crc kubenswrapper[4700]: I0227 17:45:42.527532 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-cell1-compute-config-2" (OuterVolumeSpecName: "nova-cell1-compute-config-2") pod "fff27101-a84c-478b-9160-a7a5e6a189ea" (UID: "fff27101-a84c-478b-9160-a7a5e6a189ea"). InnerVolumeSpecName "nova-cell1-compute-config-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:45:42 crc kubenswrapper[4700]: I0227 17:45:42.527666 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "fff27101-a84c-478b-9160-a7a5e6a189ea" (UID: "fff27101-a84c-478b-9160-a7a5e6a189ea"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:45:42 crc kubenswrapper[4700]: I0227 17:45:42.528124 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "fff27101-a84c-478b-9160-a7a5e6a189ea" (UID: "fff27101-a84c-478b-9160-a7a5e6a189ea"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:45:42 crc kubenswrapper[4700]: I0227 17:45:42.531510 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "fff27101-a84c-478b-9160-a7a5e6a189ea" (UID: "fff27101-a84c-478b-9160-a7a5e6a189ea"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:45:42 crc kubenswrapper[4700]: I0227 17:45:42.531652 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "fff27101-a84c-478b-9160-a7a5e6a189ea" (UID: "fff27101-a84c-478b-9160-a7a5e6a189ea"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:45:42 crc kubenswrapper[4700]: I0227 17:45:42.533918 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "fff27101-a84c-478b-9160-a7a5e6a189ea" (UID: "fff27101-a84c-478b-9160-a7a5e6a189ea"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:45:42 crc kubenswrapper[4700]: I0227 17:45:42.538789 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-cell1-compute-config-3" (OuterVolumeSpecName: "nova-cell1-compute-config-3") pod "fff27101-a84c-478b-9160-a7a5e6a189ea" (UID: "fff27101-a84c-478b-9160-a7a5e6a189ea"). InnerVolumeSpecName "nova-cell1-compute-config-3". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:45:42 crc kubenswrapper[4700]: I0227 17:45:42.549371 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-inventory" (OuterVolumeSpecName: "inventory") pod "fff27101-a84c-478b-9160-a7a5e6a189ea" (UID: "fff27101-a84c-478b-9160-a7a5e6a189ea"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:45:42 crc kubenswrapper[4700]: I0227 17:45:42.551000 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "fff27101-a84c-478b-9160-a7a5e6a189ea" (UID: "fff27101-a84c-478b-9160-a7a5e6a189ea"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:45:42 crc kubenswrapper[4700]: I0227 17:45:42.598709 4700 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 17:45:42 crc kubenswrapper[4700]: I0227 17:45:42.598750 4700 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Feb 27 17:45:42 crc kubenswrapper[4700]: I0227 17:45:42.598771 4700 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Feb 27 17:45:42 crc kubenswrapper[4700]: I0227 17:45:42.598783 4700 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-cell1-compute-config-2\") on node \"crc\" DevicePath \"\"" Feb 27 17:45:42 crc kubenswrapper[4700]: I0227 17:45:42.598795 4700 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:45:42 crc kubenswrapper[4700]: I0227 17:45:42.598806 4700 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Feb 27 17:45:42 crc kubenswrapper[4700]: I0227 17:45:42.598820 4700 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-cell1-compute-config-3\") on node \"crc\" DevicePath \"\"" Feb 27 17:45:42 crc kubenswrapper[4700]: I0227 17:45:42.598831 4700 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 17:45:42 crc kubenswrapper[4700]: I0227 17:45:42.598843 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvc6g\" (UniqueName: \"kubernetes.io/projected/fff27101-a84c-478b-9160-a7a5e6a189ea-kube-api-access-xvc6g\") on node \"crc\" DevicePath \"\"" Feb 27 17:45:42 crc kubenswrapper[4700]: I0227 17:45:42.598854 4700 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Feb 27 17:45:42 crc kubenswrapper[4700]: I0227 17:45:42.598866 4700 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/fff27101-a84c-478b-9160-a7a5e6a189ea-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Feb 27 17:45:42 crc kubenswrapper[4700]: I0227 17:45:42.924712 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" event={"ID":"fff27101-a84c-478b-9160-a7a5e6a189ea","Type":"ContainerDied","Data":"b916560a60586e58a57ac074081d24d7ab46a99788adf885e3017427c97cfad7"} Feb 27 17:45:42 crc kubenswrapper[4700]: I0227 17:45:42.924762 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b916560a60586e58a57ac074081d24d7ab46a99788adf885e3017427c97cfad7" Feb 27 17:45:42 crc kubenswrapper[4700]: I0227 17:45:42.924820 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7kqqq" Feb 27 17:45:43 crc kubenswrapper[4700]: I0227 17:45:43.064015 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn"] Feb 27 17:45:43 crc kubenswrapper[4700]: E0227 17:45:43.064667 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86fc5416-1e17-405a-a34d-49965c235b33" containerName="collect-profiles" Feb 27 17:45:43 crc kubenswrapper[4700]: I0227 17:45:43.064695 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="86fc5416-1e17-405a-a34d-49965c235b33" containerName="collect-profiles" Feb 27 17:45:43 crc kubenswrapper[4700]: E0227 17:45:43.064716 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fff27101-a84c-478b-9160-a7a5e6a189ea" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 27 17:45:43 crc kubenswrapper[4700]: I0227 17:45:43.064724 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="fff27101-a84c-478b-9160-a7a5e6a189ea" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 27 17:45:43 crc kubenswrapper[4700]: I0227 17:45:43.065086 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="fff27101-a84c-478b-9160-a7a5e6a189ea" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 27 17:45:43 crc kubenswrapper[4700]: I0227 17:45:43.065140 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="86fc5416-1e17-405a-a34d-49965c235b33" containerName="collect-profiles" Feb 27 17:45:43 crc kubenswrapper[4700]: I0227 17:45:43.066109 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn" Feb 27 17:45:43 crc kubenswrapper[4700]: I0227 17:45:43.069978 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dtdm4" Feb 27 17:45:43 crc kubenswrapper[4700]: I0227 17:45:43.070030 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Feb 27 17:45:43 crc kubenswrapper[4700]: I0227 17:45:43.070242 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 17:45:43 crc kubenswrapper[4700]: I0227 17:45:43.070327 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 17:45:43 crc kubenswrapper[4700]: I0227 17:45:43.070563 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 17:45:43 crc kubenswrapper[4700]: I0227 17:45:43.076565 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn"] Feb 27 17:45:43 crc kubenswrapper[4700]: I0227 17:45:43.212515 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d9b66d11-37c1-4826-96c3-967dd98f053c-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn\" (UID: \"d9b66d11-37c1-4826-96c3-967dd98f053c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn" Feb 27 17:45:43 crc kubenswrapper[4700]: I0227 17:45:43.212660 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9b66d11-37c1-4826-96c3-967dd98f053c-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn\" (UID: \"d9b66d11-37c1-4826-96c3-967dd98f053c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn" Feb 27 17:45:43 crc kubenswrapper[4700]: I0227 17:45:43.213070 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9b66d11-37c1-4826-96c3-967dd98f053c-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn\" (UID: \"d9b66d11-37c1-4826-96c3-967dd98f053c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn" Feb 27 17:45:43 crc kubenswrapper[4700]: I0227 17:45:43.213187 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d9b66d11-37c1-4826-96c3-967dd98f053c-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn\" (UID: \"d9b66d11-37c1-4826-96c3-967dd98f053c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn" Feb 27 17:45:43 crc kubenswrapper[4700]: I0227 17:45:43.213380 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fj4x\" (UniqueName: \"kubernetes.io/projected/d9b66d11-37c1-4826-96c3-967dd98f053c-kube-api-access-2fj4x\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn\" (UID: \"d9b66d11-37c1-4826-96c3-967dd98f053c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn" Feb 27 17:45:43 crc kubenswrapper[4700]: I0227 17:45:43.213646 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d9b66d11-37c1-4826-96c3-967dd98f053c-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn\" (UID: \"d9b66d11-37c1-4826-96c3-967dd98f053c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn" Feb 27 17:45:43 crc kubenswrapper[4700]: I0227 17:45:43.213753 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d9b66d11-37c1-4826-96c3-967dd98f053c-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn\" (UID: \"d9b66d11-37c1-4826-96c3-967dd98f053c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn" Feb 27 17:45:43 crc kubenswrapper[4700]: I0227 17:45:43.316086 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d9b66d11-37c1-4826-96c3-967dd98f053c-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn\" (UID: \"d9b66d11-37c1-4826-96c3-967dd98f053c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn" Feb 27 17:45:43 crc kubenswrapper[4700]: I0227 17:45:43.316135 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d9b66d11-37c1-4826-96c3-967dd98f053c-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn\" (UID: \"d9b66d11-37c1-4826-96c3-967dd98f053c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn" Feb 27 17:45:43 crc kubenswrapper[4700]: I0227 17:45:43.316170 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d9b66d11-37c1-4826-96c3-967dd98f053c-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn\" (UID: \"d9b66d11-37c1-4826-96c3-967dd98f053c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn" Feb 27 17:45:43 crc kubenswrapper[4700]: I0227 17:45:43.316230 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9b66d11-37c1-4826-96c3-967dd98f053c-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn\" (UID: \"d9b66d11-37c1-4826-96c3-967dd98f053c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn" Feb 27 17:45:43 crc kubenswrapper[4700]: I0227 17:45:43.316326 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9b66d11-37c1-4826-96c3-967dd98f053c-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn\" (UID: \"d9b66d11-37c1-4826-96c3-967dd98f053c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn" Feb 27 17:45:43 crc kubenswrapper[4700]: I0227 17:45:43.316364 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d9b66d11-37c1-4826-96c3-967dd98f053c-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn\" (UID: \"d9b66d11-37c1-4826-96c3-967dd98f053c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn" Feb 27 17:45:43 crc kubenswrapper[4700]: I0227 17:45:43.316421 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fj4x\" (UniqueName: \"kubernetes.io/projected/d9b66d11-37c1-4826-96c3-967dd98f053c-kube-api-access-2fj4x\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn\" (UID: \"d9b66d11-37c1-4826-96c3-967dd98f053c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn" Feb 27 17:45:43 crc kubenswrapper[4700]: I0227 17:45:43.321504 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d9b66d11-37c1-4826-96c3-967dd98f053c-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn\" (UID: \"d9b66d11-37c1-4826-96c3-967dd98f053c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn" Feb 27 17:45:43 crc kubenswrapper[4700]: I0227 17:45:43.321576 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d9b66d11-37c1-4826-96c3-967dd98f053c-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn\" (UID: \"d9b66d11-37c1-4826-96c3-967dd98f053c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn" Feb 27 17:45:43 crc kubenswrapper[4700]: I0227 17:45:43.322516 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d9b66d11-37c1-4826-96c3-967dd98f053c-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn\" (UID: \"d9b66d11-37c1-4826-96c3-967dd98f053c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn" Feb 27 17:45:43 crc kubenswrapper[4700]: I0227 17:45:43.322541 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9b66d11-37c1-4826-96c3-967dd98f053c-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn\" (UID: \"d9b66d11-37c1-4826-96c3-967dd98f053c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn" Feb 27 17:45:43 crc kubenswrapper[4700]: I0227 17:45:43.323826 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9b66d11-37c1-4826-96c3-967dd98f053c-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn\" (UID: \"d9b66d11-37c1-4826-96c3-967dd98f053c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn" Feb 27 17:45:43 crc kubenswrapper[4700]: I0227 17:45:43.323846 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d9b66d11-37c1-4826-96c3-967dd98f053c-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn\" (UID: \"d9b66d11-37c1-4826-96c3-967dd98f053c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn" Feb 27 17:45:43 crc kubenswrapper[4700]: I0227 17:45:43.344153 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fj4x\" (UniqueName: \"kubernetes.io/projected/d9b66d11-37c1-4826-96c3-967dd98f053c-kube-api-access-2fj4x\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn\" (UID: \"d9b66d11-37c1-4826-96c3-967dd98f053c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn" Feb 27 17:45:43 crc kubenswrapper[4700]: I0227 17:45:43.400447 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn" Feb 27 17:45:44 crc kubenswrapper[4700]: I0227 17:45:44.047911 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn"] Feb 27 17:45:44 crc kubenswrapper[4700]: I0227 17:45:44.963854 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn" event={"ID":"d9b66d11-37c1-4826-96c3-967dd98f053c","Type":"ContainerStarted","Data":"036b10b81dd522449d3bc6674c8f9d82fa8fda100f6b0f4274142b0d5446efcd"} Feb 27 17:45:44 crc kubenswrapper[4700]: I0227 17:45:44.964566 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn" event={"ID":"d9b66d11-37c1-4826-96c3-967dd98f053c","Type":"ContainerStarted","Data":"64f1c4919a0cfae19d3b86eb8d43ff32a9dd65e6d9c52505a0f7352cae7316e0"} Feb 27 17:45:44 crc kubenswrapper[4700]: I0227 17:45:44.995360 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn" podStartSLOduration=1.548667131 podStartE2EDuration="1.995335988s" podCreationTimestamp="2026-02-27 17:45:43 +0000 UTC" firstStartedPulling="2026-02-27 17:45:44.048232846 +0000 UTC m=+2704.033545603" lastFinishedPulling="2026-02-27 17:45:44.494901713 +0000 UTC m=+2704.480214460" observedRunningTime="2026-02-27 17:45:44.981154627 +0000 UTC m=+2704.966467454" watchObservedRunningTime="2026-02-27 17:45:44.995335988 +0000 UTC m=+2704.980648735" Feb 27 17:45:53 crc kubenswrapper[4700]: E0227 17:45:53.985085 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536904-fdk8p" podUID="cc474ff3-1a67-415b-9533-eefbe72a6822" Feb 27 17:46:00 crc kubenswrapper[4700]: I0227 17:46:00.148700 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536906-lhhr4"] Feb 27 17:46:00 crc kubenswrapper[4700]: I0227 17:46:00.151422 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536906-lhhr4" Feb 27 17:46:00 crc kubenswrapper[4700]: I0227 17:46:00.164286 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536906-lhhr4"] Feb 27 17:46:00 crc kubenswrapper[4700]: I0227 17:46:00.206078 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzxzg\" (UniqueName: \"kubernetes.io/projected/c4eb7377-a24a-46df-9aec-b1f977b56708-kube-api-access-fzxzg\") pod \"auto-csr-approver-29536906-lhhr4\" (UID: \"c4eb7377-a24a-46df-9aec-b1f977b56708\") " pod="openshift-infra/auto-csr-approver-29536906-lhhr4" Feb 27 17:46:00 crc kubenswrapper[4700]: I0227 17:46:00.308150 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzxzg\" (UniqueName: \"kubernetes.io/projected/c4eb7377-a24a-46df-9aec-b1f977b56708-kube-api-access-fzxzg\") pod \"auto-csr-approver-29536906-lhhr4\" (UID: \"c4eb7377-a24a-46df-9aec-b1f977b56708\") " pod="openshift-infra/auto-csr-approver-29536906-lhhr4" Feb 27 17:46:00 crc kubenswrapper[4700]: I0227 17:46:00.335249 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzxzg\" (UniqueName: \"kubernetes.io/projected/c4eb7377-a24a-46df-9aec-b1f977b56708-kube-api-access-fzxzg\") pod \"auto-csr-approver-29536906-lhhr4\" (UID: \"c4eb7377-a24a-46df-9aec-b1f977b56708\") " pod="openshift-infra/auto-csr-approver-29536906-lhhr4" Feb 27 17:46:00 crc kubenswrapper[4700]: I0227 17:46:00.494807 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536906-lhhr4" Feb 27 17:46:01 crc kubenswrapper[4700]: I0227 17:46:01.016280 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536906-lhhr4"] Feb 27 17:46:01 crc kubenswrapper[4700]: W0227 17:46:01.018487 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4eb7377_a24a_46df_9aec_b1f977b56708.slice/crio-bfd6a20f92dadce9fbd184f1b5108b50de11f271cd8d25f02a04e70d8c6eff18 WatchSource:0}: Error finding container bfd6a20f92dadce9fbd184f1b5108b50de11f271cd8d25f02a04e70d8c6eff18: Status 404 returned error can't find the container with id bfd6a20f92dadce9fbd184f1b5108b50de11f271cd8d25f02a04e70d8c6eff18 Feb 27 17:46:01 crc kubenswrapper[4700]: I0227 17:46:01.166028 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536906-lhhr4" event={"ID":"c4eb7377-a24a-46df-9aec-b1f977b56708","Type":"ContainerStarted","Data":"bfd6a20f92dadce9fbd184f1b5108b50de11f271cd8d25f02a04e70d8c6eff18"} Feb 27 17:46:04 crc kubenswrapper[4700]: I0227 17:46:04.209008 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536906-lhhr4" event={"ID":"c4eb7377-a24a-46df-9aec-b1f977b56708","Type":"ContainerStarted","Data":"1d5e81db8e3e48c15ad1a1c701d8c43aade148b25183151df8761b6c50825556"} Feb 27 17:46:04 crc kubenswrapper[4700]: I0227 17:46:04.228587 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536906-lhhr4" podStartSLOduration=1.414472793 podStartE2EDuration="4.228566714s" podCreationTimestamp="2026-02-27 17:46:00 +0000 UTC" firstStartedPulling="2026-02-27 17:46:01.023137745 +0000 UTC m=+2721.008450502" lastFinishedPulling="2026-02-27 17:46:03.837231676 +0000 UTC m=+2723.822544423" observedRunningTime="2026-02-27 17:46:04.223518292 +0000 UTC m=+2724.208831049" watchObservedRunningTime="2026-02-27 17:46:04.228566714 +0000 UTC m=+2724.213879481" Feb 27 17:46:05 crc kubenswrapper[4700]: I0227 17:46:05.217198 4700 generic.go:334] "Generic (PLEG): container finished" podID="c4eb7377-a24a-46df-9aec-b1f977b56708" containerID="1d5e81db8e3e48c15ad1a1c701d8c43aade148b25183151df8761b6c50825556" exitCode=0 Feb 27 17:46:05 crc kubenswrapper[4700]: I0227 17:46:05.217297 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536906-lhhr4" event={"ID":"c4eb7377-a24a-46df-9aec-b1f977b56708","Type":"ContainerDied","Data":"1d5e81db8e3e48c15ad1a1c701d8c43aade148b25183151df8761b6c50825556"} Feb 27 17:46:06 crc kubenswrapper[4700]: I0227 17:46:06.691779 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536906-lhhr4" Feb 27 17:46:06 crc kubenswrapper[4700]: I0227 17:46:06.787804 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzxzg\" (UniqueName: \"kubernetes.io/projected/c4eb7377-a24a-46df-9aec-b1f977b56708-kube-api-access-fzxzg\") pod \"c4eb7377-a24a-46df-9aec-b1f977b56708\" (UID: \"c4eb7377-a24a-46df-9aec-b1f977b56708\") " Feb 27 17:46:06 crc kubenswrapper[4700]: I0227 17:46:06.793682 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4eb7377-a24a-46df-9aec-b1f977b56708-kube-api-access-fzxzg" (OuterVolumeSpecName: "kube-api-access-fzxzg") pod "c4eb7377-a24a-46df-9aec-b1f977b56708" (UID: "c4eb7377-a24a-46df-9aec-b1f977b56708"). InnerVolumeSpecName "kube-api-access-fzxzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:46:06 crc kubenswrapper[4700]: I0227 17:46:06.890975 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzxzg\" (UniqueName: \"kubernetes.io/projected/c4eb7377-a24a-46df-9aec-b1f977b56708-kube-api-access-fzxzg\") on node \"crc\" DevicePath \"\"" Feb 27 17:46:07 crc kubenswrapper[4700]: I0227 17:46:07.263136 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536906-lhhr4" event={"ID":"c4eb7377-a24a-46df-9aec-b1f977b56708","Type":"ContainerDied","Data":"bfd6a20f92dadce9fbd184f1b5108b50de11f271cd8d25f02a04e70d8c6eff18"} Feb 27 17:46:07 crc kubenswrapper[4700]: I0227 17:46:07.263608 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bfd6a20f92dadce9fbd184f1b5108b50de11f271cd8d25f02a04e70d8c6eff18" Feb 27 17:46:07 crc kubenswrapper[4700]: I0227 17:46:07.263720 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536906-lhhr4" Feb 27 17:46:07 crc kubenswrapper[4700]: I0227 17:46:07.341495 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536898-4zjfl"] Feb 27 17:46:07 crc kubenswrapper[4700]: I0227 17:46:07.359508 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536898-4zjfl"] Feb 27 17:46:08 crc kubenswrapper[4700]: E0227 17:46:08.984415 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536904-fdk8p" podUID="cc474ff3-1a67-415b-9533-eefbe72a6822" Feb 27 17:46:08 crc kubenswrapper[4700]: I0227 17:46:08.998960 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c095a55-8ba9-44a9-b4c0-656cf3de7880" path="/var/lib/kubelet/pods/6c095a55-8ba9-44a9-b4c0-656cf3de7880/volumes" Feb 27 17:46:22 crc kubenswrapper[4700]: E0227 17:46:22.983902 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536904-fdk8p" podUID="cc474ff3-1a67-415b-9533-eefbe72a6822" Feb 27 17:46:33 crc kubenswrapper[4700]: E0227 17:46:33.984144 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536904-fdk8p" podUID="cc474ff3-1a67-415b-9533-eefbe72a6822" Feb 27 17:46:46 crc kubenswrapper[4700]: I0227 17:46:46.068361 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7zbv5"] Feb 27 17:46:46 crc kubenswrapper[4700]: E0227 17:46:46.069269 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4eb7377-a24a-46df-9aec-b1f977b56708" containerName="oc" Feb 27 17:46:46 crc kubenswrapper[4700]: I0227 17:46:46.069282 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4eb7377-a24a-46df-9aec-b1f977b56708" containerName="oc" Feb 27 17:46:46 crc kubenswrapper[4700]: I0227 17:46:46.069538 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4eb7377-a24a-46df-9aec-b1f977b56708" containerName="oc" Feb 27 17:46:46 crc kubenswrapper[4700]: I0227 17:46:46.070951 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7zbv5" Feb 27 17:46:46 crc kubenswrapper[4700]: I0227 17:46:46.094841 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7zbv5"] Feb 27 17:46:46 crc kubenswrapper[4700]: I0227 17:46:46.158172 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnkf4\" (UniqueName: \"kubernetes.io/projected/a0c58c8c-ee03-4646-b274-cf0edf8fe47b-kube-api-access-xnkf4\") pod \"redhat-operators-7zbv5\" (UID: \"a0c58c8c-ee03-4646-b274-cf0edf8fe47b\") " pod="openshift-marketplace/redhat-operators-7zbv5" Feb 27 17:46:46 crc kubenswrapper[4700]: I0227 17:46:46.158232 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0c58c8c-ee03-4646-b274-cf0edf8fe47b-catalog-content\") pod \"redhat-operators-7zbv5\" (UID: \"a0c58c8c-ee03-4646-b274-cf0edf8fe47b\") " pod="openshift-marketplace/redhat-operators-7zbv5" Feb 27 17:46:46 crc kubenswrapper[4700]: I0227 17:46:46.158542 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0c58c8c-ee03-4646-b274-cf0edf8fe47b-utilities\") pod \"redhat-operators-7zbv5\" (UID: \"a0c58c8c-ee03-4646-b274-cf0edf8fe47b\") " pod="openshift-marketplace/redhat-operators-7zbv5" Feb 27 17:46:46 crc kubenswrapper[4700]: I0227 17:46:46.260739 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0c58c8c-ee03-4646-b274-cf0edf8fe47b-utilities\") pod \"redhat-operators-7zbv5\" (UID: \"a0c58c8c-ee03-4646-b274-cf0edf8fe47b\") " pod="openshift-marketplace/redhat-operators-7zbv5" Feb 27 17:46:46 crc kubenswrapper[4700]: I0227 17:46:46.260828 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnkf4\" (UniqueName: \"kubernetes.io/projected/a0c58c8c-ee03-4646-b274-cf0edf8fe47b-kube-api-access-xnkf4\") pod \"redhat-operators-7zbv5\" (UID: \"a0c58c8c-ee03-4646-b274-cf0edf8fe47b\") " pod="openshift-marketplace/redhat-operators-7zbv5" Feb 27 17:46:46 crc kubenswrapper[4700]: I0227 17:46:46.260882 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0c58c8c-ee03-4646-b274-cf0edf8fe47b-catalog-content\") pod \"redhat-operators-7zbv5\" (UID: \"a0c58c8c-ee03-4646-b274-cf0edf8fe47b\") " pod="openshift-marketplace/redhat-operators-7zbv5" Feb 27 17:46:46 crc kubenswrapper[4700]: I0227 17:46:46.261454 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0c58c8c-ee03-4646-b274-cf0edf8fe47b-catalog-content\") pod \"redhat-operators-7zbv5\" (UID: \"a0c58c8c-ee03-4646-b274-cf0edf8fe47b\") " pod="openshift-marketplace/redhat-operators-7zbv5" Feb 27 17:46:46 crc kubenswrapper[4700]: I0227 17:46:46.261519 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0c58c8c-ee03-4646-b274-cf0edf8fe47b-utilities\") pod \"redhat-operators-7zbv5\" (UID: \"a0c58c8c-ee03-4646-b274-cf0edf8fe47b\") " pod="openshift-marketplace/redhat-operators-7zbv5" Feb 27 17:46:46 crc kubenswrapper[4700]: I0227 17:46:46.285473 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnkf4\" (UniqueName: \"kubernetes.io/projected/a0c58c8c-ee03-4646-b274-cf0edf8fe47b-kube-api-access-xnkf4\") pod \"redhat-operators-7zbv5\" (UID: \"a0c58c8c-ee03-4646-b274-cf0edf8fe47b\") " pod="openshift-marketplace/redhat-operators-7zbv5" Feb 27 17:46:46 crc kubenswrapper[4700]: I0227 17:46:46.401862 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7zbv5" Feb 27 17:46:46 crc kubenswrapper[4700]: I0227 17:46:46.852321 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7zbv5"] Feb 27 17:46:47 crc kubenswrapper[4700]: I0227 17:46:47.771346 4700 generic.go:334] "Generic (PLEG): container finished" podID="a0c58c8c-ee03-4646-b274-cf0edf8fe47b" containerID="f3d0c607ee5ee4c60c1e146c8bcd94e878d5ea8512fc5c460aac2d7a6117e257" exitCode=0 Feb 27 17:46:47 crc kubenswrapper[4700]: I0227 17:46:47.771450 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7zbv5" event={"ID":"a0c58c8c-ee03-4646-b274-cf0edf8fe47b","Type":"ContainerDied","Data":"f3d0c607ee5ee4c60c1e146c8bcd94e878d5ea8512fc5c460aac2d7a6117e257"} Feb 27 17:46:47 crc kubenswrapper[4700]: I0227 17:46:47.771752 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7zbv5" event={"ID":"a0c58c8c-ee03-4646-b274-cf0edf8fe47b","Type":"ContainerStarted","Data":"cfd0fe0bf1b46bea86c95779ad93a161702139c36f183a571ec82b74783d68ce"} Feb 27 17:46:47 crc kubenswrapper[4700]: E0227 17:46:47.983380 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536904-fdk8p" podUID="cc474ff3-1a67-415b-9533-eefbe72a6822" Feb 27 17:46:49 crc kubenswrapper[4700]: E0227 17:46:49.853970 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 27 17:46:49 crc kubenswrapper[4700]: E0227 17:46:49.854340 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xnkf4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-7zbv5_openshift-marketplace(a0c58c8c-ee03-4646-b274-cf0edf8fe47b): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 17:46:49 crc kubenswrapper[4700]: E0227 17:46:49.855579 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-operators-7zbv5" podUID="a0c58c8c-ee03-4646-b274-cf0edf8fe47b" Feb 27 17:46:50 crc kubenswrapper[4700]: E0227 17:46:50.813271 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-7zbv5" podUID="a0c58c8c-ee03-4646-b274-cf0edf8fe47b" Feb 27 17:47:01 crc kubenswrapper[4700]: I0227 17:47:01.989644 4700 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 17:47:03 crc kubenswrapper[4700]: E0227 17:47:03.542617 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 27 17:47:03 crc kubenswrapper[4700]: E0227 17:47:03.543087 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xnkf4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-7zbv5_openshift-marketplace(a0c58c8c-ee03-4646-b274-cf0edf8fe47b): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 17:47:03 crc kubenswrapper[4700]: E0227 17:47:03.544420 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-operators-7zbv5" podUID="a0c58c8c-ee03-4646-b274-cf0edf8fe47b" Feb 27 17:47:04 crc kubenswrapper[4700]: I0227 17:47:04.978191 4700 generic.go:334] "Generic (PLEG): container finished" podID="cc474ff3-1a67-415b-9533-eefbe72a6822" containerID="eb24abb5cd44aeda29252db9c820f948348c9e7e730de4c77262eebaf09babcb" exitCode=0 Feb 27 17:47:04 crc kubenswrapper[4700]: I0227 17:47:04.978283 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536904-fdk8p" event={"ID":"cc474ff3-1a67-415b-9533-eefbe72a6822","Type":"ContainerDied","Data":"eb24abb5cd44aeda29252db9c820f948348c9e7e730de4c77262eebaf09babcb"} Feb 27 17:47:06 crc kubenswrapper[4700]: I0227 17:47:06.411057 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:47:06 crc kubenswrapper[4700]: I0227 17:47:06.411402 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:47:06 crc kubenswrapper[4700]: I0227 17:47:06.417925 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536904-fdk8p" Feb 27 17:47:06 crc kubenswrapper[4700]: I0227 17:47:06.530453 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4b7pg\" (UniqueName: \"kubernetes.io/projected/cc474ff3-1a67-415b-9533-eefbe72a6822-kube-api-access-4b7pg\") pod \"cc474ff3-1a67-415b-9533-eefbe72a6822\" (UID: \"cc474ff3-1a67-415b-9533-eefbe72a6822\") " Feb 27 17:47:06 crc kubenswrapper[4700]: I0227 17:47:06.538957 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc474ff3-1a67-415b-9533-eefbe72a6822-kube-api-access-4b7pg" (OuterVolumeSpecName: "kube-api-access-4b7pg") pod "cc474ff3-1a67-415b-9533-eefbe72a6822" (UID: "cc474ff3-1a67-415b-9533-eefbe72a6822"). InnerVolumeSpecName "kube-api-access-4b7pg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:47:06 crc kubenswrapper[4700]: I0227 17:47:06.633147 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4b7pg\" (UniqueName: \"kubernetes.io/projected/cc474ff3-1a67-415b-9533-eefbe72a6822-kube-api-access-4b7pg\") on node \"crc\" DevicePath \"\"" Feb 27 17:47:07 crc kubenswrapper[4700]: I0227 17:47:07.006618 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536904-fdk8p" Feb 27 17:47:07 crc kubenswrapper[4700]: I0227 17:47:07.008371 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536904-fdk8p" event={"ID":"cc474ff3-1a67-415b-9533-eefbe72a6822","Type":"ContainerDied","Data":"80d12144da0eb9e59c18dd64d8b2598353bf3a3ff06df2ac1b49c52f19171a15"} Feb 27 17:47:07 crc kubenswrapper[4700]: I0227 17:47:07.008423 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80d12144da0eb9e59c18dd64d8b2598353bf3a3ff06df2ac1b49c52f19171a15" Feb 27 17:47:07 crc kubenswrapper[4700]: I0227 17:47:07.506800 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536900-xp5cs"] Feb 27 17:47:07 crc kubenswrapper[4700]: I0227 17:47:07.519188 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536900-xp5cs"] Feb 27 17:47:08 crc kubenswrapper[4700]: I0227 17:47:08.999579 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28371b07-fe9e-41c2-bb2d-d2a714510732" path="/var/lib/kubelet/pods/28371b07-fe9e-41c2-bb2d-d2a714510732/volumes" Feb 27 17:47:10 crc kubenswrapper[4700]: I0227 17:47:10.010566 4700 scope.go:117] "RemoveContainer" containerID="15995fc6325bfa84bcf6c7aed97aea03c4466ba91003e4703e6648dd4081aa78" Feb 27 17:47:10 crc kubenswrapper[4700]: I0227 17:47:10.073932 4700 scope.go:117] "RemoveContainer" containerID="bf1f073c4ee8e6cebb4d8470d6277d6ce6aebb31fd4dd9396283535b9ee76e1c" Feb 27 17:47:14 crc kubenswrapper[4700]: E0227 17:47:14.985081 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-7zbv5" podUID="a0c58c8c-ee03-4646-b274-cf0edf8fe47b" Feb 27 17:47:30 crc kubenswrapper[4700]: I0227 17:47:30.288527 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7zbv5" event={"ID":"a0c58c8c-ee03-4646-b274-cf0edf8fe47b","Type":"ContainerStarted","Data":"cb9ea029b232da438b5508b4b968a8c448b8e401ea26546a55bb86c0e92e6b46"} Feb 27 17:47:35 crc kubenswrapper[4700]: I0227 17:47:35.353076 4700 generic.go:334] "Generic (PLEG): container finished" podID="a0c58c8c-ee03-4646-b274-cf0edf8fe47b" containerID="cb9ea029b232da438b5508b4b968a8c448b8e401ea26546a55bb86c0e92e6b46" exitCode=0 Feb 27 17:47:35 crc kubenswrapper[4700]: I0227 17:47:35.353171 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7zbv5" event={"ID":"a0c58c8c-ee03-4646-b274-cf0edf8fe47b","Type":"ContainerDied","Data":"cb9ea029b232da438b5508b4b968a8c448b8e401ea26546a55bb86c0e92e6b46"} Feb 27 17:47:36 crc kubenswrapper[4700]: I0227 17:47:36.370907 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7zbv5" event={"ID":"a0c58c8c-ee03-4646-b274-cf0edf8fe47b","Type":"ContainerStarted","Data":"8325eb7946129083d7860130cd00d03bdeb8c6a8abc995a5771a4a827ea5a483"} Feb 27 17:47:36 crc kubenswrapper[4700]: I0227 17:47:36.402915 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7zbv5" Feb 27 17:47:36 crc kubenswrapper[4700]: I0227 17:47:36.402985 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7zbv5" Feb 27 17:47:36 crc kubenswrapper[4700]: I0227 17:47:36.410918 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:47:36 crc kubenswrapper[4700]: I0227 17:47:36.411005 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:47:36 crc kubenswrapper[4700]: I0227 17:47:36.422076 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7zbv5" podStartSLOduration=2.414524181 podStartE2EDuration="50.422037948s" podCreationTimestamp="2026-02-27 17:46:46 +0000 UTC" firstStartedPulling="2026-02-27 17:46:47.77413576 +0000 UTC m=+2767.759448517" lastFinishedPulling="2026-02-27 17:47:35.781649507 +0000 UTC m=+2815.766962284" observedRunningTime="2026-02-27 17:47:36.403682846 +0000 UTC m=+2816.388995623" watchObservedRunningTime="2026-02-27 17:47:36.422037948 +0000 UTC m=+2816.407350735" Feb 27 17:47:37 crc kubenswrapper[4700]: I0227 17:47:37.474270 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7zbv5" podUID="a0c58c8c-ee03-4646-b274-cf0edf8fe47b" containerName="registry-server" probeResult="failure" output=< Feb 27 17:47:37 crc kubenswrapper[4700]: timeout: failed to connect service ":50051" within 1s Feb 27 17:47:37 crc kubenswrapper[4700]: > Feb 27 17:47:46 crc kubenswrapper[4700]: I0227 17:47:46.465426 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7zbv5" Feb 27 17:47:46 crc kubenswrapper[4700]: I0227 17:47:46.528105 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7zbv5" Feb 27 17:47:47 crc kubenswrapper[4700]: I0227 17:47:47.294377 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7zbv5"] Feb 27 17:47:48 crc kubenswrapper[4700]: I0227 17:47:48.488092 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7zbv5" podUID="a0c58c8c-ee03-4646-b274-cf0edf8fe47b" containerName="registry-server" containerID="cri-o://8325eb7946129083d7860130cd00d03bdeb8c6a8abc995a5771a4a827ea5a483" gracePeriod=2 Feb 27 17:47:48 crc kubenswrapper[4700]: I0227 17:47:48.953796 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7zbv5" Feb 27 17:47:49 crc kubenswrapper[4700]: I0227 17:47:49.079428 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnkf4\" (UniqueName: \"kubernetes.io/projected/a0c58c8c-ee03-4646-b274-cf0edf8fe47b-kube-api-access-xnkf4\") pod \"a0c58c8c-ee03-4646-b274-cf0edf8fe47b\" (UID: \"a0c58c8c-ee03-4646-b274-cf0edf8fe47b\") " Feb 27 17:47:49 crc kubenswrapper[4700]: I0227 17:47:49.079574 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0c58c8c-ee03-4646-b274-cf0edf8fe47b-catalog-content\") pod \"a0c58c8c-ee03-4646-b274-cf0edf8fe47b\" (UID: \"a0c58c8c-ee03-4646-b274-cf0edf8fe47b\") " Feb 27 17:47:49 crc kubenswrapper[4700]: I0227 17:47:49.079678 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0c58c8c-ee03-4646-b274-cf0edf8fe47b-utilities\") pod \"a0c58c8c-ee03-4646-b274-cf0edf8fe47b\" (UID: \"a0c58c8c-ee03-4646-b274-cf0edf8fe47b\") " Feb 27 17:47:49 crc kubenswrapper[4700]: I0227 17:47:49.081257 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0c58c8c-ee03-4646-b274-cf0edf8fe47b-utilities" (OuterVolumeSpecName: "utilities") pod "a0c58c8c-ee03-4646-b274-cf0edf8fe47b" (UID: "a0c58c8c-ee03-4646-b274-cf0edf8fe47b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:47:49 crc kubenswrapper[4700]: I0227 17:47:49.082272 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0c58c8c-ee03-4646-b274-cf0edf8fe47b-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:47:49 crc kubenswrapper[4700]: I0227 17:47:49.086861 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0c58c8c-ee03-4646-b274-cf0edf8fe47b-kube-api-access-xnkf4" (OuterVolumeSpecName: "kube-api-access-xnkf4") pod "a0c58c8c-ee03-4646-b274-cf0edf8fe47b" (UID: "a0c58c8c-ee03-4646-b274-cf0edf8fe47b"). InnerVolumeSpecName "kube-api-access-xnkf4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:47:49 crc kubenswrapper[4700]: I0227 17:47:49.184087 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnkf4\" (UniqueName: \"kubernetes.io/projected/a0c58c8c-ee03-4646-b274-cf0edf8fe47b-kube-api-access-xnkf4\") on node \"crc\" DevicePath \"\"" Feb 27 17:47:49 crc kubenswrapper[4700]: I0227 17:47:49.209496 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0c58c8c-ee03-4646-b274-cf0edf8fe47b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a0c58c8c-ee03-4646-b274-cf0edf8fe47b" (UID: "a0c58c8c-ee03-4646-b274-cf0edf8fe47b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:47:49 crc kubenswrapper[4700]: I0227 17:47:49.286065 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0c58c8c-ee03-4646-b274-cf0edf8fe47b-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:47:49 crc kubenswrapper[4700]: I0227 17:47:49.499319 4700 generic.go:334] "Generic (PLEG): container finished" podID="a0c58c8c-ee03-4646-b274-cf0edf8fe47b" containerID="8325eb7946129083d7860130cd00d03bdeb8c6a8abc995a5771a4a827ea5a483" exitCode=0 Feb 27 17:47:49 crc kubenswrapper[4700]: I0227 17:47:49.499402 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7zbv5" event={"ID":"a0c58c8c-ee03-4646-b274-cf0edf8fe47b","Type":"ContainerDied","Data":"8325eb7946129083d7860130cd00d03bdeb8c6a8abc995a5771a4a827ea5a483"} Feb 27 17:47:49 crc kubenswrapper[4700]: I0227 17:47:49.499432 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7zbv5" event={"ID":"a0c58c8c-ee03-4646-b274-cf0edf8fe47b","Type":"ContainerDied","Data":"cfd0fe0bf1b46bea86c95779ad93a161702139c36f183a571ec82b74783d68ce"} Feb 27 17:47:49 crc kubenswrapper[4700]: I0227 17:47:49.499453 4700 scope.go:117] "RemoveContainer" containerID="8325eb7946129083d7860130cd00d03bdeb8c6a8abc995a5771a4a827ea5a483" Feb 27 17:47:49 crc kubenswrapper[4700]: I0227 17:47:49.499633 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7zbv5" Feb 27 17:47:49 crc kubenswrapper[4700]: I0227 17:47:49.535590 4700 scope.go:117] "RemoveContainer" containerID="cb9ea029b232da438b5508b4b968a8c448b8e401ea26546a55bb86c0e92e6b46" Feb 27 17:47:49 crc kubenswrapper[4700]: I0227 17:47:49.541386 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7zbv5"] Feb 27 17:47:49 crc kubenswrapper[4700]: I0227 17:47:49.551579 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7zbv5"] Feb 27 17:47:49 crc kubenswrapper[4700]: I0227 17:47:49.568204 4700 scope.go:117] "RemoveContainer" containerID="f3d0c607ee5ee4c60c1e146c8bcd94e878d5ea8512fc5c460aac2d7a6117e257" Feb 27 17:47:49 crc kubenswrapper[4700]: I0227 17:47:49.628154 4700 scope.go:117] "RemoveContainer" containerID="8325eb7946129083d7860130cd00d03bdeb8c6a8abc995a5771a4a827ea5a483" Feb 27 17:47:49 crc kubenswrapper[4700]: E0227 17:47:49.628661 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8325eb7946129083d7860130cd00d03bdeb8c6a8abc995a5771a4a827ea5a483\": container with ID starting with 8325eb7946129083d7860130cd00d03bdeb8c6a8abc995a5771a4a827ea5a483 not found: ID does not exist" containerID="8325eb7946129083d7860130cd00d03bdeb8c6a8abc995a5771a4a827ea5a483" Feb 27 17:47:49 crc kubenswrapper[4700]: I0227 17:47:49.628707 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8325eb7946129083d7860130cd00d03bdeb8c6a8abc995a5771a4a827ea5a483"} err="failed to get container status \"8325eb7946129083d7860130cd00d03bdeb8c6a8abc995a5771a4a827ea5a483\": rpc error: code = NotFound desc = could not find container \"8325eb7946129083d7860130cd00d03bdeb8c6a8abc995a5771a4a827ea5a483\": container with ID starting with 8325eb7946129083d7860130cd00d03bdeb8c6a8abc995a5771a4a827ea5a483 not found: ID does not exist" Feb 27 17:47:49 crc kubenswrapper[4700]: I0227 17:47:49.628736 4700 scope.go:117] "RemoveContainer" containerID="cb9ea029b232da438b5508b4b968a8c448b8e401ea26546a55bb86c0e92e6b46" Feb 27 17:47:49 crc kubenswrapper[4700]: E0227 17:47:49.629101 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb9ea029b232da438b5508b4b968a8c448b8e401ea26546a55bb86c0e92e6b46\": container with ID starting with cb9ea029b232da438b5508b4b968a8c448b8e401ea26546a55bb86c0e92e6b46 not found: ID does not exist" containerID="cb9ea029b232da438b5508b4b968a8c448b8e401ea26546a55bb86c0e92e6b46" Feb 27 17:47:49 crc kubenswrapper[4700]: I0227 17:47:49.629144 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb9ea029b232da438b5508b4b968a8c448b8e401ea26546a55bb86c0e92e6b46"} err="failed to get container status \"cb9ea029b232da438b5508b4b968a8c448b8e401ea26546a55bb86c0e92e6b46\": rpc error: code = NotFound desc = could not find container \"cb9ea029b232da438b5508b4b968a8c448b8e401ea26546a55bb86c0e92e6b46\": container with ID starting with cb9ea029b232da438b5508b4b968a8c448b8e401ea26546a55bb86c0e92e6b46 not found: ID does not exist" Feb 27 17:47:49 crc kubenswrapper[4700]: I0227 17:47:49.629166 4700 scope.go:117] "RemoveContainer" containerID="f3d0c607ee5ee4c60c1e146c8bcd94e878d5ea8512fc5c460aac2d7a6117e257" Feb 27 17:47:49 crc kubenswrapper[4700]: E0227 17:47:49.629656 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3d0c607ee5ee4c60c1e146c8bcd94e878d5ea8512fc5c460aac2d7a6117e257\": container with ID starting with f3d0c607ee5ee4c60c1e146c8bcd94e878d5ea8512fc5c460aac2d7a6117e257 not found: ID does not exist" containerID="f3d0c607ee5ee4c60c1e146c8bcd94e878d5ea8512fc5c460aac2d7a6117e257" Feb 27 17:47:49 crc kubenswrapper[4700]: I0227 17:47:49.629710 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3d0c607ee5ee4c60c1e146c8bcd94e878d5ea8512fc5c460aac2d7a6117e257"} err="failed to get container status \"f3d0c607ee5ee4c60c1e146c8bcd94e878d5ea8512fc5c460aac2d7a6117e257\": rpc error: code = NotFound desc = could not find container \"f3d0c607ee5ee4c60c1e146c8bcd94e878d5ea8512fc5c460aac2d7a6117e257\": container with ID starting with f3d0c607ee5ee4c60c1e146c8bcd94e878d5ea8512fc5c460aac2d7a6117e257 not found: ID does not exist" Feb 27 17:47:51 crc kubenswrapper[4700]: I0227 17:47:51.020719 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0c58c8c-ee03-4646-b274-cf0edf8fe47b" path="/var/lib/kubelet/pods/a0c58c8c-ee03-4646-b274-cf0edf8fe47b/volumes" Feb 27 17:47:58 crc kubenswrapper[4700]: I0227 17:47:58.586441 4700 generic.go:334] "Generic (PLEG): container finished" podID="d9b66d11-37c1-4826-96c3-967dd98f053c" containerID="036b10b81dd522449d3bc6674c8f9d82fa8fda100f6b0f4274142b0d5446efcd" exitCode=0 Feb 27 17:47:58 crc kubenswrapper[4700]: I0227 17:47:58.586523 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn" event={"ID":"d9b66d11-37c1-4826-96c3-967dd98f053c","Type":"ContainerDied","Data":"036b10b81dd522449d3bc6674c8f9d82fa8fda100f6b0f4274142b0d5446efcd"} Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.121000 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn" Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.151374 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536908-djpzb"] Feb 27 17:48:00 crc kubenswrapper[4700]: E0227 17:48:00.151902 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc474ff3-1a67-415b-9533-eefbe72a6822" containerName="oc" Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.151923 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc474ff3-1a67-415b-9533-eefbe72a6822" containerName="oc" Feb 27 17:48:00 crc kubenswrapper[4700]: E0227 17:48:00.151948 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0c58c8c-ee03-4646-b274-cf0edf8fe47b" containerName="extract-utilities" Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.151955 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0c58c8c-ee03-4646-b274-cf0edf8fe47b" containerName="extract-utilities" Feb 27 17:48:00 crc kubenswrapper[4700]: E0227 17:48:00.151971 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0c58c8c-ee03-4646-b274-cf0edf8fe47b" containerName="registry-server" Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.151978 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0c58c8c-ee03-4646-b274-cf0edf8fe47b" containerName="registry-server" Feb 27 17:48:00 crc kubenswrapper[4700]: E0227 17:48:00.151998 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9b66d11-37c1-4826-96c3-967dd98f053c" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.152005 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9b66d11-37c1-4826-96c3-967dd98f053c" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 27 17:48:00 crc kubenswrapper[4700]: E0227 17:48:00.152016 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0c58c8c-ee03-4646-b274-cf0edf8fe47b" containerName="extract-content" Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.152022 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0c58c8c-ee03-4646-b274-cf0edf8fe47b" containerName="extract-content" Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.152195 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9b66d11-37c1-4826-96c3-967dd98f053c" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.152210 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0c58c8c-ee03-4646-b274-cf0edf8fe47b" containerName="registry-server" Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.152217 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc474ff3-1a67-415b-9533-eefbe72a6822" containerName="oc" Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.152935 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536908-djpzb" Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.155728 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.155987 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d62kc" Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.157889 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.170946 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536908-djpzb"] Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.230688 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fj4x\" (UniqueName: \"kubernetes.io/projected/d9b66d11-37c1-4826-96c3-967dd98f053c-kube-api-access-2fj4x\") pod \"d9b66d11-37c1-4826-96c3-967dd98f053c\" (UID: \"d9b66d11-37c1-4826-96c3-967dd98f053c\") " Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.230748 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9b66d11-37c1-4826-96c3-967dd98f053c-telemetry-combined-ca-bundle\") pod \"d9b66d11-37c1-4826-96c3-967dd98f053c\" (UID: \"d9b66d11-37c1-4826-96c3-967dd98f053c\") " Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.230774 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d9b66d11-37c1-4826-96c3-967dd98f053c-ssh-key-openstack-edpm-ipam\") pod \"d9b66d11-37c1-4826-96c3-967dd98f053c\" (UID: \"d9b66d11-37c1-4826-96c3-967dd98f053c\") " Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.230806 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9b66d11-37c1-4826-96c3-967dd98f053c-inventory\") pod \"d9b66d11-37c1-4826-96c3-967dd98f053c\" (UID: \"d9b66d11-37c1-4826-96c3-967dd98f053c\") " Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.230834 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d9b66d11-37c1-4826-96c3-967dd98f053c-ceilometer-compute-config-data-0\") pod \"d9b66d11-37c1-4826-96c3-967dd98f053c\" (UID: \"d9b66d11-37c1-4826-96c3-967dd98f053c\") " Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.230882 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d9b66d11-37c1-4826-96c3-967dd98f053c-ceilometer-compute-config-data-2\") pod \"d9b66d11-37c1-4826-96c3-967dd98f053c\" (UID: \"d9b66d11-37c1-4826-96c3-967dd98f053c\") " Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.230946 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d9b66d11-37c1-4826-96c3-967dd98f053c-ceilometer-compute-config-data-1\") pod \"d9b66d11-37c1-4826-96c3-967dd98f053c\" (UID: \"d9b66d11-37c1-4826-96c3-967dd98f053c\") " Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.236660 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9b66d11-37c1-4826-96c3-967dd98f053c-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "d9b66d11-37c1-4826-96c3-967dd98f053c" (UID: "d9b66d11-37c1-4826-96c3-967dd98f053c"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.236874 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9b66d11-37c1-4826-96c3-967dd98f053c-kube-api-access-2fj4x" (OuterVolumeSpecName: "kube-api-access-2fj4x") pod "d9b66d11-37c1-4826-96c3-967dd98f053c" (UID: "d9b66d11-37c1-4826-96c3-967dd98f053c"). InnerVolumeSpecName "kube-api-access-2fj4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.261360 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9b66d11-37c1-4826-96c3-967dd98f053c-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "d9b66d11-37c1-4826-96c3-967dd98f053c" (UID: "d9b66d11-37c1-4826-96c3-967dd98f053c"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.264801 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9b66d11-37c1-4826-96c3-967dd98f053c-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "d9b66d11-37c1-4826-96c3-967dd98f053c" (UID: "d9b66d11-37c1-4826-96c3-967dd98f053c"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.267135 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9b66d11-37c1-4826-96c3-967dd98f053c-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "d9b66d11-37c1-4826-96c3-967dd98f053c" (UID: "d9b66d11-37c1-4826-96c3-967dd98f053c"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.269415 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9b66d11-37c1-4826-96c3-967dd98f053c-inventory" (OuterVolumeSpecName: "inventory") pod "d9b66d11-37c1-4826-96c3-967dd98f053c" (UID: "d9b66d11-37c1-4826-96c3-967dd98f053c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.278335 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9b66d11-37c1-4826-96c3-967dd98f053c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "d9b66d11-37c1-4826-96c3-967dd98f053c" (UID: "d9b66d11-37c1-4826-96c3-967dd98f053c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.335082 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzkfq\" (UniqueName: \"kubernetes.io/projected/c22f6346-6080-4170-beff-4fe69f2aa293-kube-api-access-fzkfq\") pod \"auto-csr-approver-29536908-djpzb\" (UID: \"c22f6346-6080-4170-beff-4fe69f2aa293\") " pod="openshift-infra/auto-csr-approver-29536908-djpzb" Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.335338 4700 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9b66d11-37c1-4826-96c3-967dd98f053c-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.335361 4700 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d9b66d11-37c1-4826-96c3-967dd98f053c-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.335376 4700 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d9b66d11-37c1-4826-96c3-967dd98f053c-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.335389 4700 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d9b66d11-37c1-4826-96c3-967dd98f053c-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.335406 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fj4x\" (UniqueName: \"kubernetes.io/projected/d9b66d11-37c1-4826-96c3-967dd98f053c-kube-api-access-2fj4x\") on node \"crc\" DevicePath \"\"" Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.335419 4700 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9b66d11-37c1-4826-96c3-967dd98f053c-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.335430 4700 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d9b66d11-37c1-4826-96c3-967dd98f053c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.437355 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzkfq\" (UniqueName: \"kubernetes.io/projected/c22f6346-6080-4170-beff-4fe69f2aa293-kube-api-access-fzkfq\") pod \"auto-csr-approver-29536908-djpzb\" (UID: \"c22f6346-6080-4170-beff-4fe69f2aa293\") " pod="openshift-infra/auto-csr-approver-29536908-djpzb" Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.455077 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzkfq\" (UniqueName: \"kubernetes.io/projected/c22f6346-6080-4170-beff-4fe69f2aa293-kube-api-access-fzkfq\") pod \"auto-csr-approver-29536908-djpzb\" (UID: \"c22f6346-6080-4170-beff-4fe69f2aa293\") " pod="openshift-infra/auto-csr-approver-29536908-djpzb" Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.480747 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536908-djpzb" Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.606051 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn" event={"ID":"d9b66d11-37c1-4826-96c3-967dd98f053c","Type":"ContainerDied","Data":"64f1c4919a0cfae19d3b86eb8d43ff32a9dd65e6d9c52505a0f7352cae7316e0"} Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.606300 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64f1c4919a0cfae19d3b86eb8d43ff32a9dd65e6d9c52505a0f7352cae7316e0" Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.606357 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn" Feb 27 17:48:00 crc kubenswrapper[4700]: I0227 17:48:00.995885 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536908-djpzb"] Feb 27 17:48:01 crc kubenswrapper[4700]: I0227 17:48:01.626814 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536908-djpzb" event={"ID":"c22f6346-6080-4170-beff-4fe69f2aa293","Type":"ContainerStarted","Data":"61a133c9fa354d556645b703b0da1c0bf30682eda133675eba7ebba56d771909"} Feb 27 17:48:02 crc kubenswrapper[4700]: E0227 17:48:02.022678 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 17:48:02 crc kubenswrapper[4700]: E0227 17:48:02.022940 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 17:48:02 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 17:48:02 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fzkfq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536908-djpzb_openshift-infra(c22f6346-6080-4170-beff-4fe69f2aa293): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 17:48:02 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 17:48:02 crc kubenswrapper[4700]: E0227 17:48:02.024175 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536908-djpzb" podUID="c22f6346-6080-4170-beff-4fe69f2aa293" Feb 27 17:48:02 crc kubenswrapper[4700]: E0227 17:48:02.657945 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536908-djpzb" podUID="c22f6346-6080-4170-beff-4fe69f2aa293" Feb 27 17:48:06 crc kubenswrapper[4700]: I0227 17:48:06.410714 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:48:06 crc kubenswrapper[4700]: I0227 17:48:06.411795 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:48:06 crc kubenswrapper[4700]: I0227 17:48:06.411875 4700 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" Feb 27 17:48:06 crc kubenswrapper[4700]: I0227 17:48:06.412897 4700 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9b478be1943d618189bcf0c56cb81fd36080f5aa6f5ea1159e5d5ecbda877263"} pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 17:48:06 crc kubenswrapper[4700]: I0227 17:48:06.412980 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" containerID="cri-o://9b478be1943d618189bcf0c56cb81fd36080f5aa6f5ea1159e5d5ecbda877263" gracePeriod=600 Feb 27 17:48:06 crc kubenswrapper[4700]: I0227 17:48:06.699803 4700 generic.go:334] "Generic (PLEG): container finished" podID="84b45511-b94c-479f-98db-fd2c4eceec46" containerID="9b478be1943d618189bcf0c56cb81fd36080f5aa6f5ea1159e5d5ecbda877263" exitCode=0 Feb 27 17:48:06 crc kubenswrapper[4700]: I0227 17:48:06.699873 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerDied","Data":"9b478be1943d618189bcf0c56cb81fd36080f5aa6f5ea1159e5d5ecbda877263"} Feb 27 17:48:06 crc kubenswrapper[4700]: I0227 17:48:06.699931 4700 scope.go:117] "RemoveContainer" containerID="6843fae59a5a033363e409d439168388b8322ceaaef013a2a8f29582e161bd02" Feb 27 17:48:07 crc kubenswrapper[4700]: I0227 17:48:07.712968 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerStarted","Data":"8c94c84e8861b740764e2dec223048597911c5008bf61ad17e7b62bcf54abe22"} Feb 27 17:48:14 crc kubenswrapper[4700]: E0227 17:48:14.931956 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 17:48:14 crc kubenswrapper[4700]: E0227 17:48:14.932529 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 17:48:14 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 17:48:14 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fzkfq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536908-djpzb_openshift-infra(c22f6346-6080-4170-beff-4fe69f2aa293): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 17:48:14 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 17:48:14 crc kubenswrapper[4700]: E0227 17:48:14.933694 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536908-djpzb" podUID="c22f6346-6080-4170-beff-4fe69f2aa293" Feb 27 17:48:25 crc kubenswrapper[4700]: E0227 17:48:25.987445 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536908-djpzb" podUID="c22f6346-6080-4170-beff-4fe69f2aa293" Feb 27 17:48:32 crc kubenswrapper[4700]: I0227 17:48:32.251075 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-269hv"] Feb 27 17:48:32 crc kubenswrapper[4700]: I0227 17:48:32.254739 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-269hv" Feb 27 17:48:32 crc kubenswrapper[4700]: I0227 17:48:32.290768 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-269hv"] Feb 27 17:48:32 crc kubenswrapper[4700]: I0227 17:48:32.369274 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd83e054-9228-4520-ba37-e7b5fbfa05d4-utilities\") pod \"certified-operators-269hv\" (UID: \"dd83e054-9228-4520-ba37-e7b5fbfa05d4\") " pod="openshift-marketplace/certified-operators-269hv" Feb 27 17:48:32 crc kubenswrapper[4700]: I0227 17:48:32.369682 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd83e054-9228-4520-ba37-e7b5fbfa05d4-catalog-content\") pod \"certified-operators-269hv\" (UID: \"dd83e054-9228-4520-ba37-e7b5fbfa05d4\") " pod="openshift-marketplace/certified-operators-269hv" Feb 27 17:48:32 crc kubenswrapper[4700]: I0227 17:48:32.369704 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngshj\" (UniqueName: \"kubernetes.io/projected/dd83e054-9228-4520-ba37-e7b5fbfa05d4-kube-api-access-ngshj\") pod \"certified-operators-269hv\" (UID: \"dd83e054-9228-4520-ba37-e7b5fbfa05d4\") " pod="openshift-marketplace/certified-operators-269hv" Feb 27 17:48:32 crc kubenswrapper[4700]: I0227 17:48:32.471786 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd83e054-9228-4520-ba37-e7b5fbfa05d4-utilities\") pod \"certified-operators-269hv\" (UID: \"dd83e054-9228-4520-ba37-e7b5fbfa05d4\") " pod="openshift-marketplace/certified-operators-269hv" Feb 27 17:48:32 crc kubenswrapper[4700]: I0227 17:48:32.471999 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd83e054-9228-4520-ba37-e7b5fbfa05d4-catalog-content\") pod \"certified-operators-269hv\" (UID: \"dd83e054-9228-4520-ba37-e7b5fbfa05d4\") " pod="openshift-marketplace/certified-operators-269hv" Feb 27 17:48:32 crc kubenswrapper[4700]: I0227 17:48:32.472038 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngshj\" (UniqueName: \"kubernetes.io/projected/dd83e054-9228-4520-ba37-e7b5fbfa05d4-kube-api-access-ngshj\") pod \"certified-operators-269hv\" (UID: \"dd83e054-9228-4520-ba37-e7b5fbfa05d4\") " pod="openshift-marketplace/certified-operators-269hv" Feb 27 17:48:32 crc kubenswrapper[4700]: I0227 17:48:32.473259 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd83e054-9228-4520-ba37-e7b5fbfa05d4-utilities\") pod \"certified-operators-269hv\" (UID: \"dd83e054-9228-4520-ba37-e7b5fbfa05d4\") " pod="openshift-marketplace/certified-operators-269hv" Feb 27 17:48:32 crc kubenswrapper[4700]: I0227 17:48:32.473771 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd83e054-9228-4520-ba37-e7b5fbfa05d4-catalog-content\") pod \"certified-operators-269hv\" (UID: \"dd83e054-9228-4520-ba37-e7b5fbfa05d4\") " pod="openshift-marketplace/certified-operators-269hv" Feb 27 17:48:32 crc kubenswrapper[4700]: I0227 17:48:32.492839 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngshj\" (UniqueName: \"kubernetes.io/projected/dd83e054-9228-4520-ba37-e7b5fbfa05d4-kube-api-access-ngshj\") pod \"certified-operators-269hv\" (UID: \"dd83e054-9228-4520-ba37-e7b5fbfa05d4\") " pod="openshift-marketplace/certified-operators-269hv" Feb 27 17:48:32 crc kubenswrapper[4700]: I0227 17:48:32.597982 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-269hv" Feb 27 17:48:33 crc kubenswrapper[4700]: I0227 17:48:33.190477 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-269hv"] Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.020045 4700 generic.go:334] "Generic (PLEG): container finished" podID="dd83e054-9228-4520-ba37-e7b5fbfa05d4" containerID="bba6e6cd87914bd634237083804ee060ecd6bea7eccab47c006e39431e7c7cfe" exitCode=0 Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.020115 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-269hv" event={"ID":"dd83e054-9228-4520-ba37-e7b5fbfa05d4","Type":"ContainerDied","Data":"bba6e6cd87914bd634237083804ee060ecd6bea7eccab47c006e39431e7c7cfe"} Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.020437 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-269hv" event={"ID":"dd83e054-9228-4520-ba37-e7b5fbfa05d4","Type":"ContainerStarted","Data":"2a67f31d5b33bcbeca79ecd5fabc74ee7c61fc00dc27956d47e7c7a6364f1041"} Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.501597 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.503360 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.506571 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.528112 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.617043 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-nfs-0"] Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.618149 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-sys\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.618215 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.618299 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-config-data\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.618345 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-dev\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.618389 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-config-data-custom\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.618416 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.618443 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.618490 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqjf5\" (UniqueName: \"kubernetes.io/projected/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-kube-api-access-wqjf5\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.618516 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-run\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.618539 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-etc-nvme\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.618553 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-lib-modules\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.618580 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.618602 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.618637 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.618671 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-scripts\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.618996 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.627204 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-nfs-config-data" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.636430 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-nfs-0"] Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.687556 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-nfs-2-0"] Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.695873 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.706919 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-nfs-2-config-data" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.720303 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-etc-machine-id\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.720382 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-dev\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.720539 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-run\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.720588 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-config-data-custom\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.720618 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.720647 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.720672 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-etc-nvme\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.720700 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqjf5\" (UniqueName: \"kubernetes.io/projected/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-kube-api-access-wqjf5\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.720728 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-var-lib-cinder\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.720751 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-run\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.720787 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-etc-nvme\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.720809 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-lib-modules\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.720845 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.720871 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.720907 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.720929 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-var-locks-cinder\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.720978 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-config-data-custom\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.721001 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-scripts\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.721031 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-etc-iscsi\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.721054 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-config-data\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.721092 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-scripts\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.721114 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-sys\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.721150 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.721181 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-sys\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.721215 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttd8f\" (UniqueName: \"kubernetes.io/projected/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-kube-api-access-ttd8f\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.721274 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-dev\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.721299 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-var-locks-brick\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.721324 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-combined-ca-bundle\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.721347 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-lib-modules\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.721382 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-config-data\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.721756 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.721823 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-dev\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.722185 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-sys\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.722862 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.723235 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-lib-modules\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.723288 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-run\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.723854 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.723977 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-etc-nvme\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.725408 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.725507 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.728113 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-scripts\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.728793 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-config-data-custom\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.729763 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.733552 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-nfs-2-0"] Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.741195 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqjf5\" (UniqueName: \"kubernetes.io/projected/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-kube-api-access-wqjf5\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.746889 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4891dc6b-c6d7-4468-ab14-e22ccc8462d2-config-data\") pod \"cinder-backup-0\" (UID: \"4891dc6b-c6d7-4468-ab14-e22ccc8462d2\") " pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.823080 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-var-locks-brick\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.823140 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-combined-ca-bundle\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.823173 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-var-locks-cinder\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.823200 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-lib-modules\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.823221 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-var-locks-brick\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.823259 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-etc-machine-id\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.823279 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-lib-modules\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.823330 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-config-data-custom\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.823359 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-etc-machine-id\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.823395 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-run\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.823380 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-run\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.823537 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-sys\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.823607 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-scripts\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.823746 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-etc-nvme\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.823817 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-lib-modules\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.823835 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-etc-nvme\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.823875 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-var-lib-cinder\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.823893 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-etc-nvme\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.823996 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-var-lib-cinder\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.824018 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-var-locks-cinder\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.824064 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57zmg\" (UniqueName: \"kubernetes.io/projected/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-kube-api-access-57zmg\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.824085 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-run\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.824088 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-var-locks-cinder\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.824113 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-config-data-custom\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.824147 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-var-locks-brick\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.824190 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-combined-ca-bundle\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.824208 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-etc-iscsi\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.824231 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-config-data\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.824269 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-etc-iscsi\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.824264 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-config-data\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.824301 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-etc-iscsi\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.824320 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-etc-machine-id\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.824337 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-var-lib-cinder\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.824367 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-scripts\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.824398 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-dev\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.824441 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-sys\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.824489 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttd8f\" (UniqueName: \"kubernetes.io/projected/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-kube-api-access-ttd8f\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.824598 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-dev\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.824689 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-dev\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.824714 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-sys\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.827669 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-scripts\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.827821 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-config-data\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.828020 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-combined-ca-bundle\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.828302 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-config-data-custom\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.829207 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.845250 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttd8f\" (UniqueName: \"kubernetes.io/projected/4c3e4a38-4e77-4f9d-aa8d-355b41080b84-kube-api-access-ttd8f\") pod \"cinder-volume-nfs-0\" (UID: \"4c3e4a38-4e77-4f9d-aa8d-355b41080b84\") " pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.925894 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57zmg\" (UniqueName: \"kubernetes.io/projected/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-kube-api-access-57zmg\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.925968 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-run\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.926004 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-var-locks-brick\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.926034 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-combined-ca-bundle\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.926065 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-etc-iscsi\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.926087 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-config-data\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.926108 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-var-lib-cinder\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.926132 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-etc-machine-id\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.926165 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-dev\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.926239 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-var-locks-cinder\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.926446 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-config-data-custom\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.926517 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-sys\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.926563 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-scripts\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.926775 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-lib-modules\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.926819 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-etc-nvme\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.926994 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-etc-nvme\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.927331 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-run\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.927396 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-var-locks-brick\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.927935 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-etc-iscsi\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.928122 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-var-locks-cinder\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.928188 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-var-lib-cinder\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.928246 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-etc-machine-id\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.928283 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-dev\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.928305 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-sys\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.928546 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-lib-modules\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.931527 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-config-data\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.931565 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-config-data-custom\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.932158 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-combined-ca-bundle\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.932922 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-scripts\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.947204 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:34 crc kubenswrapper[4700]: I0227 17:48:34.952843 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57zmg\" (UniqueName: \"kubernetes.io/projected/0e4cd883-f86f-43d5-b2f2-eb96bd9ed323-kube-api-access-57zmg\") pod \"cinder-volume-nfs-2-0\" (UID: \"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323\") " pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:35 crc kubenswrapper[4700]: I0227 17:48:35.024917 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:35 crc kubenswrapper[4700]: I0227 17:48:35.394663 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-nfs-0"] Feb 27 17:48:35 crc kubenswrapper[4700]: W0227 17:48:35.427160 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c3e4a38_4e77_4f9d_aa8d_355b41080b84.slice/crio-8a0bbaf7405cf09313dbef7438af20a3dab93390dd327fdce64b90a87f78cb31 WatchSource:0}: Error finding container 8a0bbaf7405cf09313dbef7438af20a3dab93390dd327fdce64b90a87f78cb31: Status 404 returned error can't find the container with id 8a0bbaf7405cf09313dbef7438af20a3dab93390dd327fdce64b90a87f78cb31 Feb 27 17:48:35 crc kubenswrapper[4700]: I0227 17:48:35.579230 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Feb 27 17:48:35 crc kubenswrapper[4700]: I0227 17:48:35.922868 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-nfs-2-0"] Feb 27 17:48:36 crc kubenswrapper[4700]: I0227 17:48:36.071008 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"4891dc6b-c6d7-4468-ab14-e22ccc8462d2","Type":"ContainerStarted","Data":"170ca62218653b1b5de9f54c23e7f2562ea1ac89b57f700bd1d19bfdfe14f274"} Feb 27 17:48:36 crc kubenswrapper[4700]: I0227 17:48:36.072105 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-nfs-2-0" event={"ID":"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323","Type":"ContainerStarted","Data":"530e3f18aff534ea7f3fffac91338e32977f4cf8c052b4198394687e3527fb07"} Feb 27 17:48:36 crc kubenswrapper[4700]: I0227 17:48:36.076510 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-269hv" event={"ID":"dd83e054-9228-4520-ba37-e7b5fbfa05d4","Type":"ContainerStarted","Data":"4bde31f8b98e3d4ff3c8c900be8a29b5d31326d7427085c2a74090c58a0d38ab"} Feb 27 17:48:36 crc kubenswrapper[4700]: I0227 17:48:36.078260 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-nfs-0" event={"ID":"4c3e4a38-4e77-4f9d-aa8d-355b41080b84","Type":"ContainerStarted","Data":"8a0bbaf7405cf09313dbef7438af20a3dab93390dd327fdce64b90a87f78cb31"} Feb 27 17:48:37 crc kubenswrapper[4700]: I0227 17:48:37.097898 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-nfs-0" event={"ID":"4c3e4a38-4e77-4f9d-aa8d-355b41080b84","Type":"ContainerStarted","Data":"40ceca25a0d2537afb9f57605d4a1e5e575f579ba804ffd9c001f9cd0d002278"} Feb 27 17:48:37 crc kubenswrapper[4700]: I0227 17:48:37.098515 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-nfs-0" event={"ID":"4c3e4a38-4e77-4f9d-aa8d-355b41080b84","Type":"ContainerStarted","Data":"abfc326b24f10cd874ecd578282cf6650f63237e8f75f7690b567d8812cced3d"} Feb 27 17:48:37 crc kubenswrapper[4700]: I0227 17:48:37.104153 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"4891dc6b-c6d7-4468-ab14-e22ccc8462d2","Type":"ContainerStarted","Data":"18f2df0101565ad6449d58b656463a023b9c65315f7b190bb8235401e2336848"} Feb 27 17:48:37 crc kubenswrapper[4700]: I0227 17:48:37.104194 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"4891dc6b-c6d7-4468-ab14-e22ccc8462d2","Type":"ContainerStarted","Data":"5ec4820bc573526eca561bfcbd009e749b3e44f8000dab93c9f6b50d9784f96b"} Feb 27 17:48:37 crc kubenswrapper[4700]: I0227 17:48:37.107579 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-nfs-2-0" event={"ID":"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323","Type":"ContainerStarted","Data":"36d8b3336c05d39a15a5bf19aeb956ccf6dfbf8a4d339901c934c578fd6bdd4d"} Feb 27 17:48:37 crc kubenswrapper[4700]: I0227 17:48:37.107639 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-nfs-2-0" event={"ID":"0e4cd883-f86f-43d5-b2f2-eb96bd9ed323","Type":"ContainerStarted","Data":"a94cb0eab9208f0eea485fd5ce83de43a3d3d2bbae2854f82d3dc74a6c41a558"} Feb 27 17:48:37 crc kubenswrapper[4700]: I0227 17:48:37.123576 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-nfs-0" podStartSLOduration=2.685583013 podStartE2EDuration="3.123544426s" podCreationTimestamp="2026-02-27 17:48:34 +0000 UTC" firstStartedPulling="2026-02-27 17:48:35.432777666 +0000 UTC m=+2875.418090413" lastFinishedPulling="2026-02-27 17:48:35.870739079 +0000 UTC m=+2875.856051826" observedRunningTime="2026-02-27 17:48:37.121589415 +0000 UTC m=+2877.106902162" watchObservedRunningTime="2026-02-27 17:48:37.123544426 +0000 UTC m=+2877.108857213" Feb 27 17:48:37 crc kubenswrapper[4700]: I0227 17:48:37.176254 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-nfs-2-0" podStartSLOduration=3.17622457 podStartE2EDuration="3.17622457s" podCreationTimestamp="2026-02-27 17:48:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:48:37.161481033 +0000 UTC m=+2877.146793800" watchObservedRunningTime="2026-02-27 17:48:37.17622457 +0000 UTC m=+2877.161537317" Feb 27 17:48:37 crc kubenswrapper[4700]: I0227 17:48:37.192363 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=2.777914607 podStartE2EDuration="3.192341663s" podCreationTimestamp="2026-02-27 17:48:34 +0000 UTC" firstStartedPulling="2026-02-27 17:48:35.58637679 +0000 UTC m=+2875.571689537" lastFinishedPulling="2026-02-27 17:48:36.000803846 +0000 UTC m=+2875.986116593" observedRunningTime="2026-02-27 17:48:37.185910824 +0000 UTC m=+2877.171223581" watchObservedRunningTime="2026-02-27 17:48:37.192341663 +0000 UTC m=+2877.177654420" Feb 27 17:48:39 crc kubenswrapper[4700]: E0227 17:48:39.125363 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 17:48:39 crc kubenswrapper[4700]: E0227 17:48:39.126141 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 17:48:39 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 17:48:39 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fzkfq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536908-djpzb_openshift-infra(c22f6346-6080-4170-beff-4fe69f2aa293): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 17:48:39 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 17:48:39 crc kubenswrapper[4700]: E0227 17:48:39.127301 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536908-djpzb" podUID="c22f6346-6080-4170-beff-4fe69f2aa293" Feb 27 17:48:39 crc kubenswrapper[4700]: I0227 17:48:39.136135 4700 generic.go:334] "Generic (PLEG): container finished" podID="dd83e054-9228-4520-ba37-e7b5fbfa05d4" containerID="4bde31f8b98e3d4ff3c8c900be8a29b5d31326d7427085c2a74090c58a0d38ab" exitCode=0 Feb 27 17:48:39 crc kubenswrapper[4700]: I0227 17:48:39.136206 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-269hv" event={"ID":"dd83e054-9228-4520-ba37-e7b5fbfa05d4","Type":"ContainerDied","Data":"4bde31f8b98e3d4ff3c8c900be8a29b5d31326d7427085c2a74090c58a0d38ab"} Feb 27 17:48:39 crc kubenswrapper[4700]: I0227 17:48:39.830256 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Feb 27 17:48:39 crc kubenswrapper[4700]: I0227 17:48:39.948562 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:40 crc kubenswrapper[4700]: I0227 17:48:40.025799 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:40 crc kubenswrapper[4700]: I0227 17:48:40.151305 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-269hv" event={"ID":"dd83e054-9228-4520-ba37-e7b5fbfa05d4","Type":"ContainerStarted","Data":"d2107fc26abec4e14a9f9971835387a9324d0e4a732b91ab7e9629da1946b8a8"} Feb 27 17:48:40 crc kubenswrapper[4700]: I0227 17:48:40.178592 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-269hv" podStartSLOduration=2.604712337 podStartE2EDuration="8.178568021s" podCreationTimestamp="2026-02-27 17:48:32 +0000 UTC" firstStartedPulling="2026-02-27 17:48:34.022251737 +0000 UTC m=+2874.007564524" lastFinishedPulling="2026-02-27 17:48:39.596107461 +0000 UTC m=+2879.581420208" observedRunningTime="2026-02-27 17:48:40.174661198 +0000 UTC m=+2880.159973945" watchObservedRunningTime="2026-02-27 17:48:40.178568021 +0000 UTC m=+2880.163880808" Feb 27 17:48:42 crc kubenswrapper[4700]: I0227 17:48:42.598814 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-269hv" Feb 27 17:48:42 crc kubenswrapper[4700]: I0227 17:48:42.599296 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-269hv" Feb 27 17:48:43 crc kubenswrapper[4700]: I0227 17:48:43.805226 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-269hv" podUID="dd83e054-9228-4520-ba37-e7b5fbfa05d4" containerName="registry-server" probeResult="failure" output=< Feb 27 17:48:43 crc kubenswrapper[4700]: timeout: failed to connect service ":50051" within 1s Feb 27 17:48:43 crc kubenswrapper[4700]: > Feb 27 17:48:45 crc kubenswrapper[4700]: I0227 17:48:45.045151 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Feb 27 17:48:45 crc kubenswrapper[4700]: I0227 17:48:45.135858 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-nfs-0" Feb 27 17:48:45 crc kubenswrapper[4700]: I0227 17:48:45.219129 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-nfs-2-0" Feb 27 17:48:53 crc kubenswrapper[4700]: I0227 17:48:53.701597 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-269hv" podUID="dd83e054-9228-4520-ba37-e7b5fbfa05d4" containerName="registry-server" probeResult="failure" output=< Feb 27 17:48:53 crc kubenswrapper[4700]: timeout: failed to connect service ":50051" within 1s Feb 27 17:48:53 crc kubenswrapper[4700]: > Feb 27 17:48:53 crc kubenswrapper[4700]: E0227 17:48:53.984188 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536908-djpzb" podUID="c22f6346-6080-4170-beff-4fe69f2aa293" Feb 27 17:49:02 crc kubenswrapper[4700]: I0227 17:49:02.693033 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-269hv" Feb 27 17:49:02 crc kubenswrapper[4700]: I0227 17:49:02.779816 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-269hv" Feb 27 17:49:03 crc kubenswrapper[4700]: I0227 17:49:03.463158 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-269hv"] Feb 27 17:49:04 crc kubenswrapper[4700]: I0227 17:49:04.440115 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-269hv" podUID="dd83e054-9228-4520-ba37-e7b5fbfa05d4" containerName="registry-server" containerID="cri-o://d2107fc26abec4e14a9f9971835387a9324d0e4a732b91ab7e9629da1946b8a8" gracePeriod=2 Feb 27 17:49:05 crc kubenswrapper[4700]: I0227 17:49:05.003043 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-269hv" Feb 27 17:49:05 crc kubenswrapper[4700]: I0227 17:49:05.065675 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd83e054-9228-4520-ba37-e7b5fbfa05d4-catalog-content\") pod \"dd83e054-9228-4520-ba37-e7b5fbfa05d4\" (UID: \"dd83e054-9228-4520-ba37-e7b5fbfa05d4\") " Feb 27 17:49:05 crc kubenswrapper[4700]: I0227 17:49:05.065733 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd83e054-9228-4520-ba37-e7b5fbfa05d4-utilities\") pod \"dd83e054-9228-4520-ba37-e7b5fbfa05d4\" (UID: \"dd83e054-9228-4520-ba37-e7b5fbfa05d4\") " Feb 27 17:49:05 crc kubenswrapper[4700]: I0227 17:49:05.065787 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngshj\" (UniqueName: \"kubernetes.io/projected/dd83e054-9228-4520-ba37-e7b5fbfa05d4-kube-api-access-ngshj\") pod \"dd83e054-9228-4520-ba37-e7b5fbfa05d4\" (UID: \"dd83e054-9228-4520-ba37-e7b5fbfa05d4\") " Feb 27 17:49:05 crc kubenswrapper[4700]: I0227 17:49:05.068499 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd83e054-9228-4520-ba37-e7b5fbfa05d4-utilities" (OuterVolumeSpecName: "utilities") pod "dd83e054-9228-4520-ba37-e7b5fbfa05d4" (UID: "dd83e054-9228-4520-ba37-e7b5fbfa05d4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:49:05 crc kubenswrapper[4700]: I0227 17:49:05.074420 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd83e054-9228-4520-ba37-e7b5fbfa05d4-kube-api-access-ngshj" (OuterVolumeSpecName: "kube-api-access-ngshj") pod "dd83e054-9228-4520-ba37-e7b5fbfa05d4" (UID: "dd83e054-9228-4520-ba37-e7b5fbfa05d4"). InnerVolumeSpecName "kube-api-access-ngshj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:49:05 crc kubenswrapper[4700]: I0227 17:49:05.138953 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd83e054-9228-4520-ba37-e7b5fbfa05d4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dd83e054-9228-4520-ba37-e7b5fbfa05d4" (UID: "dd83e054-9228-4520-ba37-e7b5fbfa05d4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:49:05 crc kubenswrapper[4700]: I0227 17:49:05.168970 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd83e054-9228-4520-ba37-e7b5fbfa05d4-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:49:05 crc kubenswrapper[4700]: I0227 17:49:05.169149 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd83e054-9228-4520-ba37-e7b5fbfa05d4-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:49:05 crc kubenswrapper[4700]: I0227 17:49:05.169170 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngshj\" (UniqueName: \"kubernetes.io/projected/dd83e054-9228-4520-ba37-e7b5fbfa05d4-kube-api-access-ngshj\") on node \"crc\" DevicePath \"\"" Feb 27 17:49:05 crc kubenswrapper[4700]: I0227 17:49:05.454309 4700 generic.go:334] "Generic (PLEG): container finished" podID="dd83e054-9228-4520-ba37-e7b5fbfa05d4" containerID="d2107fc26abec4e14a9f9971835387a9324d0e4a732b91ab7e9629da1946b8a8" exitCode=0 Feb 27 17:49:05 crc kubenswrapper[4700]: I0227 17:49:05.454389 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-269hv" event={"ID":"dd83e054-9228-4520-ba37-e7b5fbfa05d4","Type":"ContainerDied","Data":"d2107fc26abec4e14a9f9971835387a9324d0e4a732b91ab7e9629da1946b8a8"} Feb 27 17:49:05 crc kubenswrapper[4700]: I0227 17:49:05.454823 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-269hv" event={"ID":"dd83e054-9228-4520-ba37-e7b5fbfa05d4","Type":"ContainerDied","Data":"2a67f31d5b33bcbeca79ecd5fabc74ee7c61fc00dc27956d47e7c7a6364f1041"} Feb 27 17:49:05 crc kubenswrapper[4700]: I0227 17:49:05.454865 4700 scope.go:117] "RemoveContainer" containerID="d2107fc26abec4e14a9f9971835387a9324d0e4a732b91ab7e9629da1946b8a8" Feb 27 17:49:05 crc kubenswrapper[4700]: I0227 17:49:05.454438 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-269hv" Feb 27 17:49:05 crc kubenswrapper[4700]: I0227 17:49:05.488219 4700 scope.go:117] "RemoveContainer" containerID="4bde31f8b98e3d4ff3c8c900be8a29b5d31326d7427085c2a74090c58a0d38ab" Feb 27 17:49:05 crc kubenswrapper[4700]: I0227 17:49:05.501078 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-269hv"] Feb 27 17:49:05 crc kubenswrapper[4700]: I0227 17:49:05.513090 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-269hv"] Feb 27 17:49:05 crc kubenswrapper[4700]: I0227 17:49:05.535813 4700 scope.go:117] "RemoveContainer" containerID="bba6e6cd87914bd634237083804ee060ecd6bea7eccab47c006e39431e7c7cfe" Feb 27 17:49:05 crc kubenswrapper[4700]: I0227 17:49:05.585356 4700 scope.go:117] "RemoveContainer" containerID="d2107fc26abec4e14a9f9971835387a9324d0e4a732b91ab7e9629da1946b8a8" Feb 27 17:49:05 crc kubenswrapper[4700]: E0227 17:49:05.585869 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2107fc26abec4e14a9f9971835387a9324d0e4a732b91ab7e9629da1946b8a8\": container with ID starting with d2107fc26abec4e14a9f9971835387a9324d0e4a732b91ab7e9629da1946b8a8 not found: ID does not exist" containerID="d2107fc26abec4e14a9f9971835387a9324d0e4a732b91ab7e9629da1946b8a8" Feb 27 17:49:05 crc kubenswrapper[4700]: I0227 17:49:05.585925 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2107fc26abec4e14a9f9971835387a9324d0e4a732b91ab7e9629da1946b8a8"} err="failed to get container status \"d2107fc26abec4e14a9f9971835387a9324d0e4a732b91ab7e9629da1946b8a8\": rpc error: code = NotFound desc = could not find container \"d2107fc26abec4e14a9f9971835387a9324d0e4a732b91ab7e9629da1946b8a8\": container with ID starting with d2107fc26abec4e14a9f9971835387a9324d0e4a732b91ab7e9629da1946b8a8 not found: ID does not exist" Feb 27 17:49:05 crc kubenswrapper[4700]: I0227 17:49:05.586023 4700 scope.go:117] "RemoveContainer" containerID="4bde31f8b98e3d4ff3c8c900be8a29b5d31326d7427085c2a74090c58a0d38ab" Feb 27 17:49:05 crc kubenswrapper[4700]: E0227 17:49:05.586589 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bde31f8b98e3d4ff3c8c900be8a29b5d31326d7427085c2a74090c58a0d38ab\": container with ID starting with 4bde31f8b98e3d4ff3c8c900be8a29b5d31326d7427085c2a74090c58a0d38ab not found: ID does not exist" containerID="4bde31f8b98e3d4ff3c8c900be8a29b5d31326d7427085c2a74090c58a0d38ab" Feb 27 17:49:05 crc kubenswrapper[4700]: I0227 17:49:05.586630 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bde31f8b98e3d4ff3c8c900be8a29b5d31326d7427085c2a74090c58a0d38ab"} err="failed to get container status \"4bde31f8b98e3d4ff3c8c900be8a29b5d31326d7427085c2a74090c58a0d38ab\": rpc error: code = NotFound desc = could not find container \"4bde31f8b98e3d4ff3c8c900be8a29b5d31326d7427085c2a74090c58a0d38ab\": container with ID starting with 4bde31f8b98e3d4ff3c8c900be8a29b5d31326d7427085c2a74090c58a0d38ab not found: ID does not exist" Feb 27 17:49:05 crc kubenswrapper[4700]: I0227 17:49:05.586658 4700 scope.go:117] "RemoveContainer" containerID="bba6e6cd87914bd634237083804ee060ecd6bea7eccab47c006e39431e7c7cfe" Feb 27 17:49:05 crc kubenswrapper[4700]: E0227 17:49:05.586985 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bba6e6cd87914bd634237083804ee060ecd6bea7eccab47c006e39431e7c7cfe\": container with ID starting with bba6e6cd87914bd634237083804ee060ecd6bea7eccab47c006e39431e7c7cfe not found: ID does not exist" containerID="bba6e6cd87914bd634237083804ee060ecd6bea7eccab47c006e39431e7c7cfe" Feb 27 17:49:05 crc kubenswrapper[4700]: I0227 17:49:05.587034 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bba6e6cd87914bd634237083804ee060ecd6bea7eccab47c006e39431e7c7cfe"} err="failed to get container status \"bba6e6cd87914bd634237083804ee060ecd6bea7eccab47c006e39431e7c7cfe\": rpc error: code = NotFound desc = could not find container \"bba6e6cd87914bd634237083804ee060ecd6bea7eccab47c006e39431e7c7cfe\": container with ID starting with bba6e6cd87914bd634237083804ee060ecd6bea7eccab47c006e39431e7c7cfe not found: ID does not exist" Feb 27 17:49:06 crc kubenswrapper[4700]: E0227 17:49:05.993255 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536908-djpzb" podUID="c22f6346-6080-4170-beff-4fe69f2aa293" Feb 27 17:49:07 crc kubenswrapper[4700]: I0227 17:49:07.000049 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd83e054-9228-4520-ba37-e7b5fbfa05d4" path="/var/lib/kubelet/pods/dd83e054-9228-4520-ba37-e7b5fbfa05d4/volumes" Feb 27 17:49:17 crc kubenswrapper[4700]: E0227 17:49:17.984295 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536908-djpzb" podUID="c22f6346-6080-4170-beff-4fe69f2aa293" Feb 27 17:49:34 crc kubenswrapper[4700]: I0227 17:49:34.805826 4700 generic.go:334] "Generic (PLEG): container finished" podID="c22f6346-6080-4170-beff-4fe69f2aa293" containerID="4706c4faccd28b192e57611ee5997cafeb72162adc3fcbcce8a6c25a87b4f8a9" exitCode=0 Feb 27 17:49:34 crc kubenswrapper[4700]: I0227 17:49:34.805933 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536908-djpzb" event={"ID":"c22f6346-6080-4170-beff-4fe69f2aa293","Type":"ContainerDied","Data":"4706c4faccd28b192e57611ee5997cafeb72162adc3fcbcce8a6c25a87b4f8a9"} Feb 27 17:49:36 crc kubenswrapper[4700]: I0227 17:49:36.293339 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536908-djpzb" Feb 27 17:49:36 crc kubenswrapper[4700]: I0227 17:49:36.377700 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzkfq\" (UniqueName: \"kubernetes.io/projected/c22f6346-6080-4170-beff-4fe69f2aa293-kube-api-access-fzkfq\") pod \"c22f6346-6080-4170-beff-4fe69f2aa293\" (UID: \"c22f6346-6080-4170-beff-4fe69f2aa293\") " Feb 27 17:49:36 crc kubenswrapper[4700]: I0227 17:49:36.385765 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c22f6346-6080-4170-beff-4fe69f2aa293-kube-api-access-fzkfq" (OuterVolumeSpecName: "kube-api-access-fzkfq") pod "c22f6346-6080-4170-beff-4fe69f2aa293" (UID: "c22f6346-6080-4170-beff-4fe69f2aa293"). InnerVolumeSpecName "kube-api-access-fzkfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:49:36 crc kubenswrapper[4700]: I0227 17:49:36.480626 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzkfq\" (UniqueName: \"kubernetes.io/projected/c22f6346-6080-4170-beff-4fe69f2aa293-kube-api-access-fzkfq\") on node \"crc\" DevicePath \"\"" Feb 27 17:49:36 crc kubenswrapper[4700]: I0227 17:49:36.836253 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536908-djpzb" event={"ID":"c22f6346-6080-4170-beff-4fe69f2aa293","Type":"ContainerDied","Data":"61a133c9fa354d556645b703b0da1c0bf30682eda133675eba7ebba56d771909"} Feb 27 17:49:36 crc kubenswrapper[4700]: I0227 17:49:36.836701 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61a133c9fa354d556645b703b0da1c0bf30682eda133675eba7ebba56d771909" Feb 27 17:49:36 crc kubenswrapper[4700]: I0227 17:49:36.836338 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536908-djpzb" Feb 27 17:49:37 crc kubenswrapper[4700]: I0227 17:49:37.403487 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536902-vfh4z"] Feb 27 17:49:37 crc kubenswrapper[4700]: I0227 17:49:37.413419 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536902-vfh4z"] Feb 27 17:49:38 crc kubenswrapper[4700]: I0227 17:49:38.994205 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82ea0ed5-7fde-447d-8fb4-780ad1fa6215" path="/var/lib/kubelet/pods/82ea0ed5-7fde-447d-8fb4-780ad1fa6215/volumes" Feb 27 17:49:44 crc kubenswrapper[4700]: I0227 17:49:44.821275 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Feb 27 17:49:44 crc kubenswrapper[4700]: I0227 17:49:44.821933 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="dc9ec3fe-326d-4df2-b561-318f574cfdf3" containerName="prometheus" containerID="cri-o://a3ba82809ba9f29c586671d969f56e51963e184bd5c34352d55e7d387d108965" gracePeriod=600 Feb 27 17:49:44 crc kubenswrapper[4700]: I0227 17:49:44.822049 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="dc9ec3fe-326d-4df2-b561-318f574cfdf3" containerName="config-reloader" containerID="cri-o://97260215fdb80740d553458dd4cf1caafb5e06781782252f81abe8bd85e205b1" gracePeriod=600 Feb 27 17:49:44 crc kubenswrapper[4700]: I0227 17:49:44.822057 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="dc9ec3fe-326d-4df2-b561-318f574cfdf3" containerName="thanos-sidecar" containerID="cri-o://f2821b24a20a1fda4814c4839d720bb09a0da9f533505a5024cf4d4cb922684c" gracePeriod=600 Feb 27 17:49:45 crc kubenswrapper[4700]: I0227 17:49:45.357619 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="dc9ec3fe-326d-4df2-b561-318f574cfdf3" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.140:9090/-/ready\": dial tcp 10.217.0.140:9090: connect: connection refused" Feb 27 17:49:45 crc kubenswrapper[4700]: I0227 17:49:45.880904 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:45 crc kubenswrapper[4700]: I0227 17:49:45.979588 4700 generic.go:334] "Generic (PLEG): container finished" podID="dc9ec3fe-326d-4df2-b561-318f574cfdf3" containerID="f2821b24a20a1fda4814c4839d720bb09a0da9f533505a5024cf4d4cb922684c" exitCode=0 Feb 27 17:49:45 crc kubenswrapper[4700]: I0227 17:49:45.979627 4700 generic.go:334] "Generic (PLEG): container finished" podID="dc9ec3fe-326d-4df2-b561-318f574cfdf3" containerID="97260215fdb80740d553458dd4cf1caafb5e06781782252f81abe8bd85e205b1" exitCode=0 Feb 27 17:49:45 crc kubenswrapper[4700]: I0227 17:49:45.979639 4700 generic.go:334] "Generic (PLEG): container finished" podID="dc9ec3fe-326d-4df2-b561-318f574cfdf3" containerID="a3ba82809ba9f29c586671d969f56e51963e184bd5c34352d55e7d387d108965" exitCode=0 Feb 27 17:49:45 crc kubenswrapper[4700]: I0227 17:49:45.979664 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"dc9ec3fe-326d-4df2-b561-318f574cfdf3","Type":"ContainerDied","Data":"f2821b24a20a1fda4814c4839d720bb09a0da9f533505a5024cf4d4cb922684c"} Feb 27 17:49:45 crc kubenswrapper[4700]: I0227 17:49:45.979693 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"dc9ec3fe-326d-4df2-b561-318f574cfdf3","Type":"ContainerDied","Data":"97260215fdb80740d553458dd4cf1caafb5e06781782252f81abe8bd85e205b1"} Feb 27 17:49:45 crc kubenswrapper[4700]: I0227 17:49:45.979707 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"dc9ec3fe-326d-4df2-b561-318f574cfdf3","Type":"ContainerDied","Data":"a3ba82809ba9f29c586671d969f56e51963e184bd5c34352d55e7d387d108965"} Feb 27 17:49:45 crc kubenswrapper[4700]: I0227 17:49:45.979720 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"dc9ec3fe-326d-4df2-b561-318f574cfdf3","Type":"ContainerDied","Data":"f04bb2e9aebd382befff678525f9d9408ceb5ed38fb1d6f61cd1910981782d4f"} Feb 27 17:49:45 crc kubenswrapper[4700]: I0227 17:49:45.979740 4700 scope.go:117] "RemoveContainer" containerID="f2821b24a20a1fda4814c4839d720bb09a0da9f533505a5024cf4d4cb922684c" Feb 27 17:49:45 crc kubenswrapper[4700]: I0227 17:49:45.979893 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.018129 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/dc9ec3fe-326d-4df2-b561-318f574cfdf3-tls-assets\") pod \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.018193 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/dc9ec3fe-326d-4df2-b561-318f574cfdf3-config\") pod \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.018265 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/dc9ec3fe-326d-4df2-b561-318f574cfdf3-thanos-prometheus-http-client-file\") pod \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.018313 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/dc9ec3fe-326d-4df2-b561-318f574cfdf3-prometheus-metric-storage-rulefiles-1\") pod \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.018332 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc9ec3fe-326d-4df2-b561-318f574cfdf3-secret-combined-ca-bundle\") pod \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.018350 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/dc9ec3fe-326d-4df2-b561-318f574cfdf3-web-config\") pod \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.018419 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/dc9ec3fe-326d-4df2-b561-318f574cfdf3-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.019722 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bw6sb\" (UniqueName: \"kubernetes.io/projected/dc9ec3fe-326d-4df2-b561-318f574cfdf3-kube-api-access-bw6sb\") pod \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.019797 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/dc9ec3fe-326d-4df2-b561-318f574cfdf3-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.019923 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2\") pod \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.019991 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/dc9ec3fe-326d-4df2-b561-318f574cfdf3-config-out\") pod \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.020015 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/dc9ec3fe-326d-4df2-b561-318f574cfdf3-prometheus-metric-storage-rulefiles-0\") pod \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.020039 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/dc9ec3fe-326d-4df2-b561-318f574cfdf3-prometheus-metric-storage-rulefiles-2\") pod \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\" (UID: \"dc9ec3fe-326d-4df2-b561-318f574cfdf3\") " Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.021042 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc9ec3fe-326d-4df2-b561-318f574cfdf3-prometheus-metric-storage-rulefiles-2" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-2") pod "dc9ec3fe-326d-4df2-b561-318f574cfdf3" (UID: "dc9ec3fe-326d-4df2-b561-318f574cfdf3"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-2". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.025123 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc9ec3fe-326d-4df2-b561-318f574cfdf3-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "dc9ec3fe-326d-4df2-b561-318f574cfdf3" (UID: "dc9ec3fe-326d-4df2-b561-318f574cfdf3"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.025607 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc9ec3fe-326d-4df2-b561-318f574cfdf3-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d") pod "dc9ec3fe-326d-4df2-b561-318f574cfdf3" (UID: "dc9ec3fe-326d-4df2-b561-318f574cfdf3"). InnerVolumeSpecName "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.026797 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc9ec3fe-326d-4df2-b561-318f574cfdf3-config" (OuterVolumeSpecName: "config") pod "dc9ec3fe-326d-4df2-b561-318f574cfdf3" (UID: "dc9ec3fe-326d-4df2-b561-318f574cfdf3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.026823 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc9ec3fe-326d-4df2-b561-318f574cfdf3-prometheus-metric-storage-rulefiles-1" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-1") pod "dc9ec3fe-326d-4df2-b561-318f574cfdf3" (UID: "dc9ec3fe-326d-4df2-b561-318f574cfdf3"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.033751 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc9ec3fe-326d-4df2-b561-318f574cfdf3-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "dc9ec3fe-326d-4df2-b561-318f574cfdf3" (UID: "dc9ec3fe-326d-4df2-b561-318f574cfdf3"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.034640 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc9ec3fe-326d-4df2-b561-318f574cfdf3-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "dc9ec3fe-326d-4df2-b561-318f574cfdf3" (UID: "dc9ec3fe-326d-4df2-b561-318f574cfdf3"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.037519 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc9ec3fe-326d-4df2-b561-318f574cfdf3-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d") pod "dc9ec3fe-326d-4df2-b561-318f574cfdf3" (UID: "dc9ec3fe-326d-4df2-b561-318f574cfdf3"). InnerVolumeSpecName "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.038137 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc9ec3fe-326d-4df2-b561-318f574cfdf3-config-out" (OuterVolumeSpecName: "config-out") pod "dc9ec3fe-326d-4df2-b561-318f574cfdf3" (UID: "dc9ec3fe-326d-4df2-b561-318f574cfdf3"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.042537 4700 scope.go:117] "RemoveContainer" containerID="97260215fdb80740d553458dd4cf1caafb5e06781782252f81abe8bd85e205b1" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.056550 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc9ec3fe-326d-4df2-b561-318f574cfdf3-kube-api-access-bw6sb" (OuterVolumeSpecName: "kube-api-access-bw6sb") pod "dc9ec3fe-326d-4df2-b561-318f574cfdf3" (UID: "dc9ec3fe-326d-4df2-b561-318f574cfdf3"). InnerVolumeSpecName "kube-api-access-bw6sb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.063636 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc9ec3fe-326d-4df2-b561-318f574cfdf3-secret-combined-ca-bundle" (OuterVolumeSpecName: "secret-combined-ca-bundle") pod "dc9ec3fe-326d-4df2-b561-318f574cfdf3" (UID: "dc9ec3fe-326d-4df2-b561-318f574cfdf3"). InnerVolumeSpecName "secret-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.100244 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "dc9ec3fe-326d-4df2-b561-318f574cfdf3" (UID: "dc9ec3fe-326d-4df2-b561-318f574cfdf3"). InnerVolumeSpecName "pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.122441 4700 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/dc9ec3fe-326d-4df2-b561-318f574cfdf3-prometheus-metric-storage-rulefiles-1\") on node \"crc\" DevicePath \"\"" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.122489 4700 reconciler_common.go:293] "Volume detached for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc9ec3fe-326d-4df2-b561-318f574cfdf3-secret-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.122502 4700 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/dc9ec3fe-326d-4df2-b561-318f574cfdf3-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") on node \"crc\" DevicePath \"\"" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.122514 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bw6sb\" (UniqueName: \"kubernetes.io/projected/dc9ec3fe-326d-4df2-b561-318f574cfdf3-kube-api-access-bw6sb\") on node \"crc\" DevicePath \"\"" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.122523 4700 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/dc9ec3fe-326d-4df2-b561-318f574cfdf3-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") on node \"crc\" DevicePath \"\"" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.122542 4700 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2\") on node \"crc\" " Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.122554 4700 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/dc9ec3fe-326d-4df2-b561-318f574cfdf3-config-out\") on node \"crc\" DevicePath \"\"" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.122563 4700 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/dc9ec3fe-326d-4df2-b561-318f574cfdf3-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.122572 4700 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/dc9ec3fe-326d-4df2-b561-318f574cfdf3-prometheus-metric-storage-rulefiles-2\") on node \"crc\" DevicePath \"\"" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.122580 4700 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/dc9ec3fe-326d-4df2-b561-318f574cfdf3-tls-assets\") on node \"crc\" DevicePath \"\"" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.122588 4700 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/dc9ec3fe-326d-4df2-b561-318f574cfdf3-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.122597 4700 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/dc9ec3fe-326d-4df2-b561-318f574cfdf3-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.146603 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc9ec3fe-326d-4df2-b561-318f574cfdf3-web-config" (OuterVolumeSpecName: "web-config") pod "dc9ec3fe-326d-4df2-b561-318f574cfdf3" (UID: "dc9ec3fe-326d-4df2-b561-318f574cfdf3"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.211727 4700 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.212355 4700 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2") on node "crc" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.219650 4700 scope.go:117] "RemoveContainer" containerID="a3ba82809ba9f29c586671d969f56e51963e184bd5c34352d55e7d387d108965" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.224283 4700 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/dc9ec3fe-326d-4df2-b561-318f574cfdf3-web-config\") on node \"crc\" DevicePath \"\"" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.224321 4700 reconciler_common.go:293] "Volume detached for volume \"pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2\") on node \"crc\" DevicePath \"\"" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.240736 4700 scope.go:117] "RemoveContainer" containerID="2628eb269c1943b0e5e9f37bb8c617751b464272b789e5fc05c1a2050a641c11" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.317226 4700 scope.go:117] "RemoveContainer" containerID="f2821b24a20a1fda4814c4839d720bb09a0da9f533505a5024cf4d4cb922684c" Feb 27 17:49:46 crc kubenswrapper[4700]: E0227 17:49:46.317719 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2821b24a20a1fda4814c4839d720bb09a0da9f533505a5024cf4d4cb922684c\": container with ID starting with f2821b24a20a1fda4814c4839d720bb09a0da9f533505a5024cf4d4cb922684c not found: ID does not exist" containerID="f2821b24a20a1fda4814c4839d720bb09a0da9f533505a5024cf4d4cb922684c" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.317757 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2821b24a20a1fda4814c4839d720bb09a0da9f533505a5024cf4d4cb922684c"} err="failed to get container status \"f2821b24a20a1fda4814c4839d720bb09a0da9f533505a5024cf4d4cb922684c\": rpc error: code = NotFound desc = could not find container \"f2821b24a20a1fda4814c4839d720bb09a0da9f533505a5024cf4d4cb922684c\": container with ID starting with f2821b24a20a1fda4814c4839d720bb09a0da9f533505a5024cf4d4cb922684c not found: ID does not exist" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.317781 4700 scope.go:117] "RemoveContainer" containerID="97260215fdb80740d553458dd4cf1caafb5e06781782252f81abe8bd85e205b1" Feb 27 17:49:46 crc kubenswrapper[4700]: E0227 17:49:46.318075 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97260215fdb80740d553458dd4cf1caafb5e06781782252f81abe8bd85e205b1\": container with ID starting with 97260215fdb80740d553458dd4cf1caafb5e06781782252f81abe8bd85e205b1 not found: ID does not exist" containerID="97260215fdb80740d553458dd4cf1caafb5e06781782252f81abe8bd85e205b1" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.318111 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97260215fdb80740d553458dd4cf1caafb5e06781782252f81abe8bd85e205b1"} err="failed to get container status \"97260215fdb80740d553458dd4cf1caafb5e06781782252f81abe8bd85e205b1\": rpc error: code = NotFound desc = could not find container \"97260215fdb80740d553458dd4cf1caafb5e06781782252f81abe8bd85e205b1\": container with ID starting with 97260215fdb80740d553458dd4cf1caafb5e06781782252f81abe8bd85e205b1 not found: ID does not exist" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.318129 4700 scope.go:117] "RemoveContainer" containerID="a3ba82809ba9f29c586671d969f56e51963e184bd5c34352d55e7d387d108965" Feb 27 17:49:46 crc kubenswrapper[4700]: E0227 17:49:46.318520 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3ba82809ba9f29c586671d969f56e51963e184bd5c34352d55e7d387d108965\": container with ID starting with a3ba82809ba9f29c586671d969f56e51963e184bd5c34352d55e7d387d108965 not found: ID does not exist" containerID="a3ba82809ba9f29c586671d969f56e51963e184bd5c34352d55e7d387d108965" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.318566 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3ba82809ba9f29c586671d969f56e51963e184bd5c34352d55e7d387d108965"} err="failed to get container status \"a3ba82809ba9f29c586671d969f56e51963e184bd5c34352d55e7d387d108965\": rpc error: code = NotFound desc = could not find container \"a3ba82809ba9f29c586671d969f56e51963e184bd5c34352d55e7d387d108965\": container with ID starting with a3ba82809ba9f29c586671d969f56e51963e184bd5c34352d55e7d387d108965 not found: ID does not exist" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.318595 4700 scope.go:117] "RemoveContainer" containerID="2628eb269c1943b0e5e9f37bb8c617751b464272b789e5fc05c1a2050a641c11" Feb 27 17:49:46 crc kubenswrapper[4700]: E0227 17:49:46.319004 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2628eb269c1943b0e5e9f37bb8c617751b464272b789e5fc05c1a2050a641c11\": container with ID starting with 2628eb269c1943b0e5e9f37bb8c617751b464272b789e5fc05c1a2050a641c11 not found: ID does not exist" containerID="2628eb269c1943b0e5e9f37bb8c617751b464272b789e5fc05c1a2050a641c11" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.319035 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2628eb269c1943b0e5e9f37bb8c617751b464272b789e5fc05c1a2050a641c11"} err="failed to get container status \"2628eb269c1943b0e5e9f37bb8c617751b464272b789e5fc05c1a2050a641c11\": rpc error: code = NotFound desc = could not find container \"2628eb269c1943b0e5e9f37bb8c617751b464272b789e5fc05c1a2050a641c11\": container with ID starting with 2628eb269c1943b0e5e9f37bb8c617751b464272b789e5fc05c1a2050a641c11 not found: ID does not exist" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.319051 4700 scope.go:117] "RemoveContainer" containerID="f2821b24a20a1fda4814c4839d720bb09a0da9f533505a5024cf4d4cb922684c" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.319747 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2821b24a20a1fda4814c4839d720bb09a0da9f533505a5024cf4d4cb922684c"} err="failed to get container status \"f2821b24a20a1fda4814c4839d720bb09a0da9f533505a5024cf4d4cb922684c\": rpc error: code = NotFound desc = could not find container \"f2821b24a20a1fda4814c4839d720bb09a0da9f533505a5024cf4d4cb922684c\": container with ID starting with f2821b24a20a1fda4814c4839d720bb09a0da9f533505a5024cf4d4cb922684c not found: ID does not exist" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.319770 4700 scope.go:117] "RemoveContainer" containerID="97260215fdb80740d553458dd4cf1caafb5e06781782252f81abe8bd85e205b1" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.327255 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97260215fdb80740d553458dd4cf1caafb5e06781782252f81abe8bd85e205b1"} err="failed to get container status \"97260215fdb80740d553458dd4cf1caafb5e06781782252f81abe8bd85e205b1\": rpc error: code = NotFound desc = could not find container \"97260215fdb80740d553458dd4cf1caafb5e06781782252f81abe8bd85e205b1\": container with ID starting with 97260215fdb80740d553458dd4cf1caafb5e06781782252f81abe8bd85e205b1 not found: ID does not exist" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.327293 4700 scope.go:117] "RemoveContainer" containerID="a3ba82809ba9f29c586671d969f56e51963e184bd5c34352d55e7d387d108965" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.327387 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.328012 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3ba82809ba9f29c586671d969f56e51963e184bd5c34352d55e7d387d108965"} err="failed to get container status \"a3ba82809ba9f29c586671d969f56e51963e184bd5c34352d55e7d387d108965\": rpc error: code = NotFound desc = could not find container \"a3ba82809ba9f29c586671d969f56e51963e184bd5c34352d55e7d387d108965\": container with ID starting with a3ba82809ba9f29c586671d969f56e51963e184bd5c34352d55e7d387d108965 not found: ID does not exist" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.328045 4700 scope.go:117] "RemoveContainer" containerID="2628eb269c1943b0e5e9f37bb8c617751b464272b789e5fc05c1a2050a641c11" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.328421 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2628eb269c1943b0e5e9f37bb8c617751b464272b789e5fc05c1a2050a641c11"} err="failed to get container status \"2628eb269c1943b0e5e9f37bb8c617751b464272b789e5fc05c1a2050a641c11\": rpc error: code = NotFound desc = could not find container \"2628eb269c1943b0e5e9f37bb8c617751b464272b789e5fc05c1a2050a641c11\": container with ID starting with 2628eb269c1943b0e5e9f37bb8c617751b464272b789e5fc05c1a2050a641c11 not found: ID does not exist" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.328447 4700 scope.go:117] "RemoveContainer" containerID="f2821b24a20a1fda4814c4839d720bb09a0da9f533505a5024cf4d4cb922684c" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.328840 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2821b24a20a1fda4814c4839d720bb09a0da9f533505a5024cf4d4cb922684c"} err="failed to get container status \"f2821b24a20a1fda4814c4839d720bb09a0da9f533505a5024cf4d4cb922684c\": rpc error: code = NotFound desc = could not find container \"f2821b24a20a1fda4814c4839d720bb09a0da9f533505a5024cf4d4cb922684c\": container with ID starting with f2821b24a20a1fda4814c4839d720bb09a0da9f533505a5024cf4d4cb922684c not found: ID does not exist" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.328955 4700 scope.go:117] "RemoveContainer" containerID="97260215fdb80740d553458dd4cf1caafb5e06781782252f81abe8bd85e205b1" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.329367 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97260215fdb80740d553458dd4cf1caafb5e06781782252f81abe8bd85e205b1"} err="failed to get container status \"97260215fdb80740d553458dd4cf1caafb5e06781782252f81abe8bd85e205b1\": rpc error: code = NotFound desc = could not find container \"97260215fdb80740d553458dd4cf1caafb5e06781782252f81abe8bd85e205b1\": container with ID starting with 97260215fdb80740d553458dd4cf1caafb5e06781782252f81abe8bd85e205b1 not found: ID does not exist" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.329389 4700 scope.go:117] "RemoveContainer" containerID="a3ba82809ba9f29c586671d969f56e51963e184bd5c34352d55e7d387d108965" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.329620 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3ba82809ba9f29c586671d969f56e51963e184bd5c34352d55e7d387d108965"} err="failed to get container status \"a3ba82809ba9f29c586671d969f56e51963e184bd5c34352d55e7d387d108965\": rpc error: code = NotFound desc = could not find container \"a3ba82809ba9f29c586671d969f56e51963e184bd5c34352d55e7d387d108965\": container with ID starting with a3ba82809ba9f29c586671d969f56e51963e184bd5c34352d55e7d387d108965 not found: ID does not exist" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.329643 4700 scope.go:117] "RemoveContainer" containerID="2628eb269c1943b0e5e9f37bb8c617751b464272b789e5fc05c1a2050a641c11" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.329993 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2628eb269c1943b0e5e9f37bb8c617751b464272b789e5fc05c1a2050a641c11"} err="failed to get container status \"2628eb269c1943b0e5e9f37bb8c617751b464272b789e5fc05c1a2050a641c11\": rpc error: code = NotFound desc = could not find container \"2628eb269c1943b0e5e9f37bb8c617751b464272b789e5fc05c1a2050a641c11\": container with ID starting with 2628eb269c1943b0e5e9f37bb8c617751b464272b789e5fc05c1a2050a641c11 not found: ID does not exist" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.336099 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.360300 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Feb 27 17:49:46 crc kubenswrapper[4700]: E0227 17:49:46.360725 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd83e054-9228-4520-ba37-e7b5fbfa05d4" containerName="extract-utilities" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.360741 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd83e054-9228-4520-ba37-e7b5fbfa05d4" containerName="extract-utilities" Feb 27 17:49:46 crc kubenswrapper[4700]: E0227 17:49:46.360772 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc9ec3fe-326d-4df2-b561-318f574cfdf3" containerName="thanos-sidecar" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.360778 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc9ec3fe-326d-4df2-b561-318f574cfdf3" containerName="thanos-sidecar" Feb 27 17:49:46 crc kubenswrapper[4700]: E0227 17:49:46.360789 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd83e054-9228-4520-ba37-e7b5fbfa05d4" containerName="extract-content" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.360795 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd83e054-9228-4520-ba37-e7b5fbfa05d4" containerName="extract-content" Feb 27 17:49:46 crc kubenswrapper[4700]: E0227 17:49:46.360810 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c22f6346-6080-4170-beff-4fe69f2aa293" containerName="oc" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.360815 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="c22f6346-6080-4170-beff-4fe69f2aa293" containerName="oc" Feb 27 17:49:46 crc kubenswrapper[4700]: E0227 17:49:46.360826 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc9ec3fe-326d-4df2-b561-318f574cfdf3" containerName="prometheus" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.360833 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc9ec3fe-326d-4df2-b561-318f574cfdf3" containerName="prometheus" Feb 27 17:49:46 crc kubenswrapper[4700]: E0227 17:49:46.360847 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc9ec3fe-326d-4df2-b561-318f574cfdf3" containerName="init-config-reloader" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.360853 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc9ec3fe-326d-4df2-b561-318f574cfdf3" containerName="init-config-reloader" Feb 27 17:49:46 crc kubenswrapper[4700]: E0227 17:49:46.360866 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd83e054-9228-4520-ba37-e7b5fbfa05d4" containerName="registry-server" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.360873 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd83e054-9228-4520-ba37-e7b5fbfa05d4" containerName="registry-server" Feb 27 17:49:46 crc kubenswrapper[4700]: E0227 17:49:46.360883 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc9ec3fe-326d-4df2-b561-318f574cfdf3" containerName="config-reloader" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.360889 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc9ec3fe-326d-4df2-b561-318f574cfdf3" containerName="config-reloader" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.361059 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="c22f6346-6080-4170-beff-4fe69f2aa293" containerName="oc" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.361078 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc9ec3fe-326d-4df2-b561-318f574cfdf3" containerName="prometheus" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.361089 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc9ec3fe-326d-4df2-b561-318f574cfdf3" containerName="thanos-sidecar" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.361099 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc9ec3fe-326d-4df2-b561-318f574cfdf3" containerName="config-reloader" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.361107 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd83e054-9228-4520-ba37-e7b5fbfa05d4" containerName="registry-server" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.362804 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.367485 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-jdvhl" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.368051 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-1" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.368587 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.368944 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.368961 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.369191 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-2" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.369522 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.376852 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.377023 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.533496 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkgmq\" (UniqueName: \"kubernetes.io/projected/306434a9-5de0-42d0-8c10-184238aa5f97-kube-api-access-bkgmq\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.533538 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/306434a9-5de0-42d0-8c10-184238aa5f97-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.533564 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/306434a9-5de0-42d0-8c10-184238aa5f97-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.533683 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/306434a9-5de0-42d0-8c10-184238aa5f97-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.533747 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.533795 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/306434a9-5de0-42d0-8c10-184238aa5f97-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.533923 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/306434a9-5de0-42d0-8c10-184238aa5f97-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.533984 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/306434a9-5de0-42d0-8c10-184238aa5f97-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.534021 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/306434a9-5de0-42d0-8c10-184238aa5f97-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.534115 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/306434a9-5de0-42d0-8c10-184238aa5f97-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.534170 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/306434a9-5de0-42d0-8c10-184238aa5f97-config\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.534344 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/306434a9-5de0-42d0-8c10-184238aa5f97-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.534596 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/306434a9-5de0-42d0-8c10-184238aa5f97-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.636811 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/306434a9-5de0-42d0-8c10-184238aa5f97-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.636901 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.636952 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/306434a9-5de0-42d0-8c10-184238aa5f97-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.637088 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/306434a9-5de0-42d0-8c10-184238aa5f97-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.637209 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/306434a9-5de0-42d0-8c10-184238aa5f97-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.637923 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/306434a9-5de0-42d0-8c10-184238aa5f97-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.637982 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/306434a9-5de0-42d0-8c10-184238aa5f97-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.638009 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/306434a9-5de0-42d0-8c10-184238aa5f97-config\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.638118 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/306434a9-5de0-42d0-8c10-184238aa5f97-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.638260 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/306434a9-5de0-42d0-8c10-184238aa5f97-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.638330 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkgmq\" (UniqueName: \"kubernetes.io/projected/306434a9-5de0-42d0-8c10-184238aa5f97-kube-api-access-bkgmq\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.638353 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/306434a9-5de0-42d0-8c10-184238aa5f97-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.638377 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/306434a9-5de0-42d0-8c10-184238aa5f97-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.639007 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/306434a9-5de0-42d0-8c10-184238aa5f97-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.640823 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/306434a9-5de0-42d0-8c10-184238aa5f97-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.640971 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/306434a9-5de0-42d0-8c10-184238aa5f97-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.644006 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/306434a9-5de0-42d0-8c10-184238aa5f97-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.644010 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/306434a9-5de0-42d0-8c10-184238aa5f97-config\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.644049 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/306434a9-5de0-42d0-8c10-184238aa5f97-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.644714 4700 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.644738 4700 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/51157a65261d84782fc18d298283232db1d99eb5d30079bdb31ebdc3e04ec75d/globalmount\"" pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.645630 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/306434a9-5de0-42d0-8c10-184238aa5f97-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.645771 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/306434a9-5de0-42d0-8c10-184238aa5f97-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.646099 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/306434a9-5de0-42d0-8c10-184238aa5f97-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.647732 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/306434a9-5de0-42d0-8c10-184238aa5f97-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.648218 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/306434a9-5de0-42d0-8c10-184238aa5f97-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.674098 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkgmq\" (UniqueName: \"kubernetes.io/projected/306434a9-5de0-42d0-8c10-184238aa5f97-kube-api-access-bkgmq\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.704306 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2372ca89-4a20-4a69-be03-e0067f7bd2c2\") pod \"prometheus-metric-storage-0\" (UID: \"306434a9-5de0-42d0-8c10-184238aa5f97\") " pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.988168 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Feb 27 17:49:46 crc kubenswrapper[4700]: I0227 17:49:46.998598 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc9ec3fe-326d-4df2-b561-318f574cfdf3" path="/var/lib/kubelet/pods/dc9ec3fe-326d-4df2-b561-318f574cfdf3/volumes" Feb 27 17:49:47 crc kubenswrapper[4700]: I0227 17:49:47.533863 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Feb 27 17:49:48 crc kubenswrapper[4700]: I0227 17:49:48.007807 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"306434a9-5de0-42d0-8c10-184238aa5f97","Type":"ContainerStarted","Data":"f7310c1c6da1dcb1a912425e824d238b4dbc2822f64653c60c9b49f00a470e4e"} Feb 27 17:49:52 crc kubenswrapper[4700]: I0227 17:49:52.058950 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"306434a9-5de0-42d0-8c10-184238aa5f97","Type":"ContainerStarted","Data":"89a1443c95e5ed6671b71f63a4e8182c44f429450ae3fabc0749cbd93886564e"} Feb 27 17:50:00 crc kubenswrapper[4700]: I0227 17:50:00.145401 4700 generic.go:334] "Generic (PLEG): container finished" podID="306434a9-5de0-42d0-8c10-184238aa5f97" containerID="89a1443c95e5ed6671b71f63a4e8182c44f429450ae3fabc0749cbd93886564e" exitCode=0 Feb 27 17:50:00 crc kubenswrapper[4700]: I0227 17:50:00.145652 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"306434a9-5de0-42d0-8c10-184238aa5f97","Type":"ContainerDied","Data":"89a1443c95e5ed6671b71f63a4e8182c44f429450ae3fabc0749cbd93886564e"} Feb 27 17:50:00 crc kubenswrapper[4700]: I0227 17:50:00.172939 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536910-x8r7g"] Feb 27 17:50:00 crc kubenswrapper[4700]: I0227 17:50:00.174594 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536910-x8r7g" Feb 27 17:50:00 crc kubenswrapper[4700]: I0227 17:50:00.177497 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d62kc" Feb 27 17:50:00 crc kubenswrapper[4700]: I0227 17:50:00.178747 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:50:00 crc kubenswrapper[4700]: I0227 17:50:00.183312 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:50:00 crc kubenswrapper[4700]: I0227 17:50:00.199211 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536910-x8r7g"] Feb 27 17:50:00 crc kubenswrapper[4700]: I0227 17:50:00.260773 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfxrq\" (UniqueName: \"kubernetes.io/projected/8b9e50e8-ca40-48ba-93b3-66e6e0d6d8e0-kube-api-access-jfxrq\") pod \"auto-csr-approver-29536910-x8r7g\" (UID: \"8b9e50e8-ca40-48ba-93b3-66e6e0d6d8e0\") " pod="openshift-infra/auto-csr-approver-29536910-x8r7g" Feb 27 17:50:00 crc kubenswrapper[4700]: I0227 17:50:00.362735 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfxrq\" (UniqueName: \"kubernetes.io/projected/8b9e50e8-ca40-48ba-93b3-66e6e0d6d8e0-kube-api-access-jfxrq\") pod \"auto-csr-approver-29536910-x8r7g\" (UID: \"8b9e50e8-ca40-48ba-93b3-66e6e0d6d8e0\") " pod="openshift-infra/auto-csr-approver-29536910-x8r7g" Feb 27 17:50:00 crc kubenswrapper[4700]: I0227 17:50:00.379208 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfxrq\" (UniqueName: \"kubernetes.io/projected/8b9e50e8-ca40-48ba-93b3-66e6e0d6d8e0-kube-api-access-jfxrq\") pod \"auto-csr-approver-29536910-x8r7g\" (UID: \"8b9e50e8-ca40-48ba-93b3-66e6e0d6d8e0\") " pod="openshift-infra/auto-csr-approver-29536910-x8r7g" Feb 27 17:50:00 crc kubenswrapper[4700]: I0227 17:50:00.610055 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536910-x8r7g" Feb 27 17:50:01 crc kubenswrapper[4700]: W0227 17:50:01.117069 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8b9e50e8_ca40_48ba_93b3_66e6e0d6d8e0.slice/crio-b73a5b3fbb6bb624ab7221b904767a160259b2a50057e53970541c2c6d861a11 WatchSource:0}: Error finding container b73a5b3fbb6bb624ab7221b904767a160259b2a50057e53970541c2c6d861a11: Status 404 returned error can't find the container with id b73a5b3fbb6bb624ab7221b904767a160259b2a50057e53970541c2c6d861a11 Feb 27 17:50:01 crc kubenswrapper[4700]: I0227 17:50:01.117494 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536910-x8r7g"] Feb 27 17:50:01 crc kubenswrapper[4700]: I0227 17:50:01.170384 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536910-x8r7g" event={"ID":"8b9e50e8-ca40-48ba-93b3-66e6e0d6d8e0","Type":"ContainerStarted","Data":"b73a5b3fbb6bb624ab7221b904767a160259b2a50057e53970541c2c6d861a11"} Feb 27 17:50:01 crc kubenswrapper[4700]: I0227 17:50:01.176925 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"306434a9-5de0-42d0-8c10-184238aa5f97","Type":"ContainerStarted","Data":"9e76e6b1cb9bed2b790b956d45b1b23f732b03a2c1e31bfb409ba6f046dbf3f4"} Feb 27 17:50:03 crc kubenswrapper[4700]: I0227 17:50:03.229936 4700 generic.go:334] "Generic (PLEG): container finished" podID="8b9e50e8-ca40-48ba-93b3-66e6e0d6d8e0" containerID="910a60df0f8c6e7a7463b63eabc9ebbc398b40c91e0ef95c5634880606dd1655" exitCode=0 Feb 27 17:50:03 crc kubenswrapper[4700]: I0227 17:50:03.230482 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536910-x8r7g" event={"ID":"8b9e50e8-ca40-48ba-93b3-66e6e0d6d8e0","Type":"ContainerDied","Data":"910a60df0f8c6e7a7463b63eabc9ebbc398b40c91e0ef95c5634880606dd1655"} Feb 27 17:50:04 crc kubenswrapper[4700]: I0227 17:50:04.691712 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536910-x8r7g" Feb 27 17:50:04 crc kubenswrapper[4700]: I0227 17:50:04.759313 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jfxrq\" (UniqueName: \"kubernetes.io/projected/8b9e50e8-ca40-48ba-93b3-66e6e0d6d8e0-kube-api-access-jfxrq\") pod \"8b9e50e8-ca40-48ba-93b3-66e6e0d6d8e0\" (UID: \"8b9e50e8-ca40-48ba-93b3-66e6e0d6d8e0\") " Feb 27 17:50:04 crc kubenswrapper[4700]: I0227 17:50:04.769497 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b9e50e8-ca40-48ba-93b3-66e6e0d6d8e0-kube-api-access-jfxrq" (OuterVolumeSpecName: "kube-api-access-jfxrq") pod "8b9e50e8-ca40-48ba-93b3-66e6e0d6d8e0" (UID: "8b9e50e8-ca40-48ba-93b3-66e6e0d6d8e0"). InnerVolumeSpecName "kube-api-access-jfxrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:50:04 crc kubenswrapper[4700]: I0227 17:50:04.861764 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jfxrq\" (UniqueName: \"kubernetes.io/projected/8b9e50e8-ca40-48ba-93b3-66e6e0d6d8e0-kube-api-access-jfxrq\") on node \"crc\" DevicePath \"\"" Feb 27 17:50:05 crc kubenswrapper[4700]: I0227 17:50:05.256792 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"306434a9-5de0-42d0-8c10-184238aa5f97","Type":"ContainerStarted","Data":"4018f30e7afc1661c1b9fda24d1dd89787a9e1150b971e4e1db53006e3522bfe"} Feb 27 17:50:05 crc kubenswrapper[4700]: I0227 17:50:05.257170 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"306434a9-5de0-42d0-8c10-184238aa5f97","Type":"ContainerStarted","Data":"d2f70465570dd5c7d80abfdc8b6d1755526de7809f6bf2e0ccfa9180c86bf1db"} Feb 27 17:50:05 crc kubenswrapper[4700]: I0227 17:50:05.259939 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536910-x8r7g" event={"ID":"8b9e50e8-ca40-48ba-93b3-66e6e0d6d8e0","Type":"ContainerDied","Data":"b73a5b3fbb6bb624ab7221b904767a160259b2a50057e53970541c2c6d861a11"} Feb 27 17:50:05 crc kubenswrapper[4700]: I0227 17:50:05.259998 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b73a5b3fbb6bb624ab7221b904767a160259b2a50057e53970541c2c6d861a11" Feb 27 17:50:05 crc kubenswrapper[4700]: I0227 17:50:05.260024 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536910-x8r7g" Feb 27 17:50:05 crc kubenswrapper[4700]: I0227 17:50:05.301068 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=19.301041351 podStartE2EDuration="19.301041351s" podCreationTimestamp="2026-02-27 17:49:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 17:50:05.298236328 +0000 UTC m=+2965.283549115" watchObservedRunningTime="2026-02-27 17:50:05.301041351 +0000 UTC m=+2965.286354128" Feb 27 17:50:05 crc kubenswrapper[4700]: I0227 17:50:05.818990 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536904-fdk8p"] Feb 27 17:50:05 crc kubenswrapper[4700]: I0227 17:50:05.838867 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536904-fdk8p"] Feb 27 17:50:06 crc kubenswrapper[4700]: I0227 17:50:06.410550 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:50:06 crc kubenswrapper[4700]: I0227 17:50:06.410653 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:50:06 crc kubenswrapper[4700]: I0227 17:50:06.997997 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc474ff3-1a67-415b-9533-eefbe72a6822" path="/var/lib/kubelet/pods/cc474ff3-1a67-415b-9533-eefbe72a6822/volumes" Feb 27 17:50:06 crc kubenswrapper[4700]: I0227 17:50:06.999166 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Feb 27 17:50:10 crc kubenswrapper[4700]: I0227 17:50:10.294719 4700 scope.go:117] "RemoveContainer" containerID="9f3419d394ef17ef080db92c79c7ae493996b3e66af50e73d3e42294f51bf616" Feb 27 17:50:17 crc kubenswrapper[4700]: I0227 17:50:16.999807 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Feb 27 17:50:17 crc kubenswrapper[4700]: I0227 17:50:17.000407 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Feb 27 17:50:17 crc kubenswrapper[4700]: I0227 17:50:17.012704 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.288256 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Feb 27 17:50:23 crc kubenswrapper[4700]: E0227 17:50:23.289154 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b9e50e8-ca40-48ba-93b3-66e6e0d6d8e0" containerName="oc" Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.289169 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b9e50e8-ca40-48ba-93b3-66e6e0d6d8e0" containerName="oc" Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.289365 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b9e50e8-ca40-48ba-93b3-66e6e0d6d8e0" containerName="oc" Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.290098 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.295645 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.295848 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-nrmqz" Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.295970 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.296446 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.298253 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.412209 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5gwp\" (UniqueName: \"kubernetes.io/projected/805fcd3a-360c-445d-85d8-889bf15d310b-kube-api-access-k5gwp\") pod \"tempest-tests-tempest\" (UID: \"805fcd3a-360c-445d-85d8-889bf15d310b\") " pod="openstack/tempest-tests-tempest" Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.412271 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/805fcd3a-360c-445d-85d8-889bf15d310b-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"805fcd3a-360c-445d-85d8-889bf15d310b\") " pod="openstack/tempest-tests-tempest" Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.412316 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/805fcd3a-360c-445d-85d8-889bf15d310b-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"805fcd3a-360c-445d-85d8-889bf15d310b\") " pod="openstack/tempest-tests-tempest" Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.412535 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/805fcd3a-360c-445d-85d8-889bf15d310b-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"805fcd3a-360c-445d-85d8-889bf15d310b\") " pod="openstack/tempest-tests-tempest" Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.412597 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/805fcd3a-360c-445d-85d8-889bf15d310b-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"805fcd3a-360c-445d-85d8-889bf15d310b\") " pod="openstack/tempest-tests-tempest" Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.412788 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/805fcd3a-360c-445d-85d8-889bf15d310b-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"805fcd3a-360c-445d-85d8-889bf15d310b\") " pod="openstack/tempest-tests-tempest" Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.412980 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"805fcd3a-360c-445d-85d8-889bf15d310b\") " pod="openstack/tempest-tests-tempest" Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.413153 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/805fcd3a-360c-445d-85d8-889bf15d310b-config-data\") pod \"tempest-tests-tempest\" (UID: \"805fcd3a-360c-445d-85d8-889bf15d310b\") " pod="openstack/tempest-tests-tempest" Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.413272 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/805fcd3a-360c-445d-85d8-889bf15d310b-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"805fcd3a-360c-445d-85d8-889bf15d310b\") " pod="openstack/tempest-tests-tempest" Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.515447 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/805fcd3a-360c-445d-85d8-889bf15d310b-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"805fcd3a-360c-445d-85d8-889bf15d310b\") " pod="openstack/tempest-tests-tempest" Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.515529 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5gwp\" (UniqueName: \"kubernetes.io/projected/805fcd3a-360c-445d-85d8-889bf15d310b-kube-api-access-k5gwp\") pod \"tempest-tests-tempest\" (UID: \"805fcd3a-360c-445d-85d8-889bf15d310b\") " pod="openstack/tempest-tests-tempest" Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.515566 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/805fcd3a-360c-445d-85d8-889bf15d310b-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"805fcd3a-360c-445d-85d8-889bf15d310b\") " pod="openstack/tempest-tests-tempest" Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.515586 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/805fcd3a-360c-445d-85d8-889bf15d310b-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"805fcd3a-360c-445d-85d8-889bf15d310b\") " pod="openstack/tempest-tests-tempest" Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.515622 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/805fcd3a-360c-445d-85d8-889bf15d310b-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"805fcd3a-360c-445d-85d8-889bf15d310b\") " pod="openstack/tempest-tests-tempest" Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.515642 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/805fcd3a-360c-445d-85d8-889bf15d310b-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"805fcd3a-360c-445d-85d8-889bf15d310b\") " pod="openstack/tempest-tests-tempest" Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.515687 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/805fcd3a-360c-445d-85d8-889bf15d310b-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"805fcd3a-360c-445d-85d8-889bf15d310b\") " pod="openstack/tempest-tests-tempest" Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.515737 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"805fcd3a-360c-445d-85d8-889bf15d310b\") " pod="openstack/tempest-tests-tempest" Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.515779 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/805fcd3a-360c-445d-85d8-889bf15d310b-config-data\") pod \"tempest-tests-tempest\" (UID: \"805fcd3a-360c-445d-85d8-889bf15d310b\") " pod="openstack/tempest-tests-tempest" Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.515891 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/805fcd3a-360c-445d-85d8-889bf15d310b-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"805fcd3a-360c-445d-85d8-889bf15d310b\") " pod="openstack/tempest-tests-tempest" Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.516814 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/805fcd3a-360c-445d-85d8-889bf15d310b-config-data\") pod \"tempest-tests-tempest\" (UID: \"805fcd3a-360c-445d-85d8-889bf15d310b\") " pod="openstack/tempest-tests-tempest" Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.516972 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/805fcd3a-360c-445d-85d8-889bf15d310b-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"805fcd3a-360c-445d-85d8-889bf15d310b\") " pod="openstack/tempest-tests-tempest" Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.517091 4700 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"805fcd3a-360c-445d-85d8-889bf15d310b\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/tempest-tests-tempest" Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.517680 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/805fcd3a-360c-445d-85d8-889bf15d310b-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"805fcd3a-360c-445d-85d8-889bf15d310b\") " pod="openstack/tempest-tests-tempest" Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.526169 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/805fcd3a-360c-445d-85d8-889bf15d310b-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"805fcd3a-360c-445d-85d8-889bf15d310b\") " pod="openstack/tempest-tests-tempest" Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.530155 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/805fcd3a-360c-445d-85d8-889bf15d310b-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"805fcd3a-360c-445d-85d8-889bf15d310b\") " pod="openstack/tempest-tests-tempest" Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.536271 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/805fcd3a-360c-445d-85d8-889bf15d310b-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"805fcd3a-360c-445d-85d8-889bf15d310b\") " pod="openstack/tempest-tests-tempest" Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.552786 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5gwp\" (UniqueName: \"kubernetes.io/projected/805fcd3a-360c-445d-85d8-889bf15d310b-kube-api-access-k5gwp\") pod \"tempest-tests-tempest\" (UID: \"805fcd3a-360c-445d-85d8-889bf15d310b\") " pod="openstack/tempest-tests-tempest" Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.581714 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"805fcd3a-360c-445d-85d8-889bf15d310b\") " pod="openstack/tempest-tests-tempest" Feb 27 17:50:23 crc kubenswrapper[4700]: I0227 17:50:23.622943 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 27 17:50:24 crc kubenswrapper[4700]: I0227 17:50:24.121430 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Feb 27 17:50:24 crc kubenswrapper[4700]: W0227 17:50:24.124938 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod805fcd3a_360c_445d_85d8_889bf15d310b.slice/crio-c16b5ad8ef3e1c56a689da5097461cd9ac5d9508dd8e8a8766f2406d1017456f WatchSource:0}: Error finding container c16b5ad8ef3e1c56a689da5097461cd9ac5d9508dd8e8a8766f2406d1017456f: Status 404 returned error can't find the container with id c16b5ad8ef3e1c56a689da5097461cd9ac5d9508dd8e8a8766f2406d1017456f Feb 27 17:50:24 crc kubenswrapper[4700]: I0227 17:50:24.534362 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"805fcd3a-360c-445d-85d8-889bf15d310b","Type":"ContainerStarted","Data":"c16b5ad8ef3e1c56a689da5097461cd9ac5d9508dd8e8a8766f2406d1017456f"} Feb 27 17:50:36 crc kubenswrapper[4700]: I0227 17:50:36.411016 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:50:36 crc kubenswrapper[4700]: I0227 17:50:36.411613 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:50:37 crc kubenswrapper[4700]: I0227 17:50:37.701395 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"805fcd3a-360c-445d-85d8-889bf15d310b","Type":"ContainerStarted","Data":"72ffb278d0378cf551b82b535304736b4dcda281beffa87353431f5b4d160ec5"} Feb 27 17:50:37 crc kubenswrapper[4700]: I0227 17:50:37.735152 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.815653764 podStartE2EDuration="15.735118526s" podCreationTimestamp="2026-02-27 17:50:22 +0000 UTC" firstStartedPulling="2026-02-27 17:50:24.128594332 +0000 UTC m=+2984.113907079" lastFinishedPulling="2026-02-27 17:50:36.048059094 +0000 UTC m=+2996.033371841" observedRunningTime="2026-02-27 17:50:37.722432953 +0000 UTC m=+2997.707745700" watchObservedRunningTime="2026-02-27 17:50:37.735118526 +0000 UTC m=+2997.720431283" Feb 27 17:51:02 crc kubenswrapper[4700]: I0227 17:51:02.015994 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8pmlv"] Feb 27 17:51:02 crc kubenswrapper[4700]: I0227 17:51:02.019206 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8pmlv" Feb 27 17:51:02 crc kubenswrapper[4700]: I0227 17:51:02.033284 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8pmlv"] Feb 27 17:51:02 crc kubenswrapper[4700]: I0227 17:51:02.117532 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1a342f0-547f-4e51-a9fe-32115d4ad4fb-catalog-content\") pod \"redhat-marketplace-8pmlv\" (UID: \"f1a342f0-547f-4e51-a9fe-32115d4ad4fb\") " pod="openshift-marketplace/redhat-marketplace-8pmlv" Feb 27 17:51:02 crc kubenswrapper[4700]: I0227 17:51:02.117648 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1a342f0-547f-4e51-a9fe-32115d4ad4fb-utilities\") pod \"redhat-marketplace-8pmlv\" (UID: \"f1a342f0-547f-4e51-a9fe-32115d4ad4fb\") " pod="openshift-marketplace/redhat-marketplace-8pmlv" Feb 27 17:51:02 crc kubenswrapper[4700]: I0227 17:51:02.117672 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5bh5\" (UniqueName: \"kubernetes.io/projected/f1a342f0-547f-4e51-a9fe-32115d4ad4fb-kube-api-access-p5bh5\") pod \"redhat-marketplace-8pmlv\" (UID: \"f1a342f0-547f-4e51-a9fe-32115d4ad4fb\") " pod="openshift-marketplace/redhat-marketplace-8pmlv" Feb 27 17:51:02 crc kubenswrapper[4700]: I0227 17:51:02.219972 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1a342f0-547f-4e51-a9fe-32115d4ad4fb-catalog-content\") pod \"redhat-marketplace-8pmlv\" (UID: \"f1a342f0-547f-4e51-a9fe-32115d4ad4fb\") " pod="openshift-marketplace/redhat-marketplace-8pmlv" Feb 27 17:51:02 crc kubenswrapper[4700]: I0227 17:51:02.220093 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1a342f0-547f-4e51-a9fe-32115d4ad4fb-utilities\") pod \"redhat-marketplace-8pmlv\" (UID: \"f1a342f0-547f-4e51-a9fe-32115d4ad4fb\") " pod="openshift-marketplace/redhat-marketplace-8pmlv" Feb 27 17:51:02 crc kubenswrapper[4700]: I0227 17:51:02.220131 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5bh5\" (UniqueName: \"kubernetes.io/projected/f1a342f0-547f-4e51-a9fe-32115d4ad4fb-kube-api-access-p5bh5\") pod \"redhat-marketplace-8pmlv\" (UID: \"f1a342f0-547f-4e51-a9fe-32115d4ad4fb\") " pod="openshift-marketplace/redhat-marketplace-8pmlv" Feb 27 17:51:02 crc kubenswrapper[4700]: I0227 17:51:02.220830 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1a342f0-547f-4e51-a9fe-32115d4ad4fb-catalog-content\") pod \"redhat-marketplace-8pmlv\" (UID: \"f1a342f0-547f-4e51-a9fe-32115d4ad4fb\") " pod="openshift-marketplace/redhat-marketplace-8pmlv" Feb 27 17:51:02 crc kubenswrapper[4700]: I0227 17:51:02.221331 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1a342f0-547f-4e51-a9fe-32115d4ad4fb-utilities\") pod \"redhat-marketplace-8pmlv\" (UID: \"f1a342f0-547f-4e51-a9fe-32115d4ad4fb\") " pod="openshift-marketplace/redhat-marketplace-8pmlv" Feb 27 17:51:02 crc kubenswrapper[4700]: I0227 17:51:02.272285 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5bh5\" (UniqueName: \"kubernetes.io/projected/f1a342f0-547f-4e51-a9fe-32115d4ad4fb-kube-api-access-p5bh5\") pod \"redhat-marketplace-8pmlv\" (UID: \"f1a342f0-547f-4e51-a9fe-32115d4ad4fb\") " pod="openshift-marketplace/redhat-marketplace-8pmlv" Feb 27 17:51:02 crc kubenswrapper[4700]: I0227 17:51:02.350415 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8pmlv" Feb 27 17:51:02 crc kubenswrapper[4700]: I0227 17:51:02.882397 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8pmlv"] Feb 27 17:51:03 crc kubenswrapper[4700]: I0227 17:51:03.026073 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8pmlv" event={"ID":"f1a342f0-547f-4e51-a9fe-32115d4ad4fb","Type":"ContainerStarted","Data":"a3574123b59a995647b557df7e2483909fcab0b1b59c6fc817440406c073168d"} Feb 27 17:51:04 crc kubenswrapper[4700]: I0227 17:51:04.041300 4700 generic.go:334] "Generic (PLEG): container finished" podID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" containerID="f84f9dd5caea489c35ec6c527d134de5ac0bb45693599b0e9564546c217d59a8" exitCode=0 Feb 27 17:51:04 crc kubenswrapper[4700]: I0227 17:51:04.041386 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8pmlv" event={"ID":"f1a342f0-547f-4e51-a9fe-32115d4ad4fb","Type":"ContainerDied","Data":"f84f9dd5caea489c35ec6c527d134de5ac0bb45693599b0e9564546c217d59a8"} Feb 27 17:51:04 crc kubenswrapper[4700]: I0227 17:51:04.416925 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zrvhs"] Feb 27 17:51:04 crc kubenswrapper[4700]: I0227 17:51:04.421606 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zrvhs" Feb 27 17:51:04 crc kubenswrapper[4700]: I0227 17:51:04.443930 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zrvhs"] Feb 27 17:51:04 crc kubenswrapper[4700]: I0227 17:51:04.479150 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f89dd7d-bf5a-4827-82d0-a3ee8ec76429-utilities\") pod \"community-operators-zrvhs\" (UID: \"6f89dd7d-bf5a-4827-82d0-a3ee8ec76429\") " pod="openshift-marketplace/community-operators-zrvhs" Feb 27 17:51:04 crc kubenswrapper[4700]: I0227 17:51:04.479230 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f89dd7d-bf5a-4827-82d0-a3ee8ec76429-catalog-content\") pod \"community-operators-zrvhs\" (UID: \"6f89dd7d-bf5a-4827-82d0-a3ee8ec76429\") " pod="openshift-marketplace/community-operators-zrvhs" Feb 27 17:51:04 crc kubenswrapper[4700]: I0227 17:51:04.479345 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwp65\" (UniqueName: \"kubernetes.io/projected/6f89dd7d-bf5a-4827-82d0-a3ee8ec76429-kube-api-access-rwp65\") pod \"community-operators-zrvhs\" (UID: \"6f89dd7d-bf5a-4827-82d0-a3ee8ec76429\") " pod="openshift-marketplace/community-operators-zrvhs" Feb 27 17:51:04 crc kubenswrapper[4700]: I0227 17:51:04.581169 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwp65\" (UniqueName: \"kubernetes.io/projected/6f89dd7d-bf5a-4827-82d0-a3ee8ec76429-kube-api-access-rwp65\") pod \"community-operators-zrvhs\" (UID: \"6f89dd7d-bf5a-4827-82d0-a3ee8ec76429\") " pod="openshift-marketplace/community-operators-zrvhs" Feb 27 17:51:04 crc kubenswrapper[4700]: I0227 17:51:04.581396 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f89dd7d-bf5a-4827-82d0-a3ee8ec76429-utilities\") pod \"community-operators-zrvhs\" (UID: \"6f89dd7d-bf5a-4827-82d0-a3ee8ec76429\") " pod="openshift-marketplace/community-operators-zrvhs" Feb 27 17:51:04 crc kubenswrapper[4700]: I0227 17:51:04.581450 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f89dd7d-bf5a-4827-82d0-a3ee8ec76429-catalog-content\") pod \"community-operators-zrvhs\" (UID: \"6f89dd7d-bf5a-4827-82d0-a3ee8ec76429\") " pod="openshift-marketplace/community-operators-zrvhs" Feb 27 17:51:04 crc kubenswrapper[4700]: I0227 17:51:04.582119 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f89dd7d-bf5a-4827-82d0-a3ee8ec76429-utilities\") pod \"community-operators-zrvhs\" (UID: \"6f89dd7d-bf5a-4827-82d0-a3ee8ec76429\") " pod="openshift-marketplace/community-operators-zrvhs" Feb 27 17:51:04 crc kubenswrapper[4700]: I0227 17:51:04.582177 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f89dd7d-bf5a-4827-82d0-a3ee8ec76429-catalog-content\") pod \"community-operators-zrvhs\" (UID: \"6f89dd7d-bf5a-4827-82d0-a3ee8ec76429\") " pod="openshift-marketplace/community-operators-zrvhs" Feb 27 17:51:04 crc kubenswrapper[4700]: I0227 17:51:04.602286 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwp65\" (UniqueName: \"kubernetes.io/projected/6f89dd7d-bf5a-4827-82d0-a3ee8ec76429-kube-api-access-rwp65\") pod \"community-operators-zrvhs\" (UID: \"6f89dd7d-bf5a-4827-82d0-a3ee8ec76429\") " pod="openshift-marketplace/community-operators-zrvhs" Feb 27 17:51:04 crc kubenswrapper[4700]: E0227 17:51:04.697741 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 17:51:04 crc kubenswrapper[4700]: E0227 17:51:04.697902 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p5bh5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-8pmlv_openshift-marketplace(f1a342f0-547f-4e51-a9fe-32115d4ad4fb): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 17:51:04 crc kubenswrapper[4700]: E0227 17:51:04.699141 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-marketplace-8pmlv" podUID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" Feb 27 17:51:04 crc kubenswrapper[4700]: I0227 17:51:04.758234 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zrvhs" Feb 27 17:51:05 crc kubenswrapper[4700]: E0227 17:51:05.055422 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-8pmlv" podUID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" Feb 27 17:51:05 crc kubenswrapper[4700]: I0227 17:51:05.344113 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zrvhs"] Feb 27 17:51:06 crc kubenswrapper[4700]: I0227 17:51:06.063370 4700 generic.go:334] "Generic (PLEG): container finished" podID="6f89dd7d-bf5a-4827-82d0-a3ee8ec76429" containerID="7fbb9845992d74f64bd8507888897920c368ffc20e38590504de631defda1863" exitCode=0 Feb 27 17:51:06 crc kubenswrapper[4700]: I0227 17:51:06.063452 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zrvhs" event={"ID":"6f89dd7d-bf5a-4827-82d0-a3ee8ec76429","Type":"ContainerDied","Data":"7fbb9845992d74f64bd8507888897920c368ffc20e38590504de631defda1863"} Feb 27 17:51:06 crc kubenswrapper[4700]: I0227 17:51:06.063796 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zrvhs" event={"ID":"6f89dd7d-bf5a-4827-82d0-a3ee8ec76429","Type":"ContainerStarted","Data":"6204ff37e65901f83e02c0244af878bf05a4cae6b56064e91aead15a48b7fd52"} Feb 27 17:51:06 crc kubenswrapper[4700]: I0227 17:51:06.410984 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:51:06 crc kubenswrapper[4700]: I0227 17:51:06.411063 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:51:06 crc kubenswrapper[4700]: I0227 17:51:06.411125 4700 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" Feb 27 17:51:06 crc kubenswrapper[4700]: I0227 17:51:06.412206 4700 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8c94c84e8861b740764e2dec223048597911c5008bf61ad17e7b62bcf54abe22"} pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 17:51:06 crc kubenswrapper[4700]: I0227 17:51:06.412312 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" containerID="cri-o://8c94c84e8861b740764e2dec223048597911c5008bf61ad17e7b62bcf54abe22" gracePeriod=600 Feb 27 17:51:06 crc kubenswrapper[4700]: E0227 17:51:06.539951 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:51:07 crc kubenswrapper[4700]: I0227 17:51:07.082625 4700 generic.go:334] "Generic (PLEG): container finished" podID="84b45511-b94c-479f-98db-fd2c4eceec46" containerID="8c94c84e8861b740764e2dec223048597911c5008bf61ad17e7b62bcf54abe22" exitCode=0 Feb 27 17:51:07 crc kubenswrapper[4700]: I0227 17:51:07.082981 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerDied","Data":"8c94c84e8861b740764e2dec223048597911c5008bf61ad17e7b62bcf54abe22"} Feb 27 17:51:07 crc kubenswrapper[4700]: I0227 17:51:07.083020 4700 scope.go:117] "RemoveContainer" containerID="9b478be1943d618189bcf0c56cb81fd36080f5aa6f5ea1159e5d5ecbda877263" Feb 27 17:51:07 crc kubenswrapper[4700]: I0227 17:51:07.083982 4700 scope.go:117] "RemoveContainer" containerID="8c94c84e8861b740764e2dec223048597911c5008bf61ad17e7b62bcf54abe22" Feb 27 17:51:07 crc kubenswrapper[4700]: E0227 17:51:07.084381 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:51:08 crc kubenswrapper[4700]: I0227 17:51:08.102276 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zrvhs" event={"ID":"6f89dd7d-bf5a-4827-82d0-a3ee8ec76429","Type":"ContainerStarted","Data":"3ef382e44fb7d3c73fa90a29a9578cd9ece8a6c74a2661a19ca3e77ef7db5ea6"} Feb 27 17:51:09 crc kubenswrapper[4700]: I0227 17:51:09.133326 4700 generic.go:334] "Generic (PLEG): container finished" podID="6f89dd7d-bf5a-4827-82d0-a3ee8ec76429" containerID="3ef382e44fb7d3c73fa90a29a9578cd9ece8a6c74a2661a19ca3e77ef7db5ea6" exitCode=0 Feb 27 17:51:09 crc kubenswrapper[4700]: I0227 17:51:09.134006 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zrvhs" event={"ID":"6f89dd7d-bf5a-4827-82d0-a3ee8ec76429","Type":"ContainerDied","Data":"3ef382e44fb7d3c73fa90a29a9578cd9ece8a6c74a2661a19ca3e77ef7db5ea6"} Feb 27 17:51:10 crc kubenswrapper[4700]: I0227 17:51:10.151759 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zrvhs" event={"ID":"6f89dd7d-bf5a-4827-82d0-a3ee8ec76429","Type":"ContainerStarted","Data":"3311cef1f617d348e922e10392a82862ffe6be033e27fedfec90af2e840da16e"} Feb 27 17:51:10 crc kubenswrapper[4700]: I0227 17:51:10.182324 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zrvhs" podStartSLOduration=2.59520383 podStartE2EDuration="6.18230549s" podCreationTimestamp="2026-02-27 17:51:04 +0000 UTC" firstStartedPulling="2026-02-27 17:51:06.065222519 +0000 UTC m=+3026.050535266" lastFinishedPulling="2026-02-27 17:51:09.652324149 +0000 UTC m=+3029.637636926" observedRunningTime="2026-02-27 17:51:10.18042358 +0000 UTC m=+3030.165736337" watchObservedRunningTime="2026-02-27 17:51:10.18230549 +0000 UTC m=+3030.167618247" Feb 27 17:51:14 crc kubenswrapper[4700]: I0227 17:51:14.759415 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zrvhs" Feb 27 17:51:14 crc kubenswrapper[4700]: I0227 17:51:14.762796 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zrvhs" Feb 27 17:51:14 crc kubenswrapper[4700]: I0227 17:51:14.847963 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zrvhs" Feb 27 17:51:15 crc kubenswrapper[4700]: I0227 17:51:15.261796 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zrvhs" Feb 27 17:51:16 crc kubenswrapper[4700]: E0227 17:51:16.577922 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 17:51:16 crc kubenswrapper[4700]: E0227 17:51:16.578585 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p5bh5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-8pmlv_openshift-marketplace(f1a342f0-547f-4e51-a9fe-32115d4ad4fb): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 17:51:16 crc kubenswrapper[4700]: E0227 17:51:16.579771 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-marketplace-8pmlv" podUID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" Feb 27 17:51:21 crc kubenswrapper[4700]: I0227 17:51:21.000098 4700 scope.go:117] "RemoveContainer" containerID="8c94c84e8861b740764e2dec223048597911c5008bf61ad17e7b62bcf54abe22" Feb 27 17:51:21 crc kubenswrapper[4700]: E0227 17:51:21.001031 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:51:21 crc kubenswrapper[4700]: I0227 17:51:21.200647 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zrvhs"] Feb 27 17:51:21 crc kubenswrapper[4700]: I0227 17:51:21.201053 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zrvhs" podUID="6f89dd7d-bf5a-4827-82d0-a3ee8ec76429" containerName="registry-server" containerID="cri-o://3311cef1f617d348e922e10392a82862ffe6be033e27fedfec90af2e840da16e" gracePeriod=2 Feb 27 17:51:21 crc kubenswrapper[4700]: I0227 17:51:21.774431 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zrvhs" Feb 27 17:51:21 crc kubenswrapper[4700]: I0227 17:51:21.945247 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwp65\" (UniqueName: \"kubernetes.io/projected/6f89dd7d-bf5a-4827-82d0-a3ee8ec76429-kube-api-access-rwp65\") pod \"6f89dd7d-bf5a-4827-82d0-a3ee8ec76429\" (UID: \"6f89dd7d-bf5a-4827-82d0-a3ee8ec76429\") " Feb 27 17:51:21 crc kubenswrapper[4700]: I0227 17:51:21.945596 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f89dd7d-bf5a-4827-82d0-a3ee8ec76429-utilities\") pod \"6f89dd7d-bf5a-4827-82d0-a3ee8ec76429\" (UID: \"6f89dd7d-bf5a-4827-82d0-a3ee8ec76429\") " Feb 27 17:51:21 crc kubenswrapper[4700]: I0227 17:51:21.945853 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f89dd7d-bf5a-4827-82d0-a3ee8ec76429-catalog-content\") pod \"6f89dd7d-bf5a-4827-82d0-a3ee8ec76429\" (UID: \"6f89dd7d-bf5a-4827-82d0-a3ee8ec76429\") " Feb 27 17:51:21 crc kubenswrapper[4700]: I0227 17:51:21.948005 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f89dd7d-bf5a-4827-82d0-a3ee8ec76429-utilities" (OuterVolumeSpecName: "utilities") pod "6f89dd7d-bf5a-4827-82d0-a3ee8ec76429" (UID: "6f89dd7d-bf5a-4827-82d0-a3ee8ec76429"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:51:21 crc kubenswrapper[4700]: I0227 17:51:21.958792 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f89dd7d-bf5a-4827-82d0-a3ee8ec76429-kube-api-access-rwp65" (OuterVolumeSpecName: "kube-api-access-rwp65") pod "6f89dd7d-bf5a-4827-82d0-a3ee8ec76429" (UID: "6f89dd7d-bf5a-4827-82d0-a3ee8ec76429"). InnerVolumeSpecName "kube-api-access-rwp65". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:51:22 crc kubenswrapper[4700]: I0227 17:51:22.018263 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f89dd7d-bf5a-4827-82d0-a3ee8ec76429-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6f89dd7d-bf5a-4827-82d0-a3ee8ec76429" (UID: "6f89dd7d-bf5a-4827-82d0-a3ee8ec76429"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:51:22 crc kubenswrapper[4700]: I0227 17:51:22.050546 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f89dd7d-bf5a-4827-82d0-a3ee8ec76429-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:51:22 crc kubenswrapper[4700]: I0227 17:51:22.050594 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwp65\" (UniqueName: \"kubernetes.io/projected/6f89dd7d-bf5a-4827-82d0-a3ee8ec76429-kube-api-access-rwp65\") on node \"crc\" DevicePath \"\"" Feb 27 17:51:22 crc kubenswrapper[4700]: I0227 17:51:22.050616 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f89dd7d-bf5a-4827-82d0-a3ee8ec76429-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:51:22 crc kubenswrapper[4700]: I0227 17:51:22.295350 4700 generic.go:334] "Generic (PLEG): container finished" podID="6f89dd7d-bf5a-4827-82d0-a3ee8ec76429" containerID="3311cef1f617d348e922e10392a82862ffe6be033e27fedfec90af2e840da16e" exitCode=0 Feb 27 17:51:22 crc kubenswrapper[4700]: I0227 17:51:22.295482 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zrvhs" Feb 27 17:51:22 crc kubenswrapper[4700]: I0227 17:51:22.295452 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zrvhs" event={"ID":"6f89dd7d-bf5a-4827-82d0-a3ee8ec76429","Type":"ContainerDied","Data":"3311cef1f617d348e922e10392a82862ffe6be033e27fedfec90af2e840da16e"} Feb 27 17:51:22 crc kubenswrapper[4700]: I0227 17:51:22.296303 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zrvhs" event={"ID":"6f89dd7d-bf5a-4827-82d0-a3ee8ec76429","Type":"ContainerDied","Data":"6204ff37e65901f83e02c0244af878bf05a4cae6b56064e91aead15a48b7fd52"} Feb 27 17:51:22 crc kubenswrapper[4700]: I0227 17:51:22.296339 4700 scope.go:117] "RemoveContainer" containerID="3311cef1f617d348e922e10392a82862ffe6be033e27fedfec90af2e840da16e" Feb 27 17:51:22 crc kubenswrapper[4700]: I0227 17:51:22.330778 4700 scope.go:117] "RemoveContainer" containerID="3ef382e44fb7d3c73fa90a29a9578cd9ece8a6c74a2661a19ca3e77ef7db5ea6" Feb 27 17:51:22 crc kubenswrapper[4700]: I0227 17:51:22.337802 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zrvhs"] Feb 27 17:51:22 crc kubenswrapper[4700]: I0227 17:51:22.347112 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zrvhs"] Feb 27 17:51:22 crc kubenswrapper[4700]: I0227 17:51:22.367996 4700 scope.go:117] "RemoveContainer" containerID="7fbb9845992d74f64bd8507888897920c368ffc20e38590504de631defda1863" Feb 27 17:51:22 crc kubenswrapper[4700]: I0227 17:51:22.408217 4700 scope.go:117] "RemoveContainer" containerID="3311cef1f617d348e922e10392a82862ffe6be033e27fedfec90af2e840da16e" Feb 27 17:51:22 crc kubenswrapper[4700]: E0227 17:51:22.408732 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3311cef1f617d348e922e10392a82862ffe6be033e27fedfec90af2e840da16e\": container with ID starting with 3311cef1f617d348e922e10392a82862ffe6be033e27fedfec90af2e840da16e not found: ID does not exist" containerID="3311cef1f617d348e922e10392a82862ffe6be033e27fedfec90af2e840da16e" Feb 27 17:51:22 crc kubenswrapper[4700]: I0227 17:51:22.408782 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3311cef1f617d348e922e10392a82862ffe6be033e27fedfec90af2e840da16e"} err="failed to get container status \"3311cef1f617d348e922e10392a82862ffe6be033e27fedfec90af2e840da16e\": rpc error: code = NotFound desc = could not find container \"3311cef1f617d348e922e10392a82862ffe6be033e27fedfec90af2e840da16e\": container with ID starting with 3311cef1f617d348e922e10392a82862ffe6be033e27fedfec90af2e840da16e not found: ID does not exist" Feb 27 17:51:22 crc kubenswrapper[4700]: I0227 17:51:22.408812 4700 scope.go:117] "RemoveContainer" containerID="3ef382e44fb7d3c73fa90a29a9578cd9ece8a6c74a2661a19ca3e77ef7db5ea6" Feb 27 17:51:22 crc kubenswrapper[4700]: E0227 17:51:22.409258 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ef382e44fb7d3c73fa90a29a9578cd9ece8a6c74a2661a19ca3e77ef7db5ea6\": container with ID starting with 3ef382e44fb7d3c73fa90a29a9578cd9ece8a6c74a2661a19ca3e77ef7db5ea6 not found: ID does not exist" containerID="3ef382e44fb7d3c73fa90a29a9578cd9ece8a6c74a2661a19ca3e77ef7db5ea6" Feb 27 17:51:22 crc kubenswrapper[4700]: I0227 17:51:22.409365 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ef382e44fb7d3c73fa90a29a9578cd9ece8a6c74a2661a19ca3e77ef7db5ea6"} err="failed to get container status \"3ef382e44fb7d3c73fa90a29a9578cd9ece8a6c74a2661a19ca3e77ef7db5ea6\": rpc error: code = NotFound desc = could not find container \"3ef382e44fb7d3c73fa90a29a9578cd9ece8a6c74a2661a19ca3e77ef7db5ea6\": container with ID starting with 3ef382e44fb7d3c73fa90a29a9578cd9ece8a6c74a2661a19ca3e77ef7db5ea6 not found: ID does not exist" Feb 27 17:51:22 crc kubenswrapper[4700]: I0227 17:51:22.409481 4700 scope.go:117] "RemoveContainer" containerID="7fbb9845992d74f64bd8507888897920c368ffc20e38590504de631defda1863" Feb 27 17:51:22 crc kubenswrapper[4700]: E0227 17:51:22.409990 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fbb9845992d74f64bd8507888897920c368ffc20e38590504de631defda1863\": container with ID starting with 7fbb9845992d74f64bd8507888897920c368ffc20e38590504de631defda1863 not found: ID does not exist" containerID="7fbb9845992d74f64bd8507888897920c368ffc20e38590504de631defda1863" Feb 27 17:51:22 crc kubenswrapper[4700]: I0227 17:51:22.410076 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fbb9845992d74f64bd8507888897920c368ffc20e38590504de631defda1863"} err="failed to get container status \"7fbb9845992d74f64bd8507888897920c368ffc20e38590504de631defda1863\": rpc error: code = NotFound desc = could not find container \"7fbb9845992d74f64bd8507888897920c368ffc20e38590504de631defda1863\": container with ID starting with 7fbb9845992d74f64bd8507888897920c368ffc20e38590504de631defda1863 not found: ID does not exist" Feb 27 17:51:22 crc kubenswrapper[4700]: I0227 17:51:22.995053 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f89dd7d-bf5a-4827-82d0-a3ee8ec76429" path="/var/lib/kubelet/pods/6f89dd7d-bf5a-4827-82d0-a3ee8ec76429/volumes" Feb 27 17:51:29 crc kubenswrapper[4700]: E0227 17:51:29.984274 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-8pmlv" podUID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" Feb 27 17:51:34 crc kubenswrapper[4700]: I0227 17:51:34.981213 4700 scope.go:117] "RemoveContainer" containerID="8c94c84e8861b740764e2dec223048597911c5008bf61ad17e7b62bcf54abe22" Feb 27 17:51:34 crc kubenswrapper[4700]: E0227 17:51:34.982131 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:51:49 crc kubenswrapper[4700]: I0227 17:51:49.981298 4700 scope.go:117] "RemoveContainer" containerID="8c94c84e8861b740764e2dec223048597911c5008bf61ad17e7b62bcf54abe22" Feb 27 17:51:49 crc kubenswrapper[4700]: E0227 17:51:49.982357 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:52:00 crc kubenswrapper[4700]: I0227 17:52:00.142271 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536912-f45j2"] Feb 27 17:52:00 crc kubenswrapper[4700]: E0227 17:52:00.143290 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f89dd7d-bf5a-4827-82d0-a3ee8ec76429" containerName="registry-server" Feb 27 17:52:00 crc kubenswrapper[4700]: I0227 17:52:00.143306 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f89dd7d-bf5a-4827-82d0-a3ee8ec76429" containerName="registry-server" Feb 27 17:52:00 crc kubenswrapper[4700]: E0227 17:52:00.143337 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f89dd7d-bf5a-4827-82d0-a3ee8ec76429" containerName="extract-content" Feb 27 17:52:00 crc kubenswrapper[4700]: I0227 17:52:00.143343 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f89dd7d-bf5a-4827-82d0-a3ee8ec76429" containerName="extract-content" Feb 27 17:52:00 crc kubenswrapper[4700]: E0227 17:52:00.143357 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f89dd7d-bf5a-4827-82d0-a3ee8ec76429" containerName="extract-utilities" Feb 27 17:52:00 crc kubenswrapper[4700]: I0227 17:52:00.143363 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f89dd7d-bf5a-4827-82d0-a3ee8ec76429" containerName="extract-utilities" Feb 27 17:52:00 crc kubenswrapper[4700]: I0227 17:52:00.143559 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f89dd7d-bf5a-4827-82d0-a3ee8ec76429" containerName="registry-server" Feb 27 17:52:00 crc kubenswrapper[4700]: I0227 17:52:00.144203 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536912-f45j2" Feb 27 17:52:00 crc kubenswrapper[4700]: I0227 17:52:00.147700 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:52:00 crc kubenswrapper[4700]: I0227 17:52:00.148955 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d62kc" Feb 27 17:52:00 crc kubenswrapper[4700]: I0227 17:52:00.150373 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:52:00 crc kubenswrapper[4700]: I0227 17:52:00.152298 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536912-f45j2"] Feb 27 17:52:00 crc kubenswrapper[4700]: I0227 17:52:00.334516 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9csm\" (UniqueName: \"kubernetes.io/projected/4c9585c5-3cb1-4475-a63c-cfcd8a868cb8-kube-api-access-v9csm\") pod \"auto-csr-approver-29536912-f45j2\" (UID: \"4c9585c5-3cb1-4475-a63c-cfcd8a868cb8\") " pod="openshift-infra/auto-csr-approver-29536912-f45j2" Feb 27 17:52:00 crc kubenswrapper[4700]: I0227 17:52:00.436981 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9csm\" (UniqueName: \"kubernetes.io/projected/4c9585c5-3cb1-4475-a63c-cfcd8a868cb8-kube-api-access-v9csm\") pod \"auto-csr-approver-29536912-f45j2\" (UID: \"4c9585c5-3cb1-4475-a63c-cfcd8a868cb8\") " pod="openshift-infra/auto-csr-approver-29536912-f45j2" Feb 27 17:52:00 crc kubenswrapper[4700]: I0227 17:52:00.458580 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9csm\" (UniqueName: \"kubernetes.io/projected/4c9585c5-3cb1-4475-a63c-cfcd8a868cb8-kube-api-access-v9csm\") pod \"auto-csr-approver-29536912-f45j2\" (UID: \"4c9585c5-3cb1-4475-a63c-cfcd8a868cb8\") " pod="openshift-infra/auto-csr-approver-29536912-f45j2" Feb 27 17:52:00 crc kubenswrapper[4700]: I0227 17:52:00.467638 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536912-f45j2" Feb 27 17:52:00 crc kubenswrapper[4700]: I0227 17:52:00.966127 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536912-f45j2"] Feb 27 17:52:01 crc kubenswrapper[4700]: W0227 17:52:01.014409 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c9585c5_3cb1_4475_a63c_cfcd8a868cb8.slice/crio-050b3e820006e3368d08265c9eccefe2308312bff32745822226b14e871d84f9 WatchSource:0}: Error finding container 050b3e820006e3368d08265c9eccefe2308312bff32745822226b14e871d84f9: Status 404 returned error can't find the container with id 050b3e820006e3368d08265c9eccefe2308312bff32745822226b14e871d84f9 Feb 27 17:52:01 crc kubenswrapper[4700]: I0227 17:52:01.795095 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536912-f45j2" event={"ID":"4c9585c5-3cb1-4475-a63c-cfcd8a868cb8","Type":"ContainerStarted","Data":"050b3e820006e3368d08265c9eccefe2308312bff32745822226b14e871d84f9"} Feb 27 17:52:02 crc kubenswrapper[4700]: I0227 17:52:02.811980 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536912-f45j2" event={"ID":"4c9585c5-3cb1-4475-a63c-cfcd8a868cb8","Type":"ContainerStarted","Data":"1e2bae2f36963bb8444d08508fc7917dc2e55f2a7a74aab3fbc9b71b585d285b"} Feb 27 17:52:02 crc kubenswrapper[4700]: I0227 17:52:02.840605 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536912-f45j2" podStartSLOduration=1.794177517 podStartE2EDuration="2.840573683s" podCreationTimestamp="2026-02-27 17:52:00 +0000 UTC" firstStartedPulling="2026-02-27 17:52:01.022453798 +0000 UTC m=+3081.007766555" lastFinishedPulling="2026-02-27 17:52:02.068849954 +0000 UTC m=+3082.054162721" observedRunningTime="2026-02-27 17:52:02.829868902 +0000 UTC m=+3082.815181659" watchObservedRunningTime="2026-02-27 17:52:02.840573683 +0000 UTC m=+3082.825886430" Feb 27 17:52:03 crc kubenswrapper[4700]: I0227 17:52:03.821922 4700 generic.go:334] "Generic (PLEG): container finished" podID="4c9585c5-3cb1-4475-a63c-cfcd8a868cb8" containerID="1e2bae2f36963bb8444d08508fc7917dc2e55f2a7a74aab3fbc9b71b585d285b" exitCode=0 Feb 27 17:52:03 crc kubenswrapper[4700]: I0227 17:52:03.822043 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536912-f45j2" event={"ID":"4c9585c5-3cb1-4475-a63c-cfcd8a868cb8","Type":"ContainerDied","Data":"1e2bae2f36963bb8444d08508fc7917dc2e55f2a7a74aab3fbc9b71b585d285b"} Feb 27 17:52:03 crc kubenswrapper[4700]: I0227 17:52:03.981206 4700 scope.go:117] "RemoveContainer" containerID="8c94c84e8861b740764e2dec223048597911c5008bf61ad17e7b62bcf54abe22" Feb 27 17:52:03 crc kubenswrapper[4700]: E0227 17:52:03.981862 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:52:05 crc kubenswrapper[4700]: I0227 17:52:05.282928 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536912-f45j2" Feb 27 17:52:05 crc kubenswrapper[4700]: I0227 17:52:05.469484 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9csm\" (UniqueName: \"kubernetes.io/projected/4c9585c5-3cb1-4475-a63c-cfcd8a868cb8-kube-api-access-v9csm\") pod \"4c9585c5-3cb1-4475-a63c-cfcd8a868cb8\" (UID: \"4c9585c5-3cb1-4475-a63c-cfcd8a868cb8\") " Feb 27 17:52:05 crc kubenswrapper[4700]: I0227 17:52:05.484276 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c9585c5-3cb1-4475-a63c-cfcd8a868cb8-kube-api-access-v9csm" (OuterVolumeSpecName: "kube-api-access-v9csm") pod "4c9585c5-3cb1-4475-a63c-cfcd8a868cb8" (UID: "4c9585c5-3cb1-4475-a63c-cfcd8a868cb8"). InnerVolumeSpecName "kube-api-access-v9csm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:52:05 crc kubenswrapper[4700]: I0227 17:52:05.573073 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9csm\" (UniqueName: \"kubernetes.io/projected/4c9585c5-3cb1-4475-a63c-cfcd8a868cb8-kube-api-access-v9csm\") on node \"crc\" DevicePath \"\"" Feb 27 17:52:05 crc kubenswrapper[4700]: I0227 17:52:05.857329 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536912-f45j2" event={"ID":"4c9585c5-3cb1-4475-a63c-cfcd8a868cb8","Type":"ContainerDied","Data":"050b3e820006e3368d08265c9eccefe2308312bff32745822226b14e871d84f9"} Feb 27 17:52:05 crc kubenswrapper[4700]: I0227 17:52:05.857385 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="050b3e820006e3368d08265c9eccefe2308312bff32745822226b14e871d84f9" Feb 27 17:52:05 crc kubenswrapper[4700]: I0227 17:52:05.857489 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536912-f45j2" Feb 27 17:52:05 crc kubenswrapper[4700]: I0227 17:52:05.928286 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536906-lhhr4"] Feb 27 17:52:05 crc kubenswrapper[4700]: I0227 17:52:05.940612 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536906-lhhr4"] Feb 27 17:52:06 crc kubenswrapper[4700]: I0227 17:52:06.993423 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4eb7377-a24a-46df-9aec-b1f977b56708" path="/var/lib/kubelet/pods/c4eb7377-a24a-46df-9aec-b1f977b56708/volumes" Feb 27 17:52:10 crc kubenswrapper[4700]: I0227 17:52:10.515495 4700 scope.go:117] "RemoveContainer" containerID="1d5e81db8e3e48c15ad1a1c701d8c43aade148b25183151df8761b6c50825556" Feb 27 17:52:17 crc kubenswrapper[4700]: E0227 17:52:17.867301 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 17:52:17 crc kubenswrapper[4700]: E0227 17:52:17.868138 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p5bh5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-8pmlv_openshift-marketplace(f1a342f0-547f-4e51-a9fe-32115d4ad4fb): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 17:52:17 crc kubenswrapper[4700]: E0227 17:52:17.869435 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-marketplace-8pmlv" podUID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" Feb 27 17:52:17 crc kubenswrapper[4700]: I0227 17:52:17.980766 4700 scope.go:117] "RemoveContainer" containerID="8c94c84e8861b740764e2dec223048597911c5008bf61ad17e7b62bcf54abe22" Feb 27 17:52:17 crc kubenswrapper[4700]: E0227 17:52:17.981129 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:52:29 crc kubenswrapper[4700]: E0227 17:52:29.983683 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-8pmlv" podUID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" Feb 27 17:52:32 crc kubenswrapper[4700]: I0227 17:52:32.981802 4700 scope.go:117] "RemoveContainer" containerID="8c94c84e8861b740764e2dec223048597911c5008bf61ad17e7b62bcf54abe22" Feb 27 17:52:32 crc kubenswrapper[4700]: E0227 17:52:32.982906 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:52:41 crc kubenswrapper[4700]: E0227 17:52:41.984281 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-8pmlv" podUID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" Feb 27 17:52:46 crc kubenswrapper[4700]: I0227 17:52:46.982437 4700 scope.go:117] "RemoveContainer" containerID="8c94c84e8861b740764e2dec223048597911c5008bf61ad17e7b62bcf54abe22" Feb 27 17:52:46 crc kubenswrapper[4700]: E0227 17:52:46.983391 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:52:53 crc kubenswrapper[4700]: E0227 17:52:53.985626 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-8pmlv" podUID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" Feb 27 17:52:57 crc kubenswrapper[4700]: I0227 17:52:57.981602 4700 scope.go:117] "RemoveContainer" containerID="8c94c84e8861b740764e2dec223048597911c5008bf61ad17e7b62bcf54abe22" Feb 27 17:52:57 crc kubenswrapper[4700]: E0227 17:52:57.982309 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:53:05 crc kubenswrapper[4700]: I0227 17:53:05.984504 4700 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 17:53:06 crc kubenswrapper[4700]: E0227 17:53:06.753750 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 17:53:06 crc kubenswrapper[4700]: E0227 17:53:06.754220 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p5bh5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-8pmlv_openshift-marketplace(f1a342f0-547f-4e51-a9fe-32115d4ad4fb): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 17:53:06 crc kubenswrapper[4700]: E0227 17:53:06.755484 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-marketplace-8pmlv" podUID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" Feb 27 17:53:10 crc kubenswrapper[4700]: I0227 17:53:10.640492 4700 scope.go:117] "RemoveContainer" containerID="eb24abb5cd44aeda29252db9c820f948348c9e7e730de4c77262eebaf09babcb" Feb 27 17:53:10 crc kubenswrapper[4700]: I0227 17:53:10.984425 4700 scope.go:117] "RemoveContainer" containerID="8c94c84e8861b740764e2dec223048597911c5008bf61ad17e7b62bcf54abe22" Feb 27 17:53:10 crc kubenswrapper[4700]: E0227 17:53:10.985891 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:53:17 crc kubenswrapper[4700]: E0227 17:53:17.984096 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-8pmlv" podUID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" Feb 27 17:53:21 crc kubenswrapper[4700]: I0227 17:53:21.981250 4700 scope.go:117] "RemoveContainer" containerID="8c94c84e8861b740764e2dec223048597911c5008bf61ad17e7b62bcf54abe22" Feb 27 17:53:21 crc kubenswrapper[4700]: E0227 17:53:21.981819 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:53:28 crc kubenswrapper[4700]: E0227 17:53:28.984918 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-8pmlv" podUID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" Feb 27 17:53:33 crc kubenswrapper[4700]: I0227 17:53:33.981197 4700 scope.go:117] "RemoveContainer" containerID="8c94c84e8861b740764e2dec223048597911c5008bf61ad17e7b62bcf54abe22" Feb 27 17:53:33 crc kubenswrapper[4700]: E0227 17:53:33.982308 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:53:43 crc kubenswrapper[4700]: E0227 17:53:43.984656 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-8pmlv" podUID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" Feb 27 17:53:48 crc kubenswrapper[4700]: I0227 17:53:48.980941 4700 scope.go:117] "RemoveContainer" containerID="8c94c84e8861b740764e2dec223048597911c5008bf61ad17e7b62bcf54abe22" Feb 27 17:53:48 crc kubenswrapper[4700]: E0227 17:53:48.981760 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:53:57 crc kubenswrapper[4700]: E0227 17:53:57.983864 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-8pmlv" podUID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" Feb 27 17:54:00 crc kubenswrapper[4700]: I0227 17:54:00.166864 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536914-fbzc2"] Feb 27 17:54:00 crc kubenswrapper[4700]: E0227 17:54:00.167921 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c9585c5-3cb1-4475-a63c-cfcd8a868cb8" containerName="oc" Feb 27 17:54:00 crc kubenswrapper[4700]: I0227 17:54:00.167938 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c9585c5-3cb1-4475-a63c-cfcd8a868cb8" containerName="oc" Feb 27 17:54:00 crc kubenswrapper[4700]: I0227 17:54:00.168199 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c9585c5-3cb1-4475-a63c-cfcd8a868cb8" containerName="oc" Feb 27 17:54:00 crc kubenswrapper[4700]: I0227 17:54:00.169103 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536914-fbzc2" Feb 27 17:54:00 crc kubenswrapper[4700]: I0227 17:54:00.172319 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:54:00 crc kubenswrapper[4700]: I0227 17:54:00.172615 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d62kc" Feb 27 17:54:00 crc kubenswrapper[4700]: I0227 17:54:00.172620 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:54:00 crc kubenswrapper[4700]: I0227 17:54:00.182912 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536914-fbzc2"] Feb 27 17:54:00 crc kubenswrapper[4700]: I0227 17:54:00.290667 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmpt5\" (UniqueName: \"kubernetes.io/projected/1ce1cc33-dc54-4b17-8333-fbfdd703c410-kube-api-access-lmpt5\") pod \"auto-csr-approver-29536914-fbzc2\" (UID: \"1ce1cc33-dc54-4b17-8333-fbfdd703c410\") " pod="openshift-infra/auto-csr-approver-29536914-fbzc2" Feb 27 17:54:00 crc kubenswrapper[4700]: I0227 17:54:00.393210 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmpt5\" (UniqueName: \"kubernetes.io/projected/1ce1cc33-dc54-4b17-8333-fbfdd703c410-kube-api-access-lmpt5\") pod \"auto-csr-approver-29536914-fbzc2\" (UID: \"1ce1cc33-dc54-4b17-8333-fbfdd703c410\") " pod="openshift-infra/auto-csr-approver-29536914-fbzc2" Feb 27 17:54:00 crc kubenswrapper[4700]: I0227 17:54:00.432007 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmpt5\" (UniqueName: \"kubernetes.io/projected/1ce1cc33-dc54-4b17-8333-fbfdd703c410-kube-api-access-lmpt5\") pod \"auto-csr-approver-29536914-fbzc2\" (UID: \"1ce1cc33-dc54-4b17-8333-fbfdd703c410\") " pod="openshift-infra/auto-csr-approver-29536914-fbzc2" Feb 27 17:54:00 crc kubenswrapper[4700]: I0227 17:54:00.505329 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536914-fbzc2" Feb 27 17:54:01 crc kubenswrapper[4700]: I0227 17:54:01.010659 4700 scope.go:117] "RemoveContainer" containerID="8c94c84e8861b740764e2dec223048597911c5008bf61ad17e7b62bcf54abe22" Feb 27 17:54:01 crc kubenswrapper[4700]: E0227 17:54:01.011399 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:54:01 crc kubenswrapper[4700]: I0227 17:54:01.014197 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536914-fbzc2"] Feb 27 17:54:01 crc kubenswrapper[4700]: I0227 17:54:01.600115 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536914-fbzc2" event={"ID":"1ce1cc33-dc54-4b17-8333-fbfdd703c410","Type":"ContainerStarted","Data":"bd9ae8cf3fc1fe6e6b35f8193af25ae9d0f904b4a81373d565f45a3697e99dd6"} Feb 27 17:54:02 crc kubenswrapper[4700]: I0227 17:54:02.614077 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536914-fbzc2" event={"ID":"1ce1cc33-dc54-4b17-8333-fbfdd703c410","Type":"ContainerStarted","Data":"61220a5aa3ff7592bbb0f6058f3dbbe3b93fb0e9d72d764a6ca0d75cda8fdeff"} Feb 27 17:54:02 crc kubenswrapper[4700]: I0227 17:54:02.633823 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536914-fbzc2" podStartSLOduration=1.5047854840000001 podStartE2EDuration="2.633795259s" podCreationTimestamp="2026-02-27 17:54:00 +0000 UTC" firstStartedPulling="2026-02-27 17:54:01.028261957 +0000 UTC m=+3201.013574714" lastFinishedPulling="2026-02-27 17:54:02.157271732 +0000 UTC m=+3202.142584489" observedRunningTime="2026-02-27 17:54:02.627595186 +0000 UTC m=+3202.612907943" watchObservedRunningTime="2026-02-27 17:54:02.633795259 +0000 UTC m=+3202.619108016" Feb 27 17:54:03 crc kubenswrapper[4700]: I0227 17:54:03.625948 4700 generic.go:334] "Generic (PLEG): container finished" podID="1ce1cc33-dc54-4b17-8333-fbfdd703c410" containerID="61220a5aa3ff7592bbb0f6058f3dbbe3b93fb0e9d72d764a6ca0d75cda8fdeff" exitCode=0 Feb 27 17:54:03 crc kubenswrapper[4700]: I0227 17:54:03.626023 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536914-fbzc2" event={"ID":"1ce1cc33-dc54-4b17-8333-fbfdd703c410","Type":"ContainerDied","Data":"61220a5aa3ff7592bbb0f6058f3dbbe3b93fb0e9d72d764a6ca0d75cda8fdeff"} Feb 27 17:54:05 crc kubenswrapper[4700]: I0227 17:54:05.024515 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536914-fbzc2" Feb 27 17:54:05 crc kubenswrapper[4700]: I0227 17:54:05.088622 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmpt5\" (UniqueName: \"kubernetes.io/projected/1ce1cc33-dc54-4b17-8333-fbfdd703c410-kube-api-access-lmpt5\") pod \"1ce1cc33-dc54-4b17-8333-fbfdd703c410\" (UID: \"1ce1cc33-dc54-4b17-8333-fbfdd703c410\") " Feb 27 17:54:05 crc kubenswrapper[4700]: I0227 17:54:05.095549 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ce1cc33-dc54-4b17-8333-fbfdd703c410-kube-api-access-lmpt5" (OuterVolumeSpecName: "kube-api-access-lmpt5") pod "1ce1cc33-dc54-4b17-8333-fbfdd703c410" (UID: "1ce1cc33-dc54-4b17-8333-fbfdd703c410"). InnerVolumeSpecName "kube-api-access-lmpt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:54:05 crc kubenswrapper[4700]: I0227 17:54:05.191494 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmpt5\" (UniqueName: \"kubernetes.io/projected/1ce1cc33-dc54-4b17-8333-fbfdd703c410-kube-api-access-lmpt5\") on node \"crc\" DevicePath \"\"" Feb 27 17:54:05 crc kubenswrapper[4700]: I0227 17:54:05.650634 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536914-fbzc2" event={"ID":"1ce1cc33-dc54-4b17-8333-fbfdd703c410","Type":"ContainerDied","Data":"bd9ae8cf3fc1fe6e6b35f8193af25ae9d0f904b4a81373d565f45a3697e99dd6"} Feb 27 17:54:05 crc kubenswrapper[4700]: I0227 17:54:05.651051 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd9ae8cf3fc1fe6e6b35f8193af25ae9d0f904b4a81373d565f45a3697e99dd6" Feb 27 17:54:05 crc kubenswrapper[4700]: I0227 17:54:05.650815 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536914-fbzc2" Feb 27 17:54:05 crc kubenswrapper[4700]: I0227 17:54:05.718251 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536908-djpzb"] Feb 27 17:54:05 crc kubenswrapper[4700]: I0227 17:54:05.728909 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536908-djpzb"] Feb 27 17:54:06 crc kubenswrapper[4700]: I0227 17:54:06.996780 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c22f6346-6080-4170-beff-4fe69f2aa293" path="/var/lib/kubelet/pods/c22f6346-6080-4170-beff-4fe69f2aa293/volumes" Feb 27 17:54:12 crc kubenswrapper[4700]: E0227 17:54:12.987411 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-8pmlv" podUID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" Feb 27 17:54:14 crc kubenswrapper[4700]: I0227 17:54:14.982242 4700 scope.go:117] "RemoveContainer" containerID="8c94c84e8861b740764e2dec223048597911c5008bf61ad17e7b62bcf54abe22" Feb 27 17:54:14 crc kubenswrapper[4700]: E0227 17:54:14.982973 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:54:24 crc kubenswrapper[4700]: E0227 17:54:24.985536 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-8pmlv" podUID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" Feb 27 17:54:25 crc kubenswrapper[4700]: I0227 17:54:25.983393 4700 scope.go:117] "RemoveContainer" containerID="8c94c84e8861b740764e2dec223048597911c5008bf61ad17e7b62bcf54abe22" Feb 27 17:54:25 crc kubenswrapper[4700]: E0227 17:54:25.984926 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:54:40 crc kubenswrapper[4700]: E0227 17:54:40.639357 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 17:54:40 crc kubenswrapper[4700]: E0227 17:54:40.639916 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p5bh5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-8pmlv_openshift-marketplace(f1a342f0-547f-4e51-a9fe-32115d4ad4fb): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 17:54:40 crc kubenswrapper[4700]: E0227 17:54:40.641223 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-marketplace-8pmlv" podUID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" Feb 27 17:54:40 crc kubenswrapper[4700]: I0227 17:54:40.986977 4700 scope.go:117] "RemoveContainer" containerID="8c94c84e8861b740764e2dec223048597911c5008bf61ad17e7b62bcf54abe22" Feb 27 17:54:40 crc kubenswrapper[4700]: E0227 17:54:40.988007 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:54:54 crc kubenswrapper[4700]: E0227 17:54:54.984791 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-8pmlv" podUID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" Feb 27 17:54:55 crc kubenswrapper[4700]: I0227 17:54:55.981808 4700 scope.go:117] "RemoveContainer" containerID="8c94c84e8861b740764e2dec223048597911c5008bf61ad17e7b62bcf54abe22" Feb 27 17:54:55 crc kubenswrapper[4700]: E0227 17:54:55.982339 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:55:07 crc kubenswrapper[4700]: E0227 17:55:07.988117 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-8pmlv" podUID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" Feb 27 17:55:09 crc kubenswrapper[4700]: I0227 17:55:09.982127 4700 scope.go:117] "RemoveContainer" containerID="8c94c84e8861b740764e2dec223048597911c5008bf61ad17e7b62bcf54abe22" Feb 27 17:55:09 crc kubenswrapper[4700]: E0227 17:55:09.983077 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:55:21 crc kubenswrapper[4700]: E0227 17:55:21.985800 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-8pmlv" podUID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" Feb 27 17:55:24 crc kubenswrapper[4700]: I0227 17:55:24.982221 4700 scope.go:117] "RemoveContainer" containerID="8c94c84e8861b740764e2dec223048597911c5008bf61ad17e7b62bcf54abe22" Feb 27 17:55:24 crc kubenswrapper[4700]: E0227 17:55:24.983002 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:55:33 crc kubenswrapper[4700]: E0227 17:55:33.987725 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-8pmlv" podUID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" Feb 27 17:55:39 crc kubenswrapper[4700]: I0227 17:55:39.982525 4700 scope.go:117] "RemoveContainer" containerID="8c94c84e8861b740764e2dec223048597911c5008bf61ad17e7b62bcf54abe22" Feb 27 17:55:39 crc kubenswrapper[4700]: E0227 17:55:39.983615 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:55:44 crc kubenswrapper[4700]: E0227 17:55:44.985175 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-8pmlv" podUID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" Feb 27 17:55:50 crc kubenswrapper[4700]: I0227 17:55:50.990337 4700 scope.go:117] "RemoveContainer" containerID="8c94c84e8861b740764e2dec223048597911c5008bf61ad17e7b62bcf54abe22" Feb 27 17:55:50 crc kubenswrapper[4700]: E0227 17:55:50.991266 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:55:57 crc kubenswrapper[4700]: E0227 17:55:57.983877 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-8pmlv" podUID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" Feb 27 17:56:00 crc kubenswrapper[4700]: I0227 17:56:00.233364 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536916-l7sxp"] Feb 27 17:56:00 crc kubenswrapper[4700]: E0227 17:56:00.234525 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ce1cc33-dc54-4b17-8333-fbfdd703c410" containerName="oc" Feb 27 17:56:00 crc kubenswrapper[4700]: I0227 17:56:00.234549 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ce1cc33-dc54-4b17-8333-fbfdd703c410" containerName="oc" Feb 27 17:56:00 crc kubenswrapper[4700]: I0227 17:56:00.234935 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ce1cc33-dc54-4b17-8333-fbfdd703c410" containerName="oc" Feb 27 17:56:00 crc kubenswrapper[4700]: I0227 17:56:00.236108 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536916-l7sxp" Feb 27 17:56:00 crc kubenswrapper[4700]: I0227 17:56:00.239790 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:56:00 crc kubenswrapper[4700]: I0227 17:56:00.240057 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d62kc" Feb 27 17:56:00 crc kubenswrapper[4700]: I0227 17:56:00.240086 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:56:00 crc kubenswrapper[4700]: I0227 17:56:00.248270 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536916-l7sxp"] Feb 27 17:56:00 crc kubenswrapper[4700]: I0227 17:56:00.326022 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pv6fg\" (UniqueName: \"kubernetes.io/projected/cd4abc15-d07f-46ee-9f1c-98164a4b2783-kube-api-access-pv6fg\") pod \"auto-csr-approver-29536916-l7sxp\" (UID: \"cd4abc15-d07f-46ee-9f1c-98164a4b2783\") " pod="openshift-infra/auto-csr-approver-29536916-l7sxp" Feb 27 17:56:00 crc kubenswrapper[4700]: I0227 17:56:00.428865 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pv6fg\" (UniqueName: \"kubernetes.io/projected/cd4abc15-d07f-46ee-9f1c-98164a4b2783-kube-api-access-pv6fg\") pod \"auto-csr-approver-29536916-l7sxp\" (UID: \"cd4abc15-d07f-46ee-9f1c-98164a4b2783\") " pod="openshift-infra/auto-csr-approver-29536916-l7sxp" Feb 27 17:56:00 crc kubenswrapper[4700]: I0227 17:56:00.456197 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pv6fg\" (UniqueName: \"kubernetes.io/projected/cd4abc15-d07f-46ee-9f1c-98164a4b2783-kube-api-access-pv6fg\") pod \"auto-csr-approver-29536916-l7sxp\" (UID: \"cd4abc15-d07f-46ee-9f1c-98164a4b2783\") " pod="openshift-infra/auto-csr-approver-29536916-l7sxp" Feb 27 17:56:00 crc kubenswrapper[4700]: I0227 17:56:00.578899 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536916-l7sxp" Feb 27 17:56:01 crc kubenswrapper[4700]: I0227 17:56:01.059821 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536916-l7sxp"] Feb 27 17:56:01 crc kubenswrapper[4700]: W0227 17:56:01.068171 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd4abc15_d07f_46ee_9f1c_98164a4b2783.slice/crio-9d68e81af18b76abf72f157aed94ce7325c5cb62633d710a9c1d7fdc532df20c WatchSource:0}: Error finding container 9d68e81af18b76abf72f157aed94ce7325c5cb62633d710a9c1d7fdc532df20c: Status 404 returned error can't find the container with id 9d68e81af18b76abf72f157aed94ce7325c5cb62633d710a9c1d7fdc532df20c Feb 27 17:56:01 crc kubenswrapper[4700]: E0227 17:56:01.965106 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 17:56:01 crc kubenswrapper[4700]: E0227 17:56:01.965581 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 17:56:01 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 17:56:01 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pv6fg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536916-l7sxp_openshift-infra(cd4abc15-d07f-46ee-9f1c-98164a4b2783): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 17:56:01 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 17:56:01 crc kubenswrapper[4700]: E0227 17:56:01.966782 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536916-l7sxp" podUID="cd4abc15-d07f-46ee-9f1c-98164a4b2783" Feb 27 17:56:02 crc kubenswrapper[4700]: I0227 17:56:02.044715 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536916-l7sxp" event={"ID":"cd4abc15-d07f-46ee-9f1c-98164a4b2783","Type":"ContainerStarted","Data":"9d68e81af18b76abf72f157aed94ce7325c5cb62633d710a9c1d7fdc532df20c"} Feb 27 17:56:02 crc kubenswrapper[4700]: E0227 17:56:02.046587 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536916-l7sxp" podUID="cd4abc15-d07f-46ee-9f1c-98164a4b2783" Feb 27 17:56:03 crc kubenswrapper[4700]: E0227 17:56:03.058497 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536916-l7sxp" podUID="cd4abc15-d07f-46ee-9f1c-98164a4b2783" Feb 27 17:56:03 crc kubenswrapper[4700]: I0227 17:56:03.981201 4700 scope.go:117] "RemoveContainer" containerID="8c94c84e8861b740764e2dec223048597911c5008bf61ad17e7b62bcf54abe22" Feb 27 17:56:03 crc kubenswrapper[4700]: E0227 17:56:03.981490 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 17:56:10 crc kubenswrapper[4700]: I0227 17:56:10.826098 4700 scope.go:117] "RemoveContainer" containerID="4706c4faccd28b192e57611ee5997cafeb72162adc3fcbcce8a6c25a87b4f8a9" Feb 27 17:56:10 crc kubenswrapper[4700]: E0227 17:56:10.989418 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-8pmlv" podUID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" Feb 27 17:56:16 crc kubenswrapper[4700]: I0227 17:56:16.982427 4700 scope.go:117] "RemoveContainer" containerID="8c94c84e8861b740764e2dec223048597911c5008bf61ad17e7b62bcf54abe22" Feb 27 17:56:17 crc kubenswrapper[4700]: I0227 17:56:17.213575 4700 generic.go:334] "Generic (PLEG): container finished" podID="cd4abc15-d07f-46ee-9f1c-98164a4b2783" containerID="c8775f85417900834955f825c0a0230a38074772d1cf1c592d6c5dfd1c92d9fc" exitCode=0 Feb 27 17:56:17 crc kubenswrapper[4700]: I0227 17:56:17.213698 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536916-l7sxp" event={"ID":"cd4abc15-d07f-46ee-9f1c-98164a4b2783","Type":"ContainerDied","Data":"c8775f85417900834955f825c0a0230a38074772d1cf1c592d6c5dfd1c92d9fc"} Feb 27 17:56:18 crc kubenswrapper[4700]: I0227 17:56:18.233963 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerStarted","Data":"85becbca57d9b923c2d58f8b905a4fa4ab2c81b51b8ed23006735a99d384929c"} Feb 27 17:56:18 crc kubenswrapper[4700]: I0227 17:56:18.638438 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536916-l7sxp" Feb 27 17:56:18 crc kubenswrapper[4700]: I0227 17:56:18.771724 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pv6fg\" (UniqueName: \"kubernetes.io/projected/cd4abc15-d07f-46ee-9f1c-98164a4b2783-kube-api-access-pv6fg\") pod \"cd4abc15-d07f-46ee-9f1c-98164a4b2783\" (UID: \"cd4abc15-d07f-46ee-9f1c-98164a4b2783\") " Feb 27 17:56:18 crc kubenswrapper[4700]: I0227 17:56:18.792656 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd4abc15-d07f-46ee-9f1c-98164a4b2783-kube-api-access-pv6fg" (OuterVolumeSpecName: "kube-api-access-pv6fg") pod "cd4abc15-d07f-46ee-9f1c-98164a4b2783" (UID: "cd4abc15-d07f-46ee-9f1c-98164a4b2783"). InnerVolumeSpecName "kube-api-access-pv6fg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:56:18 crc kubenswrapper[4700]: I0227 17:56:18.874665 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pv6fg\" (UniqueName: \"kubernetes.io/projected/cd4abc15-d07f-46ee-9f1c-98164a4b2783-kube-api-access-pv6fg\") on node \"crc\" DevicePath \"\"" Feb 27 17:56:19 crc kubenswrapper[4700]: I0227 17:56:19.265856 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536916-l7sxp" event={"ID":"cd4abc15-d07f-46ee-9f1c-98164a4b2783","Type":"ContainerDied","Data":"9d68e81af18b76abf72f157aed94ce7325c5cb62633d710a9c1d7fdc532df20c"} Feb 27 17:56:19 crc kubenswrapper[4700]: I0227 17:56:19.265954 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d68e81af18b76abf72f157aed94ce7325c5cb62633d710a9c1d7fdc532df20c" Feb 27 17:56:19 crc kubenswrapper[4700]: I0227 17:56:19.265972 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536916-l7sxp" Feb 27 17:56:19 crc kubenswrapper[4700]: I0227 17:56:19.735542 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536910-x8r7g"] Feb 27 17:56:19 crc kubenswrapper[4700]: I0227 17:56:19.748069 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536910-x8r7g"] Feb 27 17:56:21 crc kubenswrapper[4700]: I0227 17:56:21.001249 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b9e50e8-ca40-48ba-93b3-66e6e0d6d8e0" path="/var/lib/kubelet/pods/8b9e50e8-ca40-48ba-93b3-66e6e0d6d8e0/volumes" Feb 27 17:56:22 crc kubenswrapper[4700]: E0227 17:56:22.984389 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-8pmlv" podUID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" Feb 27 17:56:36 crc kubenswrapper[4700]: E0227 17:56:36.986175 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-8pmlv" podUID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" Feb 27 17:56:47 crc kubenswrapper[4700]: E0227 17:56:47.984402 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-8pmlv" podUID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" Feb 27 17:56:58 crc kubenswrapper[4700]: E0227 17:56:58.986343 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-8pmlv" podUID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" Feb 27 17:57:10 crc kubenswrapper[4700]: I0227 17:57:10.929745 4700 scope.go:117] "RemoveContainer" containerID="910a60df0f8c6e7a7463b63eabc9ebbc398b40c91e0ef95c5634880606dd1655" Feb 27 17:57:11 crc kubenswrapper[4700]: E0227 17:57:11.023550 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-8pmlv" podUID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" Feb 27 17:57:19 crc kubenswrapper[4700]: I0227 17:57:19.229808 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tjnm7"] Feb 27 17:57:19 crc kubenswrapper[4700]: E0227 17:57:19.231487 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd4abc15-d07f-46ee-9f1c-98164a4b2783" containerName="oc" Feb 27 17:57:19 crc kubenswrapper[4700]: I0227 17:57:19.231514 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd4abc15-d07f-46ee-9f1c-98164a4b2783" containerName="oc" Feb 27 17:57:19 crc kubenswrapper[4700]: I0227 17:57:19.231812 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd4abc15-d07f-46ee-9f1c-98164a4b2783" containerName="oc" Feb 27 17:57:19 crc kubenswrapper[4700]: I0227 17:57:19.233993 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tjnm7" Feb 27 17:57:19 crc kubenswrapper[4700]: I0227 17:57:19.254338 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tjnm7"] Feb 27 17:57:19 crc kubenswrapper[4700]: I0227 17:57:19.423359 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/628b6871-f589-40e3-9126-57f0cb0e3713-utilities\") pod \"redhat-operators-tjnm7\" (UID: \"628b6871-f589-40e3-9126-57f0cb0e3713\") " pod="openshift-marketplace/redhat-operators-tjnm7" Feb 27 17:57:19 crc kubenswrapper[4700]: I0227 17:57:19.423552 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khs8l\" (UniqueName: \"kubernetes.io/projected/628b6871-f589-40e3-9126-57f0cb0e3713-kube-api-access-khs8l\") pod \"redhat-operators-tjnm7\" (UID: \"628b6871-f589-40e3-9126-57f0cb0e3713\") " pod="openshift-marketplace/redhat-operators-tjnm7" Feb 27 17:57:19 crc kubenswrapper[4700]: I0227 17:57:19.423599 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/628b6871-f589-40e3-9126-57f0cb0e3713-catalog-content\") pod \"redhat-operators-tjnm7\" (UID: \"628b6871-f589-40e3-9126-57f0cb0e3713\") " pod="openshift-marketplace/redhat-operators-tjnm7" Feb 27 17:57:19 crc kubenswrapper[4700]: I0227 17:57:19.525405 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khs8l\" (UniqueName: \"kubernetes.io/projected/628b6871-f589-40e3-9126-57f0cb0e3713-kube-api-access-khs8l\") pod \"redhat-operators-tjnm7\" (UID: \"628b6871-f589-40e3-9126-57f0cb0e3713\") " pod="openshift-marketplace/redhat-operators-tjnm7" Feb 27 17:57:19 crc kubenswrapper[4700]: I0227 17:57:19.525489 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/628b6871-f589-40e3-9126-57f0cb0e3713-catalog-content\") pod \"redhat-operators-tjnm7\" (UID: \"628b6871-f589-40e3-9126-57f0cb0e3713\") " pod="openshift-marketplace/redhat-operators-tjnm7" Feb 27 17:57:19 crc kubenswrapper[4700]: I0227 17:57:19.525559 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/628b6871-f589-40e3-9126-57f0cb0e3713-utilities\") pod \"redhat-operators-tjnm7\" (UID: \"628b6871-f589-40e3-9126-57f0cb0e3713\") " pod="openshift-marketplace/redhat-operators-tjnm7" Feb 27 17:57:19 crc kubenswrapper[4700]: I0227 17:57:19.525949 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/628b6871-f589-40e3-9126-57f0cb0e3713-catalog-content\") pod \"redhat-operators-tjnm7\" (UID: \"628b6871-f589-40e3-9126-57f0cb0e3713\") " pod="openshift-marketplace/redhat-operators-tjnm7" Feb 27 17:57:19 crc kubenswrapper[4700]: I0227 17:57:19.525995 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/628b6871-f589-40e3-9126-57f0cb0e3713-utilities\") pod \"redhat-operators-tjnm7\" (UID: \"628b6871-f589-40e3-9126-57f0cb0e3713\") " pod="openshift-marketplace/redhat-operators-tjnm7" Feb 27 17:57:19 crc kubenswrapper[4700]: I0227 17:57:19.549695 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khs8l\" (UniqueName: \"kubernetes.io/projected/628b6871-f589-40e3-9126-57f0cb0e3713-kube-api-access-khs8l\") pod \"redhat-operators-tjnm7\" (UID: \"628b6871-f589-40e3-9126-57f0cb0e3713\") " pod="openshift-marketplace/redhat-operators-tjnm7" Feb 27 17:57:19 crc kubenswrapper[4700]: I0227 17:57:19.587563 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tjnm7" Feb 27 17:57:20 crc kubenswrapper[4700]: I0227 17:57:20.084890 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tjnm7"] Feb 27 17:57:21 crc kubenswrapper[4700]: I0227 17:57:21.010674 4700 generic.go:334] "Generic (PLEG): container finished" podID="628b6871-f589-40e3-9126-57f0cb0e3713" containerID="ac4339eca097a12c774634aa577a958ffc837e1201eea66a5a9ba9d9b8c89cf4" exitCode=0 Feb 27 17:57:21 crc kubenswrapper[4700]: I0227 17:57:21.010823 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tjnm7" event={"ID":"628b6871-f589-40e3-9126-57f0cb0e3713","Type":"ContainerDied","Data":"ac4339eca097a12c774634aa577a958ffc837e1201eea66a5a9ba9d9b8c89cf4"} Feb 27 17:57:21 crc kubenswrapper[4700]: I0227 17:57:21.011187 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tjnm7" event={"ID":"628b6871-f589-40e3-9126-57f0cb0e3713","Type":"ContainerStarted","Data":"9f5439b0841864abf34b7a1723faa4b116c76b104b5e2d813be9fb08a011691d"} Feb 27 17:57:21 crc kubenswrapper[4700]: E0227 17:57:21.651502 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 27 17:57:21 crc kubenswrapper[4700]: E0227 17:57:21.652102 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-khs8l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-tjnm7_openshift-marketplace(628b6871-f589-40e3-9126-57f0cb0e3713): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 17:57:21 crc kubenswrapper[4700]: E0227 17:57:21.653658 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-operators-tjnm7" podUID="628b6871-f589-40e3-9126-57f0cb0e3713" Feb 27 17:57:22 crc kubenswrapper[4700]: E0227 17:57:22.026119 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-tjnm7" podUID="628b6871-f589-40e3-9126-57f0cb0e3713" Feb 27 17:57:25 crc kubenswrapper[4700]: I0227 17:57:25.060073 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8pmlv" event={"ID":"f1a342f0-547f-4e51-a9fe-32115d4ad4fb","Type":"ContainerStarted","Data":"fbb100652760b73904e65c4bbdb5cfddce07a28559121384c6fb8e10c278ab36"} Feb 27 17:57:26 crc kubenswrapper[4700]: I0227 17:57:26.074572 4700 generic.go:334] "Generic (PLEG): container finished" podID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" containerID="fbb100652760b73904e65c4bbdb5cfddce07a28559121384c6fb8e10c278ab36" exitCode=0 Feb 27 17:57:26 crc kubenswrapper[4700]: I0227 17:57:26.074654 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8pmlv" event={"ID":"f1a342f0-547f-4e51-a9fe-32115d4ad4fb","Type":"ContainerDied","Data":"fbb100652760b73904e65c4bbdb5cfddce07a28559121384c6fb8e10c278ab36"} Feb 27 17:57:27 crc kubenswrapper[4700]: I0227 17:57:27.094115 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8pmlv" event={"ID":"f1a342f0-547f-4e51-a9fe-32115d4ad4fb","Type":"ContainerStarted","Data":"e068d6bb97cb8036739ec549c03c69659f6f1aa5364834555a448bdc7168a28e"} Feb 27 17:57:27 crc kubenswrapper[4700]: I0227 17:57:27.124438 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8pmlv" podStartSLOduration=3.662561413 podStartE2EDuration="6m26.124403863s" podCreationTimestamp="2026-02-27 17:51:01 +0000 UTC" firstStartedPulling="2026-02-27 17:51:04.044712378 +0000 UTC m=+3024.030025145" lastFinishedPulling="2026-02-27 17:57:26.506554808 +0000 UTC m=+3406.491867595" observedRunningTime="2026-02-27 17:57:27.119173897 +0000 UTC m=+3407.104486654" watchObservedRunningTime="2026-02-27 17:57:27.124403863 +0000 UTC m=+3407.109716650" Feb 27 17:57:32 crc kubenswrapper[4700]: I0227 17:57:32.350759 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8pmlv" Feb 27 17:57:32 crc kubenswrapper[4700]: I0227 17:57:32.351928 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8pmlv" Feb 27 17:57:32 crc kubenswrapper[4700]: I0227 17:57:32.403797 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8pmlv" Feb 27 17:57:33 crc kubenswrapper[4700]: I0227 17:57:33.235333 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8pmlv" Feb 27 17:57:34 crc kubenswrapper[4700]: I0227 17:57:34.497219 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8pmlv"] Feb 27 17:57:36 crc kubenswrapper[4700]: I0227 17:57:36.208550 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8pmlv" podUID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" containerName="registry-server" containerID="cri-o://e068d6bb97cb8036739ec549c03c69659f6f1aa5364834555a448bdc7168a28e" gracePeriod=2 Feb 27 17:57:36 crc kubenswrapper[4700]: I0227 17:57:36.768881 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8pmlv" Feb 27 17:57:36 crc kubenswrapper[4700]: I0227 17:57:36.882090 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p5bh5\" (UniqueName: \"kubernetes.io/projected/f1a342f0-547f-4e51-a9fe-32115d4ad4fb-kube-api-access-p5bh5\") pod \"f1a342f0-547f-4e51-a9fe-32115d4ad4fb\" (UID: \"f1a342f0-547f-4e51-a9fe-32115d4ad4fb\") " Feb 27 17:57:36 crc kubenswrapper[4700]: I0227 17:57:36.882151 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1a342f0-547f-4e51-a9fe-32115d4ad4fb-catalog-content\") pod \"f1a342f0-547f-4e51-a9fe-32115d4ad4fb\" (UID: \"f1a342f0-547f-4e51-a9fe-32115d4ad4fb\") " Feb 27 17:57:36 crc kubenswrapper[4700]: I0227 17:57:36.882249 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1a342f0-547f-4e51-a9fe-32115d4ad4fb-utilities\") pod \"f1a342f0-547f-4e51-a9fe-32115d4ad4fb\" (UID: \"f1a342f0-547f-4e51-a9fe-32115d4ad4fb\") " Feb 27 17:57:36 crc kubenswrapper[4700]: I0227 17:57:36.883787 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1a342f0-547f-4e51-a9fe-32115d4ad4fb-utilities" (OuterVolumeSpecName: "utilities") pod "f1a342f0-547f-4e51-a9fe-32115d4ad4fb" (UID: "f1a342f0-547f-4e51-a9fe-32115d4ad4fb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:57:36 crc kubenswrapper[4700]: I0227 17:57:36.896177 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1a342f0-547f-4e51-a9fe-32115d4ad4fb-kube-api-access-p5bh5" (OuterVolumeSpecName: "kube-api-access-p5bh5") pod "f1a342f0-547f-4e51-a9fe-32115d4ad4fb" (UID: "f1a342f0-547f-4e51-a9fe-32115d4ad4fb"). InnerVolumeSpecName "kube-api-access-p5bh5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:57:36 crc kubenswrapper[4700]: I0227 17:57:36.921408 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1a342f0-547f-4e51-a9fe-32115d4ad4fb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f1a342f0-547f-4e51-a9fe-32115d4ad4fb" (UID: "f1a342f0-547f-4e51-a9fe-32115d4ad4fb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:57:36 crc kubenswrapper[4700]: I0227 17:57:36.985499 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p5bh5\" (UniqueName: \"kubernetes.io/projected/f1a342f0-547f-4e51-a9fe-32115d4ad4fb-kube-api-access-p5bh5\") on node \"crc\" DevicePath \"\"" Feb 27 17:57:36 crc kubenswrapper[4700]: I0227 17:57:36.985568 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1a342f0-547f-4e51-a9fe-32115d4ad4fb-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:57:36 crc kubenswrapper[4700]: I0227 17:57:36.985595 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1a342f0-547f-4e51-a9fe-32115d4ad4fb-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:57:37 crc kubenswrapper[4700]: I0227 17:57:37.229441 4700 generic.go:334] "Generic (PLEG): container finished" podID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" containerID="e068d6bb97cb8036739ec549c03c69659f6f1aa5364834555a448bdc7168a28e" exitCode=0 Feb 27 17:57:37 crc kubenswrapper[4700]: I0227 17:57:37.229534 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8pmlv" event={"ID":"f1a342f0-547f-4e51-a9fe-32115d4ad4fb","Type":"ContainerDied","Data":"e068d6bb97cb8036739ec549c03c69659f6f1aa5364834555a448bdc7168a28e"} Feb 27 17:57:37 crc kubenswrapper[4700]: I0227 17:57:37.229605 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8pmlv" Feb 27 17:57:37 crc kubenswrapper[4700]: I0227 17:57:37.229652 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8pmlv" event={"ID":"f1a342f0-547f-4e51-a9fe-32115d4ad4fb","Type":"ContainerDied","Data":"a3574123b59a995647b557df7e2483909fcab0b1b59c6fc817440406c073168d"} Feb 27 17:57:37 crc kubenswrapper[4700]: I0227 17:57:37.229684 4700 scope.go:117] "RemoveContainer" containerID="e068d6bb97cb8036739ec549c03c69659f6f1aa5364834555a448bdc7168a28e" Feb 27 17:57:37 crc kubenswrapper[4700]: I0227 17:57:37.263769 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8pmlv"] Feb 27 17:57:37 crc kubenswrapper[4700]: I0227 17:57:37.271583 4700 scope.go:117] "RemoveContainer" containerID="fbb100652760b73904e65c4bbdb5cfddce07a28559121384c6fb8e10c278ab36" Feb 27 17:57:37 crc kubenswrapper[4700]: I0227 17:57:37.279399 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8pmlv"] Feb 27 17:57:37 crc kubenswrapper[4700]: I0227 17:57:37.302549 4700 scope.go:117] "RemoveContainer" containerID="f84f9dd5caea489c35ec6c527d134de5ac0bb45693599b0e9564546c217d59a8" Feb 27 17:57:37 crc kubenswrapper[4700]: I0227 17:57:37.370920 4700 scope.go:117] "RemoveContainer" containerID="e068d6bb97cb8036739ec549c03c69659f6f1aa5364834555a448bdc7168a28e" Feb 27 17:57:37 crc kubenswrapper[4700]: E0227 17:57:37.371827 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e068d6bb97cb8036739ec549c03c69659f6f1aa5364834555a448bdc7168a28e\": container with ID starting with e068d6bb97cb8036739ec549c03c69659f6f1aa5364834555a448bdc7168a28e not found: ID does not exist" containerID="e068d6bb97cb8036739ec549c03c69659f6f1aa5364834555a448bdc7168a28e" Feb 27 17:57:37 crc kubenswrapper[4700]: I0227 17:57:37.371883 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e068d6bb97cb8036739ec549c03c69659f6f1aa5364834555a448bdc7168a28e"} err="failed to get container status \"e068d6bb97cb8036739ec549c03c69659f6f1aa5364834555a448bdc7168a28e\": rpc error: code = NotFound desc = could not find container \"e068d6bb97cb8036739ec549c03c69659f6f1aa5364834555a448bdc7168a28e\": container with ID starting with e068d6bb97cb8036739ec549c03c69659f6f1aa5364834555a448bdc7168a28e not found: ID does not exist" Feb 27 17:57:37 crc kubenswrapper[4700]: I0227 17:57:37.371919 4700 scope.go:117] "RemoveContainer" containerID="fbb100652760b73904e65c4bbdb5cfddce07a28559121384c6fb8e10c278ab36" Feb 27 17:57:37 crc kubenswrapper[4700]: E0227 17:57:37.372433 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbb100652760b73904e65c4bbdb5cfddce07a28559121384c6fb8e10c278ab36\": container with ID starting with fbb100652760b73904e65c4bbdb5cfddce07a28559121384c6fb8e10c278ab36 not found: ID does not exist" containerID="fbb100652760b73904e65c4bbdb5cfddce07a28559121384c6fb8e10c278ab36" Feb 27 17:57:37 crc kubenswrapper[4700]: I0227 17:57:37.372475 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbb100652760b73904e65c4bbdb5cfddce07a28559121384c6fb8e10c278ab36"} err="failed to get container status \"fbb100652760b73904e65c4bbdb5cfddce07a28559121384c6fb8e10c278ab36\": rpc error: code = NotFound desc = could not find container \"fbb100652760b73904e65c4bbdb5cfddce07a28559121384c6fb8e10c278ab36\": container with ID starting with fbb100652760b73904e65c4bbdb5cfddce07a28559121384c6fb8e10c278ab36 not found: ID does not exist" Feb 27 17:57:37 crc kubenswrapper[4700]: I0227 17:57:37.372497 4700 scope.go:117] "RemoveContainer" containerID="f84f9dd5caea489c35ec6c527d134de5ac0bb45693599b0e9564546c217d59a8" Feb 27 17:57:37 crc kubenswrapper[4700]: E0227 17:57:37.372863 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f84f9dd5caea489c35ec6c527d134de5ac0bb45693599b0e9564546c217d59a8\": container with ID starting with f84f9dd5caea489c35ec6c527d134de5ac0bb45693599b0e9564546c217d59a8 not found: ID does not exist" containerID="f84f9dd5caea489c35ec6c527d134de5ac0bb45693599b0e9564546c217d59a8" Feb 27 17:57:37 crc kubenswrapper[4700]: I0227 17:57:37.372900 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f84f9dd5caea489c35ec6c527d134de5ac0bb45693599b0e9564546c217d59a8"} err="failed to get container status \"f84f9dd5caea489c35ec6c527d134de5ac0bb45693599b0e9564546c217d59a8\": rpc error: code = NotFound desc = could not find container \"f84f9dd5caea489c35ec6c527d134de5ac0bb45693599b0e9564546c217d59a8\": container with ID starting with f84f9dd5caea489c35ec6c527d134de5ac0bb45693599b0e9564546c217d59a8 not found: ID does not exist" Feb 27 17:57:38 crc kubenswrapper[4700]: I0227 17:57:38.995943 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" path="/var/lib/kubelet/pods/f1a342f0-547f-4e51-a9fe-32115d4ad4fb/volumes" Feb 27 17:57:50 crc kubenswrapper[4700]: I0227 17:57:50.407738 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tjnm7" event={"ID":"628b6871-f589-40e3-9126-57f0cb0e3713","Type":"ContainerStarted","Data":"c4965181c872744c47f80a8db1a482594a57048012aa5477d47137762bcd342c"} Feb 27 17:57:57 crc kubenswrapper[4700]: I0227 17:57:57.503983 4700 generic.go:334] "Generic (PLEG): container finished" podID="628b6871-f589-40e3-9126-57f0cb0e3713" containerID="c4965181c872744c47f80a8db1a482594a57048012aa5477d47137762bcd342c" exitCode=0 Feb 27 17:57:57 crc kubenswrapper[4700]: I0227 17:57:57.504108 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tjnm7" event={"ID":"628b6871-f589-40e3-9126-57f0cb0e3713","Type":"ContainerDied","Data":"c4965181c872744c47f80a8db1a482594a57048012aa5477d47137762bcd342c"} Feb 27 17:57:58 crc kubenswrapper[4700]: I0227 17:57:58.520579 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tjnm7" event={"ID":"628b6871-f589-40e3-9126-57f0cb0e3713","Type":"ContainerStarted","Data":"2cb8041f93bf6bd408856c60e4b08fd7a2a87d797cf9685370395669640dc675"} Feb 27 17:57:58 crc kubenswrapper[4700]: I0227 17:57:58.546820 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tjnm7" podStartSLOduration=2.5468184259999997 podStartE2EDuration="39.546796641s" podCreationTimestamp="2026-02-27 17:57:19 +0000 UTC" firstStartedPulling="2026-02-27 17:57:21.015255342 +0000 UTC m=+3401.000568119" lastFinishedPulling="2026-02-27 17:57:58.015233567 +0000 UTC m=+3438.000546334" observedRunningTime="2026-02-27 17:57:58.543791633 +0000 UTC m=+3438.529104380" watchObservedRunningTime="2026-02-27 17:57:58.546796641 +0000 UTC m=+3438.532109388" Feb 27 17:57:59 crc kubenswrapper[4700]: I0227 17:57:59.588366 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tjnm7" Feb 27 17:57:59 crc kubenswrapper[4700]: I0227 17:57:59.589119 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tjnm7" Feb 27 17:58:00 crc kubenswrapper[4700]: I0227 17:58:00.170369 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536918-njfw6"] Feb 27 17:58:00 crc kubenswrapper[4700]: E0227 17:58:00.171216 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" containerName="extract-content" Feb 27 17:58:00 crc kubenswrapper[4700]: I0227 17:58:00.171246 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" containerName="extract-content" Feb 27 17:58:00 crc kubenswrapper[4700]: E0227 17:58:00.171287 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" containerName="registry-server" Feb 27 17:58:00 crc kubenswrapper[4700]: I0227 17:58:00.171300 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" containerName="registry-server" Feb 27 17:58:00 crc kubenswrapper[4700]: E0227 17:58:00.171329 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" containerName="extract-utilities" Feb 27 17:58:00 crc kubenswrapper[4700]: I0227 17:58:00.171342 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" containerName="extract-utilities" Feb 27 17:58:00 crc kubenswrapper[4700]: I0227 17:58:00.171813 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1a342f0-547f-4e51-a9fe-32115d4ad4fb" containerName="registry-server" Feb 27 17:58:00 crc kubenswrapper[4700]: I0227 17:58:00.173171 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536918-njfw6" Feb 27 17:58:00 crc kubenswrapper[4700]: I0227 17:58:00.176559 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 17:58:00 crc kubenswrapper[4700]: I0227 17:58:00.177445 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 17:58:00 crc kubenswrapper[4700]: I0227 17:58:00.179430 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d62kc" Feb 27 17:58:00 crc kubenswrapper[4700]: I0227 17:58:00.187723 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536918-njfw6"] Feb 27 17:58:00 crc kubenswrapper[4700]: I0227 17:58:00.301967 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzxfh\" (UniqueName: \"kubernetes.io/projected/8fe3904b-bcc3-473c-ba68-94922e3915cb-kube-api-access-nzxfh\") pod \"auto-csr-approver-29536918-njfw6\" (UID: \"8fe3904b-bcc3-473c-ba68-94922e3915cb\") " pod="openshift-infra/auto-csr-approver-29536918-njfw6" Feb 27 17:58:00 crc kubenswrapper[4700]: I0227 17:58:00.405104 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzxfh\" (UniqueName: \"kubernetes.io/projected/8fe3904b-bcc3-473c-ba68-94922e3915cb-kube-api-access-nzxfh\") pod \"auto-csr-approver-29536918-njfw6\" (UID: \"8fe3904b-bcc3-473c-ba68-94922e3915cb\") " pod="openshift-infra/auto-csr-approver-29536918-njfw6" Feb 27 17:58:00 crc kubenswrapper[4700]: I0227 17:58:00.434200 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzxfh\" (UniqueName: \"kubernetes.io/projected/8fe3904b-bcc3-473c-ba68-94922e3915cb-kube-api-access-nzxfh\") pod \"auto-csr-approver-29536918-njfw6\" (UID: \"8fe3904b-bcc3-473c-ba68-94922e3915cb\") " pod="openshift-infra/auto-csr-approver-29536918-njfw6" Feb 27 17:58:00 crc kubenswrapper[4700]: I0227 17:58:00.495370 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536918-njfw6" Feb 27 17:58:00 crc kubenswrapper[4700]: I0227 17:58:00.670198 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tjnm7" podUID="628b6871-f589-40e3-9126-57f0cb0e3713" containerName="registry-server" probeResult="failure" output=< Feb 27 17:58:00 crc kubenswrapper[4700]: timeout: failed to connect service ":50051" within 1s Feb 27 17:58:00 crc kubenswrapper[4700]: > Feb 27 17:58:00 crc kubenswrapper[4700]: I0227 17:58:00.868503 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536918-njfw6"] Feb 27 17:58:01 crc kubenswrapper[4700]: I0227 17:58:01.559295 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536918-njfw6" event={"ID":"8fe3904b-bcc3-473c-ba68-94922e3915cb","Type":"ContainerStarted","Data":"9bf9f89b6cbaacc37eb31685f2336e018659f77dfc180c65b3791bc46a418344"} Feb 27 17:58:02 crc kubenswrapper[4700]: I0227 17:58:02.570930 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536918-njfw6" event={"ID":"8fe3904b-bcc3-473c-ba68-94922e3915cb","Type":"ContainerStarted","Data":"144b2eabdc01738aa139e79f03eb42021d0d81450267e2c35e9c082b88d79cb3"} Feb 27 17:58:02 crc kubenswrapper[4700]: I0227 17:58:02.591295 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536918-njfw6" podStartSLOduration=1.4431759259999999 podStartE2EDuration="2.591277071s" podCreationTimestamp="2026-02-27 17:58:00 +0000 UTC" firstStartedPulling="2026-02-27 17:58:00.894011902 +0000 UTC m=+3440.879324649" lastFinishedPulling="2026-02-27 17:58:02.042113007 +0000 UTC m=+3442.027425794" observedRunningTime="2026-02-27 17:58:02.589300519 +0000 UTC m=+3442.574613266" watchObservedRunningTime="2026-02-27 17:58:02.591277071 +0000 UTC m=+3442.576589818" Feb 27 17:58:03 crc kubenswrapper[4700]: I0227 17:58:03.583606 4700 generic.go:334] "Generic (PLEG): container finished" podID="8fe3904b-bcc3-473c-ba68-94922e3915cb" containerID="144b2eabdc01738aa139e79f03eb42021d0d81450267e2c35e9c082b88d79cb3" exitCode=0 Feb 27 17:58:03 crc kubenswrapper[4700]: I0227 17:58:03.583660 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536918-njfw6" event={"ID":"8fe3904b-bcc3-473c-ba68-94922e3915cb","Type":"ContainerDied","Data":"144b2eabdc01738aa139e79f03eb42021d0d81450267e2c35e9c082b88d79cb3"} Feb 27 17:58:05 crc kubenswrapper[4700]: I0227 17:58:05.139851 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536918-njfw6" Feb 27 17:58:05 crc kubenswrapper[4700]: I0227 17:58:05.227779 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzxfh\" (UniqueName: \"kubernetes.io/projected/8fe3904b-bcc3-473c-ba68-94922e3915cb-kube-api-access-nzxfh\") pod \"8fe3904b-bcc3-473c-ba68-94922e3915cb\" (UID: \"8fe3904b-bcc3-473c-ba68-94922e3915cb\") " Feb 27 17:58:05 crc kubenswrapper[4700]: I0227 17:58:05.245197 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fe3904b-bcc3-473c-ba68-94922e3915cb-kube-api-access-nzxfh" (OuterVolumeSpecName: "kube-api-access-nzxfh") pod "8fe3904b-bcc3-473c-ba68-94922e3915cb" (UID: "8fe3904b-bcc3-473c-ba68-94922e3915cb"). InnerVolumeSpecName "kube-api-access-nzxfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:58:05 crc kubenswrapper[4700]: I0227 17:58:05.330411 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzxfh\" (UniqueName: \"kubernetes.io/projected/8fe3904b-bcc3-473c-ba68-94922e3915cb-kube-api-access-nzxfh\") on node \"crc\" DevicePath \"\"" Feb 27 17:58:05 crc kubenswrapper[4700]: I0227 17:58:05.615639 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536918-njfw6" event={"ID":"8fe3904b-bcc3-473c-ba68-94922e3915cb","Type":"ContainerDied","Data":"9bf9f89b6cbaacc37eb31685f2336e018659f77dfc180c65b3791bc46a418344"} Feb 27 17:58:05 crc kubenswrapper[4700]: I0227 17:58:05.615693 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9bf9f89b6cbaacc37eb31685f2336e018659f77dfc180c65b3791bc46a418344" Feb 27 17:58:05 crc kubenswrapper[4700]: I0227 17:58:05.615755 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536918-njfw6" Feb 27 17:58:05 crc kubenswrapper[4700]: I0227 17:58:05.685158 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536912-f45j2"] Feb 27 17:58:05 crc kubenswrapper[4700]: I0227 17:58:05.697331 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536912-f45j2"] Feb 27 17:58:07 crc kubenswrapper[4700]: I0227 17:58:07.003112 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c9585c5-3cb1-4475-a63c-cfcd8a868cb8" path="/var/lib/kubelet/pods/4c9585c5-3cb1-4475-a63c-cfcd8a868cb8/volumes" Feb 27 17:58:09 crc kubenswrapper[4700]: I0227 17:58:09.669419 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tjnm7" Feb 27 17:58:09 crc kubenswrapper[4700]: I0227 17:58:09.726636 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tjnm7" Feb 27 17:58:09 crc kubenswrapper[4700]: I0227 17:58:09.928138 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tjnm7"] Feb 27 17:58:11 crc kubenswrapper[4700]: I0227 17:58:11.064215 4700 scope.go:117] "RemoveContainer" containerID="1e2bae2f36963bb8444d08508fc7917dc2e55f2a7a74aab3fbc9b71b585d285b" Feb 27 17:58:11 crc kubenswrapper[4700]: I0227 17:58:11.681330 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tjnm7" podUID="628b6871-f589-40e3-9126-57f0cb0e3713" containerName="registry-server" containerID="cri-o://2cb8041f93bf6bd408856c60e4b08fd7a2a87d797cf9685370395669640dc675" gracePeriod=2 Feb 27 17:58:12 crc kubenswrapper[4700]: I0227 17:58:12.322680 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tjnm7" Feb 27 17:58:12 crc kubenswrapper[4700]: I0227 17:58:12.392144 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khs8l\" (UniqueName: \"kubernetes.io/projected/628b6871-f589-40e3-9126-57f0cb0e3713-kube-api-access-khs8l\") pod \"628b6871-f589-40e3-9126-57f0cb0e3713\" (UID: \"628b6871-f589-40e3-9126-57f0cb0e3713\") " Feb 27 17:58:12 crc kubenswrapper[4700]: I0227 17:58:12.392307 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/628b6871-f589-40e3-9126-57f0cb0e3713-catalog-content\") pod \"628b6871-f589-40e3-9126-57f0cb0e3713\" (UID: \"628b6871-f589-40e3-9126-57f0cb0e3713\") " Feb 27 17:58:12 crc kubenswrapper[4700]: I0227 17:58:12.392413 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/628b6871-f589-40e3-9126-57f0cb0e3713-utilities\") pod \"628b6871-f589-40e3-9126-57f0cb0e3713\" (UID: \"628b6871-f589-40e3-9126-57f0cb0e3713\") " Feb 27 17:58:12 crc kubenswrapper[4700]: I0227 17:58:12.393842 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/628b6871-f589-40e3-9126-57f0cb0e3713-utilities" (OuterVolumeSpecName: "utilities") pod "628b6871-f589-40e3-9126-57f0cb0e3713" (UID: "628b6871-f589-40e3-9126-57f0cb0e3713"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:58:12 crc kubenswrapper[4700]: I0227 17:58:12.399020 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/628b6871-f589-40e3-9126-57f0cb0e3713-kube-api-access-khs8l" (OuterVolumeSpecName: "kube-api-access-khs8l") pod "628b6871-f589-40e3-9126-57f0cb0e3713" (UID: "628b6871-f589-40e3-9126-57f0cb0e3713"). InnerVolumeSpecName "kube-api-access-khs8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 17:58:12 crc kubenswrapper[4700]: I0227 17:58:12.498716 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/628b6871-f589-40e3-9126-57f0cb0e3713-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 17:58:12 crc kubenswrapper[4700]: I0227 17:58:12.499154 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khs8l\" (UniqueName: \"kubernetes.io/projected/628b6871-f589-40e3-9126-57f0cb0e3713-kube-api-access-khs8l\") on node \"crc\" DevicePath \"\"" Feb 27 17:58:12 crc kubenswrapper[4700]: I0227 17:58:12.539923 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/628b6871-f589-40e3-9126-57f0cb0e3713-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "628b6871-f589-40e3-9126-57f0cb0e3713" (UID: "628b6871-f589-40e3-9126-57f0cb0e3713"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 17:58:12 crc kubenswrapper[4700]: I0227 17:58:12.600878 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/628b6871-f589-40e3-9126-57f0cb0e3713-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 17:58:12 crc kubenswrapper[4700]: I0227 17:58:12.695821 4700 generic.go:334] "Generic (PLEG): container finished" podID="628b6871-f589-40e3-9126-57f0cb0e3713" containerID="2cb8041f93bf6bd408856c60e4b08fd7a2a87d797cf9685370395669640dc675" exitCode=0 Feb 27 17:58:12 crc kubenswrapper[4700]: I0227 17:58:12.695925 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tjnm7" event={"ID":"628b6871-f589-40e3-9126-57f0cb0e3713","Type":"ContainerDied","Data":"2cb8041f93bf6bd408856c60e4b08fd7a2a87d797cf9685370395669640dc675"} Feb 27 17:58:12 crc kubenswrapper[4700]: I0227 17:58:12.695963 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tjnm7" Feb 27 17:58:12 crc kubenswrapper[4700]: I0227 17:58:12.696013 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tjnm7" event={"ID":"628b6871-f589-40e3-9126-57f0cb0e3713","Type":"ContainerDied","Data":"9f5439b0841864abf34b7a1723faa4b116c76b104b5e2d813be9fb08a011691d"} Feb 27 17:58:12 crc kubenswrapper[4700]: I0227 17:58:12.696048 4700 scope.go:117] "RemoveContainer" containerID="2cb8041f93bf6bd408856c60e4b08fd7a2a87d797cf9685370395669640dc675" Feb 27 17:58:12 crc kubenswrapper[4700]: I0227 17:58:12.726968 4700 scope.go:117] "RemoveContainer" containerID="c4965181c872744c47f80a8db1a482594a57048012aa5477d47137762bcd342c" Feb 27 17:58:12 crc kubenswrapper[4700]: I0227 17:58:12.758389 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tjnm7"] Feb 27 17:58:12 crc kubenswrapper[4700]: I0227 17:58:12.772003 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tjnm7"] Feb 27 17:58:12 crc kubenswrapper[4700]: I0227 17:58:12.778879 4700 scope.go:117] "RemoveContainer" containerID="ac4339eca097a12c774634aa577a958ffc837e1201eea66a5a9ba9d9b8c89cf4" Feb 27 17:58:12 crc kubenswrapper[4700]: I0227 17:58:12.828394 4700 scope.go:117] "RemoveContainer" containerID="2cb8041f93bf6bd408856c60e4b08fd7a2a87d797cf9685370395669640dc675" Feb 27 17:58:12 crc kubenswrapper[4700]: E0227 17:58:12.829005 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2cb8041f93bf6bd408856c60e4b08fd7a2a87d797cf9685370395669640dc675\": container with ID starting with 2cb8041f93bf6bd408856c60e4b08fd7a2a87d797cf9685370395669640dc675 not found: ID does not exist" containerID="2cb8041f93bf6bd408856c60e4b08fd7a2a87d797cf9685370395669640dc675" Feb 27 17:58:12 crc kubenswrapper[4700]: I0227 17:58:12.829068 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cb8041f93bf6bd408856c60e4b08fd7a2a87d797cf9685370395669640dc675"} err="failed to get container status \"2cb8041f93bf6bd408856c60e4b08fd7a2a87d797cf9685370395669640dc675\": rpc error: code = NotFound desc = could not find container \"2cb8041f93bf6bd408856c60e4b08fd7a2a87d797cf9685370395669640dc675\": container with ID starting with 2cb8041f93bf6bd408856c60e4b08fd7a2a87d797cf9685370395669640dc675 not found: ID does not exist" Feb 27 17:58:12 crc kubenswrapper[4700]: I0227 17:58:12.829102 4700 scope.go:117] "RemoveContainer" containerID="c4965181c872744c47f80a8db1a482594a57048012aa5477d47137762bcd342c" Feb 27 17:58:12 crc kubenswrapper[4700]: E0227 17:58:12.829773 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4965181c872744c47f80a8db1a482594a57048012aa5477d47137762bcd342c\": container with ID starting with c4965181c872744c47f80a8db1a482594a57048012aa5477d47137762bcd342c not found: ID does not exist" containerID="c4965181c872744c47f80a8db1a482594a57048012aa5477d47137762bcd342c" Feb 27 17:58:12 crc kubenswrapper[4700]: I0227 17:58:12.829825 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4965181c872744c47f80a8db1a482594a57048012aa5477d47137762bcd342c"} err="failed to get container status \"c4965181c872744c47f80a8db1a482594a57048012aa5477d47137762bcd342c\": rpc error: code = NotFound desc = could not find container \"c4965181c872744c47f80a8db1a482594a57048012aa5477d47137762bcd342c\": container with ID starting with c4965181c872744c47f80a8db1a482594a57048012aa5477d47137762bcd342c not found: ID does not exist" Feb 27 17:58:12 crc kubenswrapper[4700]: I0227 17:58:12.829855 4700 scope.go:117] "RemoveContainer" containerID="ac4339eca097a12c774634aa577a958ffc837e1201eea66a5a9ba9d9b8c89cf4" Feb 27 17:58:12 crc kubenswrapper[4700]: E0227 17:58:12.830353 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac4339eca097a12c774634aa577a958ffc837e1201eea66a5a9ba9d9b8c89cf4\": container with ID starting with ac4339eca097a12c774634aa577a958ffc837e1201eea66a5a9ba9d9b8c89cf4 not found: ID does not exist" containerID="ac4339eca097a12c774634aa577a958ffc837e1201eea66a5a9ba9d9b8c89cf4" Feb 27 17:58:12 crc kubenswrapper[4700]: I0227 17:58:12.830386 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac4339eca097a12c774634aa577a958ffc837e1201eea66a5a9ba9d9b8c89cf4"} err="failed to get container status \"ac4339eca097a12c774634aa577a958ffc837e1201eea66a5a9ba9d9b8c89cf4\": rpc error: code = NotFound desc = could not find container \"ac4339eca097a12c774634aa577a958ffc837e1201eea66a5a9ba9d9b8c89cf4\": container with ID starting with ac4339eca097a12c774634aa577a958ffc837e1201eea66a5a9ba9d9b8c89cf4 not found: ID does not exist" Feb 27 17:58:12 crc kubenswrapper[4700]: I0227 17:58:12.992647 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="628b6871-f589-40e3-9126-57f0cb0e3713" path="/var/lib/kubelet/pods/628b6871-f589-40e3-9126-57f0cb0e3713/volumes" Feb 27 17:58:36 crc kubenswrapper[4700]: I0227 17:58:36.410881 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:58:36 crc kubenswrapper[4700]: I0227 17:58:36.411544 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:59:06 crc kubenswrapper[4700]: I0227 17:59:06.411446 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:59:06 crc kubenswrapper[4700]: I0227 17:59:06.412117 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:59:36 crc kubenswrapper[4700]: I0227 17:59:36.410961 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 17:59:36 crc kubenswrapper[4700]: I0227 17:59:36.412804 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 17:59:36 crc kubenswrapper[4700]: I0227 17:59:36.412986 4700 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" Feb 27 17:59:36 crc kubenswrapper[4700]: I0227 17:59:36.414015 4700 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"85becbca57d9b923c2d58f8b905a4fa4ab2c81b51b8ed23006735a99d384929c"} pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 17:59:36 crc kubenswrapper[4700]: I0227 17:59:36.414225 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" containerID="cri-o://85becbca57d9b923c2d58f8b905a4fa4ab2c81b51b8ed23006735a99d384929c" gracePeriod=600 Feb 27 17:59:36 crc kubenswrapper[4700]: I0227 17:59:36.720053 4700 generic.go:334] "Generic (PLEG): container finished" podID="84b45511-b94c-479f-98db-fd2c4eceec46" containerID="85becbca57d9b923c2d58f8b905a4fa4ab2c81b51b8ed23006735a99d384929c" exitCode=0 Feb 27 17:59:36 crc kubenswrapper[4700]: I0227 17:59:36.720309 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerDied","Data":"85becbca57d9b923c2d58f8b905a4fa4ab2c81b51b8ed23006735a99d384929c"} Feb 27 17:59:36 crc kubenswrapper[4700]: I0227 17:59:36.720343 4700 scope.go:117] "RemoveContainer" containerID="8c94c84e8861b740764e2dec223048597911c5008bf61ad17e7b62bcf54abe22" Feb 27 17:59:37 crc kubenswrapper[4700]: I0227 17:59:37.734170 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerStarted","Data":"5496fef1f08034cab735e8cc9a87daee6229da7616d276a174410ba8eacb94a6"} Feb 27 18:00:00 crc kubenswrapper[4700]: I0227 18:00:00.154177 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536920-s86bn"] Feb 27 18:00:00 crc kubenswrapper[4700]: E0227 18:00:00.155038 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="628b6871-f589-40e3-9126-57f0cb0e3713" containerName="extract-content" Feb 27 18:00:00 crc kubenswrapper[4700]: I0227 18:00:00.155245 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="628b6871-f589-40e3-9126-57f0cb0e3713" containerName="extract-content" Feb 27 18:00:00 crc kubenswrapper[4700]: E0227 18:00:00.155269 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fe3904b-bcc3-473c-ba68-94922e3915cb" containerName="oc" Feb 27 18:00:00 crc kubenswrapper[4700]: I0227 18:00:00.155286 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fe3904b-bcc3-473c-ba68-94922e3915cb" containerName="oc" Feb 27 18:00:00 crc kubenswrapper[4700]: E0227 18:00:00.155302 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="628b6871-f589-40e3-9126-57f0cb0e3713" containerName="extract-utilities" Feb 27 18:00:00 crc kubenswrapper[4700]: I0227 18:00:00.155310 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="628b6871-f589-40e3-9126-57f0cb0e3713" containerName="extract-utilities" Feb 27 18:00:00 crc kubenswrapper[4700]: E0227 18:00:00.155322 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="628b6871-f589-40e3-9126-57f0cb0e3713" containerName="registry-server" Feb 27 18:00:00 crc kubenswrapper[4700]: I0227 18:00:00.155328 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="628b6871-f589-40e3-9126-57f0cb0e3713" containerName="registry-server" Feb 27 18:00:00 crc kubenswrapper[4700]: I0227 18:00:00.155519 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="628b6871-f589-40e3-9126-57f0cb0e3713" containerName="registry-server" Feb 27 18:00:00 crc kubenswrapper[4700]: I0227 18:00:00.155534 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fe3904b-bcc3-473c-ba68-94922e3915cb" containerName="oc" Feb 27 18:00:00 crc kubenswrapper[4700]: I0227 18:00:00.156193 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536920-s86bn" Feb 27 18:00:00 crc kubenswrapper[4700]: I0227 18:00:00.157946 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 18:00:00 crc kubenswrapper[4700]: I0227 18:00:00.158241 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 18:00:00 crc kubenswrapper[4700]: I0227 18:00:00.167260 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d62kc" Feb 27 18:00:00 crc kubenswrapper[4700]: I0227 18:00:00.171401 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536920-j946s"] Feb 27 18:00:00 crc kubenswrapper[4700]: I0227 18:00:00.181931 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536920-j946s" Feb 27 18:00:00 crc kubenswrapper[4700]: I0227 18:00:00.188447 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 27 18:00:00 crc kubenswrapper[4700]: I0227 18:00:00.188740 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 27 18:00:00 crc kubenswrapper[4700]: I0227 18:00:00.200234 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536920-s86bn"] Feb 27 18:00:00 crc kubenswrapper[4700]: I0227 18:00:00.210601 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536920-j946s"] Feb 27 18:00:00 crc kubenswrapper[4700]: I0227 18:00:00.230338 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px9xr\" (UniqueName: \"kubernetes.io/projected/b1a1cead-0efc-4a4c-9c7e-5ebdfb14086a-kube-api-access-px9xr\") pod \"auto-csr-approver-29536920-s86bn\" (UID: \"b1a1cead-0efc-4a4c-9c7e-5ebdfb14086a\") " pod="openshift-infra/auto-csr-approver-29536920-s86bn" Feb 27 18:00:00 crc kubenswrapper[4700]: I0227 18:00:00.230897 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e133aefd-7a5c-4ff6-8db8-f755e7a72327-config-volume\") pod \"collect-profiles-29536920-j946s\" (UID: \"e133aefd-7a5c-4ff6-8db8-f755e7a72327\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536920-j946s" Feb 27 18:00:00 crc kubenswrapper[4700]: I0227 18:00:00.231031 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzmnm\" (UniqueName: \"kubernetes.io/projected/e133aefd-7a5c-4ff6-8db8-f755e7a72327-kube-api-access-kzmnm\") pod \"collect-profiles-29536920-j946s\" (UID: \"e133aefd-7a5c-4ff6-8db8-f755e7a72327\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536920-j946s" Feb 27 18:00:00 crc kubenswrapper[4700]: I0227 18:00:00.231166 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e133aefd-7a5c-4ff6-8db8-f755e7a72327-secret-volume\") pod \"collect-profiles-29536920-j946s\" (UID: \"e133aefd-7a5c-4ff6-8db8-f755e7a72327\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536920-j946s" Feb 27 18:00:00 crc kubenswrapper[4700]: I0227 18:00:00.332626 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px9xr\" (UniqueName: \"kubernetes.io/projected/b1a1cead-0efc-4a4c-9c7e-5ebdfb14086a-kube-api-access-px9xr\") pod \"auto-csr-approver-29536920-s86bn\" (UID: \"b1a1cead-0efc-4a4c-9c7e-5ebdfb14086a\") " pod="openshift-infra/auto-csr-approver-29536920-s86bn" Feb 27 18:00:00 crc kubenswrapper[4700]: I0227 18:00:00.332769 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e133aefd-7a5c-4ff6-8db8-f755e7a72327-config-volume\") pod \"collect-profiles-29536920-j946s\" (UID: \"e133aefd-7a5c-4ff6-8db8-f755e7a72327\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536920-j946s" Feb 27 18:00:00 crc kubenswrapper[4700]: I0227 18:00:00.332800 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzmnm\" (UniqueName: \"kubernetes.io/projected/e133aefd-7a5c-4ff6-8db8-f755e7a72327-kube-api-access-kzmnm\") pod \"collect-profiles-29536920-j946s\" (UID: \"e133aefd-7a5c-4ff6-8db8-f755e7a72327\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536920-j946s" Feb 27 18:00:00 crc kubenswrapper[4700]: I0227 18:00:00.332835 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e133aefd-7a5c-4ff6-8db8-f755e7a72327-secret-volume\") pod \"collect-profiles-29536920-j946s\" (UID: \"e133aefd-7a5c-4ff6-8db8-f755e7a72327\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536920-j946s" Feb 27 18:00:00 crc kubenswrapper[4700]: I0227 18:00:00.334676 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e133aefd-7a5c-4ff6-8db8-f755e7a72327-config-volume\") pod \"collect-profiles-29536920-j946s\" (UID: \"e133aefd-7a5c-4ff6-8db8-f755e7a72327\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536920-j946s" Feb 27 18:00:00 crc kubenswrapper[4700]: I0227 18:00:00.340419 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e133aefd-7a5c-4ff6-8db8-f755e7a72327-secret-volume\") pod \"collect-profiles-29536920-j946s\" (UID: \"e133aefd-7a5c-4ff6-8db8-f755e7a72327\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536920-j946s" Feb 27 18:00:00 crc kubenswrapper[4700]: I0227 18:00:00.352684 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px9xr\" (UniqueName: \"kubernetes.io/projected/b1a1cead-0efc-4a4c-9c7e-5ebdfb14086a-kube-api-access-px9xr\") pod \"auto-csr-approver-29536920-s86bn\" (UID: \"b1a1cead-0efc-4a4c-9c7e-5ebdfb14086a\") " pod="openshift-infra/auto-csr-approver-29536920-s86bn" Feb 27 18:00:00 crc kubenswrapper[4700]: I0227 18:00:00.352725 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzmnm\" (UniqueName: \"kubernetes.io/projected/e133aefd-7a5c-4ff6-8db8-f755e7a72327-kube-api-access-kzmnm\") pod \"collect-profiles-29536920-j946s\" (UID: \"e133aefd-7a5c-4ff6-8db8-f755e7a72327\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536920-j946s" Feb 27 18:00:00 crc kubenswrapper[4700]: I0227 18:00:00.478167 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536920-s86bn" Feb 27 18:00:00 crc kubenswrapper[4700]: I0227 18:00:00.506747 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536920-j946s" Feb 27 18:00:01 crc kubenswrapper[4700]: I0227 18:00:01.045492 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536920-s86bn"] Feb 27 18:00:01 crc kubenswrapper[4700]: I0227 18:00:01.056437 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536920-j946s"] Feb 27 18:00:01 crc kubenswrapper[4700]: I0227 18:00:01.064794 4700 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 18:00:01 crc kubenswrapper[4700]: I0227 18:00:01.990600 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536920-s86bn" event={"ID":"b1a1cead-0efc-4a4c-9c7e-5ebdfb14086a","Type":"ContainerStarted","Data":"9c425c6d77b69e48a0b9665cff93eb2ce684a0ea1319f0fcaffacd777734860b"} Feb 27 18:00:01 crc kubenswrapper[4700]: I0227 18:00:01.992830 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536920-j946s" event={"ID":"e133aefd-7a5c-4ff6-8db8-f755e7a72327","Type":"ContainerStarted","Data":"b6054d92e394579612e9125748b8a31d02b92057bdd6b80d9c35254d3dd86e0a"} Feb 27 18:00:01 crc kubenswrapper[4700]: I0227 18:00:01.992893 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536920-j946s" event={"ID":"e133aefd-7a5c-4ff6-8db8-f755e7a72327","Type":"ContainerStarted","Data":"52853435be7a156544db0217ac7766d14022730982a74b8d312952bd5ae612e1"} Feb 27 18:00:02 crc kubenswrapper[4700]: I0227 18:00:02.017819 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29536920-j946s" podStartSLOduration=2.017789581 podStartE2EDuration="2.017789581s" podCreationTimestamp="2026-02-27 18:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 18:00:02.009481114 +0000 UTC m=+3561.994793871" watchObservedRunningTime="2026-02-27 18:00:02.017789581 +0000 UTC m=+3562.003102348" Feb 27 18:00:03 crc kubenswrapper[4700]: I0227 18:00:03.001719 4700 generic.go:334] "Generic (PLEG): container finished" podID="e133aefd-7a5c-4ff6-8db8-f755e7a72327" containerID="b6054d92e394579612e9125748b8a31d02b92057bdd6b80d9c35254d3dd86e0a" exitCode=0 Feb 27 18:00:03 crc kubenswrapper[4700]: I0227 18:00:03.001809 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536920-j946s" event={"ID":"e133aefd-7a5c-4ff6-8db8-f755e7a72327","Type":"ContainerDied","Data":"b6054d92e394579612e9125748b8a31d02b92057bdd6b80d9c35254d3dd86e0a"} Feb 27 18:00:04 crc kubenswrapper[4700]: I0227 18:00:04.432591 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536920-j946s" Feb 27 18:00:04 crc kubenswrapper[4700]: I0227 18:00:04.528799 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzmnm\" (UniqueName: \"kubernetes.io/projected/e133aefd-7a5c-4ff6-8db8-f755e7a72327-kube-api-access-kzmnm\") pod \"e133aefd-7a5c-4ff6-8db8-f755e7a72327\" (UID: \"e133aefd-7a5c-4ff6-8db8-f755e7a72327\") " Feb 27 18:00:04 crc kubenswrapper[4700]: I0227 18:00:04.529177 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e133aefd-7a5c-4ff6-8db8-f755e7a72327-secret-volume\") pod \"e133aefd-7a5c-4ff6-8db8-f755e7a72327\" (UID: \"e133aefd-7a5c-4ff6-8db8-f755e7a72327\") " Feb 27 18:00:04 crc kubenswrapper[4700]: I0227 18:00:04.529225 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e133aefd-7a5c-4ff6-8db8-f755e7a72327-config-volume\") pod \"e133aefd-7a5c-4ff6-8db8-f755e7a72327\" (UID: \"e133aefd-7a5c-4ff6-8db8-f755e7a72327\") " Feb 27 18:00:04 crc kubenswrapper[4700]: I0227 18:00:04.529733 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e133aefd-7a5c-4ff6-8db8-f755e7a72327-config-volume" (OuterVolumeSpecName: "config-volume") pod "e133aefd-7a5c-4ff6-8db8-f755e7a72327" (UID: "e133aefd-7a5c-4ff6-8db8-f755e7a72327"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 18:00:04 crc kubenswrapper[4700]: I0227 18:00:04.530284 4700 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e133aefd-7a5c-4ff6-8db8-f755e7a72327-config-volume\") on node \"crc\" DevicePath \"\"" Feb 27 18:00:04 crc kubenswrapper[4700]: I0227 18:00:04.540899 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e133aefd-7a5c-4ff6-8db8-f755e7a72327-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e133aefd-7a5c-4ff6-8db8-f755e7a72327" (UID: "e133aefd-7a5c-4ff6-8db8-f755e7a72327"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 18:00:04 crc kubenswrapper[4700]: I0227 18:00:04.546542 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e133aefd-7a5c-4ff6-8db8-f755e7a72327-kube-api-access-kzmnm" (OuterVolumeSpecName: "kube-api-access-kzmnm") pod "e133aefd-7a5c-4ff6-8db8-f755e7a72327" (UID: "e133aefd-7a5c-4ff6-8db8-f755e7a72327"). InnerVolumeSpecName "kube-api-access-kzmnm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:00:04 crc kubenswrapper[4700]: I0227 18:00:04.632362 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzmnm\" (UniqueName: \"kubernetes.io/projected/e133aefd-7a5c-4ff6-8db8-f755e7a72327-kube-api-access-kzmnm\") on node \"crc\" DevicePath \"\"" Feb 27 18:00:04 crc kubenswrapper[4700]: I0227 18:00:04.632425 4700 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e133aefd-7a5c-4ff6-8db8-f755e7a72327-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 27 18:00:05 crc kubenswrapper[4700]: I0227 18:00:05.031543 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536920-j946s" event={"ID":"e133aefd-7a5c-4ff6-8db8-f755e7a72327","Type":"ContainerDied","Data":"52853435be7a156544db0217ac7766d14022730982a74b8d312952bd5ae612e1"} Feb 27 18:00:05 crc kubenswrapper[4700]: I0227 18:00:05.031606 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52853435be7a156544db0217ac7766d14022730982a74b8d312952bd5ae612e1" Feb 27 18:00:05 crc kubenswrapper[4700]: I0227 18:00:05.031695 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536920-j946s" Feb 27 18:00:05 crc kubenswrapper[4700]: I0227 18:00:05.089005 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536875-hzbmd"] Feb 27 18:00:05 crc kubenswrapper[4700]: I0227 18:00:05.097638 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536875-hzbmd"] Feb 27 18:00:06 crc kubenswrapper[4700]: I0227 18:00:06.042107 4700 generic.go:334] "Generic (PLEG): container finished" podID="b1a1cead-0efc-4a4c-9c7e-5ebdfb14086a" containerID="30ba2e2521752f4b2c379c95364dca6e343fd4dbbcfe85d1673eb48311cf395e" exitCode=0 Feb 27 18:00:06 crc kubenswrapper[4700]: I0227 18:00:06.042212 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536920-s86bn" event={"ID":"b1a1cead-0efc-4a4c-9c7e-5ebdfb14086a","Type":"ContainerDied","Data":"30ba2e2521752f4b2c379c95364dca6e343fd4dbbcfe85d1673eb48311cf395e"} Feb 27 18:00:07 crc kubenswrapper[4700]: I0227 18:00:07.001566 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31b14208-847f-485b-90c9-fb6bb3332143" path="/var/lib/kubelet/pods/31b14208-847f-485b-90c9-fb6bb3332143/volumes" Feb 27 18:00:07 crc kubenswrapper[4700]: I0227 18:00:07.459473 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536920-s86bn" Feb 27 18:00:07 crc kubenswrapper[4700]: I0227 18:00:07.610268 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-px9xr\" (UniqueName: \"kubernetes.io/projected/b1a1cead-0efc-4a4c-9c7e-5ebdfb14086a-kube-api-access-px9xr\") pod \"b1a1cead-0efc-4a4c-9c7e-5ebdfb14086a\" (UID: \"b1a1cead-0efc-4a4c-9c7e-5ebdfb14086a\") " Feb 27 18:00:07 crc kubenswrapper[4700]: I0227 18:00:07.628071 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1a1cead-0efc-4a4c-9c7e-5ebdfb14086a-kube-api-access-px9xr" (OuterVolumeSpecName: "kube-api-access-px9xr") pod "b1a1cead-0efc-4a4c-9c7e-5ebdfb14086a" (UID: "b1a1cead-0efc-4a4c-9c7e-5ebdfb14086a"). InnerVolumeSpecName "kube-api-access-px9xr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:00:07 crc kubenswrapper[4700]: I0227 18:00:07.712608 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-px9xr\" (UniqueName: \"kubernetes.io/projected/b1a1cead-0efc-4a4c-9c7e-5ebdfb14086a-kube-api-access-px9xr\") on node \"crc\" DevicePath \"\"" Feb 27 18:00:08 crc kubenswrapper[4700]: I0227 18:00:08.063260 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536920-s86bn" event={"ID":"b1a1cead-0efc-4a4c-9c7e-5ebdfb14086a","Type":"ContainerDied","Data":"9c425c6d77b69e48a0b9665cff93eb2ce684a0ea1319f0fcaffacd777734860b"} Feb 27 18:00:08 crc kubenswrapper[4700]: I0227 18:00:08.063297 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9c425c6d77b69e48a0b9665cff93eb2ce684a0ea1319f0fcaffacd777734860b" Feb 27 18:00:08 crc kubenswrapper[4700]: I0227 18:00:08.063356 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536920-s86bn" Feb 27 18:00:08 crc kubenswrapper[4700]: E0227 18:00:08.273765 4700 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1a1cead_0efc_4a4c_9c7e_5ebdfb14086a.slice/crio-9c425c6d77b69e48a0b9665cff93eb2ce684a0ea1319f0fcaffacd777734860b\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1a1cead_0efc_4a4c_9c7e_5ebdfb14086a.slice\": RecentStats: unable to find data in memory cache]" Feb 27 18:00:08 crc kubenswrapper[4700]: I0227 18:00:08.524649 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536914-fbzc2"] Feb 27 18:00:08 crc kubenswrapper[4700]: I0227 18:00:08.534803 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536914-fbzc2"] Feb 27 18:00:08 crc kubenswrapper[4700]: I0227 18:00:08.991978 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ce1cc33-dc54-4b17-8333-fbfdd703c410" path="/var/lib/kubelet/pods/1ce1cc33-dc54-4b17-8333-fbfdd703c410/volumes" Feb 27 18:00:10 crc kubenswrapper[4700]: I0227 18:00:10.240931 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2cflz"] Feb 27 18:00:10 crc kubenswrapper[4700]: E0227 18:00:10.242041 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e133aefd-7a5c-4ff6-8db8-f755e7a72327" containerName="collect-profiles" Feb 27 18:00:10 crc kubenswrapper[4700]: I0227 18:00:10.242080 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="e133aefd-7a5c-4ff6-8db8-f755e7a72327" containerName="collect-profiles" Feb 27 18:00:10 crc kubenswrapper[4700]: E0227 18:00:10.242099 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1a1cead-0efc-4a4c-9c7e-5ebdfb14086a" containerName="oc" Feb 27 18:00:10 crc kubenswrapper[4700]: I0227 18:00:10.242107 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1a1cead-0efc-4a4c-9c7e-5ebdfb14086a" containerName="oc" Feb 27 18:00:10 crc kubenswrapper[4700]: I0227 18:00:10.242365 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="e133aefd-7a5c-4ff6-8db8-f755e7a72327" containerName="collect-profiles" Feb 27 18:00:10 crc kubenswrapper[4700]: I0227 18:00:10.242392 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1a1cead-0efc-4a4c-9c7e-5ebdfb14086a" containerName="oc" Feb 27 18:00:10 crc kubenswrapper[4700]: I0227 18:00:10.244413 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2cflz" Feb 27 18:00:10 crc kubenswrapper[4700]: I0227 18:00:10.252946 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2cflz"] Feb 27 18:00:10 crc kubenswrapper[4700]: I0227 18:00:10.269477 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9n8w\" (UniqueName: \"kubernetes.io/projected/8e084226-907c-4278-8223-34f6fa78a9a4-kube-api-access-x9n8w\") pod \"certified-operators-2cflz\" (UID: \"8e084226-907c-4278-8223-34f6fa78a9a4\") " pod="openshift-marketplace/certified-operators-2cflz" Feb 27 18:00:10 crc kubenswrapper[4700]: I0227 18:00:10.269540 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e084226-907c-4278-8223-34f6fa78a9a4-catalog-content\") pod \"certified-operators-2cflz\" (UID: \"8e084226-907c-4278-8223-34f6fa78a9a4\") " pod="openshift-marketplace/certified-operators-2cflz" Feb 27 18:00:10 crc kubenswrapper[4700]: I0227 18:00:10.269694 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e084226-907c-4278-8223-34f6fa78a9a4-utilities\") pod \"certified-operators-2cflz\" (UID: \"8e084226-907c-4278-8223-34f6fa78a9a4\") " pod="openshift-marketplace/certified-operators-2cflz" Feb 27 18:00:10 crc kubenswrapper[4700]: I0227 18:00:10.370989 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9n8w\" (UniqueName: \"kubernetes.io/projected/8e084226-907c-4278-8223-34f6fa78a9a4-kube-api-access-x9n8w\") pod \"certified-operators-2cflz\" (UID: \"8e084226-907c-4278-8223-34f6fa78a9a4\") " pod="openshift-marketplace/certified-operators-2cflz" Feb 27 18:00:10 crc kubenswrapper[4700]: I0227 18:00:10.371240 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e084226-907c-4278-8223-34f6fa78a9a4-catalog-content\") pod \"certified-operators-2cflz\" (UID: \"8e084226-907c-4278-8223-34f6fa78a9a4\") " pod="openshift-marketplace/certified-operators-2cflz" Feb 27 18:00:10 crc kubenswrapper[4700]: I0227 18:00:10.371412 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e084226-907c-4278-8223-34f6fa78a9a4-utilities\") pod \"certified-operators-2cflz\" (UID: \"8e084226-907c-4278-8223-34f6fa78a9a4\") " pod="openshift-marketplace/certified-operators-2cflz" Feb 27 18:00:10 crc kubenswrapper[4700]: I0227 18:00:10.371765 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e084226-907c-4278-8223-34f6fa78a9a4-catalog-content\") pod \"certified-operators-2cflz\" (UID: \"8e084226-907c-4278-8223-34f6fa78a9a4\") " pod="openshift-marketplace/certified-operators-2cflz" Feb 27 18:00:10 crc kubenswrapper[4700]: I0227 18:00:10.371808 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e084226-907c-4278-8223-34f6fa78a9a4-utilities\") pod \"certified-operators-2cflz\" (UID: \"8e084226-907c-4278-8223-34f6fa78a9a4\") " pod="openshift-marketplace/certified-operators-2cflz" Feb 27 18:00:10 crc kubenswrapper[4700]: I0227 18:00:10.399099 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9n8w\" (UniqueName: \"kubernetes.io/projected/8e084226-907c-4278-8223-34f6fa78a9a4-kube-api-access-x9n8w\") pod \"certified-operators-2cflz\" (UID: \"8e084226-907c-4278-8223-34f6fa78a9a4\") " pod="openshift-marketplace/certified-operators-2cflz" Feb 27 18:00:10 crc kubenswrapper[4700]: I0227 18:00:10.571052 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2cflz" Feb 27 18:00:11 crc kubenswrapper[4700]: I0227 18:00:11.136920 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2cflz"] Feb 27 18:00:11 crc kubenswrapper[4700]: I0227 18:00:11.250544 4700 scope.go:117] "RemoveContainer" containerID="61220a5aa3ff7592bbb0f6058f3dbbe3b93fb0e9d72d764a6ca0d75cda8fdeff" Feb 27 18:00:11 crc kubenswrapper[4700]: I0227 18:00:11.300915 4700 scope.go:117] "RemoveContainer" containerID="219100e53b01503afacb5172c0c80e3e9b408ba296258bd707778732d4f9a196" Feb 27 18:00:12 crc kubenswrapper[4700]: I0227 18:00:12.110597 4700 generic.go:334] "Generic (PLEG): container finished" podID="8e084226-907c-4278-8223-34f6fa78a9a4" containerID="58779ac75cbde384a3aa88df750835504c26ee66a6901a033019925eba11d358" exitCode=0 Feb 27 18:00:12 crc kubenswrapper[4700]: I0227 18:00:12.110705 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2cflz" event={"ID":"8e084226-907c-4278-8223-34f6fa78a9a4","Type":"ContainerDied","Data":"58779ac75cbde384a3aa88df750835504c26ee66a6901a033019925eba11d358"} Feb 27 18:00:12 crc kubenswrapper[4700]: I0227 18:00:12.110982 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2cflz" event={"ID":"8e084226-907c-4278-8223-34f6fa78a9a4","Type":"ContainerStarted","Data":"ffe84c0e3d835a0c0d373caf392b80b4ed1a617e3f109e293e535841ff269329"} Feb 27 18:00:13 crc kubenswrapper[4700]: E0227 18:00:13.074785 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 27 18:00:13 crc kubenswrapper[4700]: E0227 18:00:13.075037 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x9n8w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-2cflz_openshift-marketplace(8e084226-907c-4278-8223-34f6fa78a9a4): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 18:00:13 crc kubenswrapper[4700]: E0227 18:00:13.076291 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/certified-operators-2cflz" podUID="8e084226-907c-4278-8223-34f6fa78a9a4" Feb 27 18:00:13 crc kubenswrapper[4700]: E0227 18:00:13.124726 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-2cflz" podUID="8e084226-907c-4278-8223-34f6fa78a9a4" Feb 27 18:00:26 crc kubenswrapper[4700]: E0227 18:00:26.595924 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 27 18:00:26 crc kubenswrapper[4700]: E0227 18:00:26.597067 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x9n8w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-2cflz_openshift-marketplace(8e084226-907c-4278-8223-34f6fa78a9a4): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 18:00:26 crc kubenswrapper[4700]: E0227 18:00:26.598354 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/certified-operators-2cflz" podUID="8e084226-907c-4278-8223-34f6fa78a9a4" Feb 27 18:00:38 crc kubenswrapper[4700]: E0227 18:00:38.984442 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-2cflz" podUID="8e084226-907c-4278-8223-34f6fa78a9a4" Feb 27 18:00:50 crc kubenswrapper[4700]: E0227 18:00:50.536564 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 27 18:00:50 crc kubenswrapper[4700]: E0227 18:00:50.537495 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x9n8w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-2cflz_openshift-marketplace(8e084226-907c-4278-8223-34f6fa78a9a4): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 18:00:50 crc kubenswrapper[4700]: E0227 18:00:50.538797 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/certified-operators-2cflz" podUID="8e084226-907c-4278-8223-34f6fa78a9a4" Feb 27 18:01:00 crc kubenswrapper[4700]: I0227 18:01:00.182257 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29536921-kzv47"] Feb 27 18:01:00 crc kubenswrapper[4700]: I0227 18:01:00.185325 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29536921-kzv47" Feb 27 18:01:00 crc kubenswrapper[4700]: I0227 18:01:00.206279 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29536921-kzv47"] Feb 27 18:01:00 crc kubenswrapper[4700]: I0227 18:01:00.242481 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/167e1521-363c-4a1a-9ff0-67981b7ebf37-fernet-keys\") pod \"keystone-cron-29536921-kzv47\" (UID: \"167e1521-363c-4a1a-9ff0-67981b7ebf37\") " pod="openstack/keystone-cron-29536921-kzv47" Feb 27 18:01:00 crc kubenswrapper[4700]: I0227 18:01:00.242712 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/167e1521-363c-4a1a-9ff0-67981b7ebf37-config-data\") pod \"keystone-cron-29536921-kzv47\" (UID: \"167e1521-363c-4a1a-9ff0-67981b7ebf37\") " pod="openstack/keystone-cron-29536921-kzv47" Feb 27 18:01:00 crc kubenswrapper[4700]: I0227 18:01:00.242749 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwjdt\" (UniqueName: \"kubernetes.io/projected/167e1521-363c-4a1a-9ff0-67981b7ebf37-kube-api-access-fwjdt\") pod \"keystone-cron-29536921-kzv47\" (UID: \"167e1521-363c-4a1a-9ff0-67981b7ebf37\") " pod="openstack/keystone-cron-29536921-kzv47" Feb 27 18:01:00 crc kubenswrapper[4700]: I0227 18:01:00.242835 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/167e1521-363c-4a1a-9ff0-67981b7ebf37-combined-ca-bundle\") pod \"keystone-cron-29536921-kzv47\" (UID: \"167e1521-363c-4a1a-9ff0-67981b7ebf37\") " pod="openstack/keystone-cron-29536921-kzv47" Feb 27 18:01:00 crc kubenswrapper[4700]: I0227 18:01:00.345690 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/167e1521-363c-4a1a-9ff0-67981b7ebf37-config-data\") pod \"keystone-cron-29536921-kzv47\" (UID: \"167e1521-363c-4a1a-9ff0-67981b7ebf37\") " pod="openstack/keystone-cron-29536921-kzv47" Feb 27 18:01:00 crc kubenswrapper[4700]: I0227 18:01:00.345783 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwjdt\" (UniqueName: \"kubernetes.io/projected/167e1521-363c-4a1a-9ff0-67981b7ebf37-kube-api-access-fwjdt\") pod \"keystone-cron-29536921-kzv47\" (UID: \"167e1521-363c-4a1a-9ff0-67981b7ebf37\") " pod="openstack/keystone-cron-29536921-kzv47" Feb 27 18:01:00 crc kubenswrapper[4700]: I0227 18:01:00.345939 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/167e1521-363c-4a1a-9ff0-67981b7ebf37-combined-ca-bundle\") pod \"keystone-cron-29536921-kzv47\" (UID: \"167e1521-363c-4a1a-9ff0-67981b7ebf37\") " pod="openstack/keystone-cron-29536921-kzv47" Feb 27 18:01:00 crc kubenswrapper[4700]: I0227 18:01:00.346056 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/167e1521-363c-4a1a-9ff0-67981b7ebf37-fernet-keys\") pod \"keystone-cron-29536921-kzv47\" (UID: \"167e1521-363c-4a1a-9ff0-67981b7ebf37\") " pod="openstack/keystone-cron-29536921-kzv47" Feb 27 18:01:00 crc kubenswrapper[4700]: I0227 18:01:00.356584 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/167e1521-363c-4a1a-9ff0-67981b7ebf37-config-data\") pod \"keystone-cron-29536921-kzv47\" (UID: \"167e1521-363c-4a1a-9ff0-67981b7ebf37\") " pod="openstack/keystone-cron-29536921-kzv47" Feb 27 18:01:00 crc kubenswrapper[4700]: I0227 18:01:00.359611 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/167e1521-363c-4a1a-9ff0-67981b7ebf37-fernet-keys\") pod \"keystone-cron-29536921-kzv47\" (UID: \"167e1521-363c-4a1a-9ff0-67981b7ebf37\") " pod="openstack/keystone-cron-29536921-kzv47" Feb 27 18:01:00 crc kubenswrapper[4700]: I0227 18:01:00.364403 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/167e1521-363c-4a1a-9ff0-67981b7ebf37-combined-ca-bundle\") pod \"keystone-cron-29536921-kzv47\" (UID: \"167e1521-363c-4a1a-9ff0-67981b7ebf37\") " pod="openstack/keystone-cron-29536921-kzv47" Feb 27 18:01:00 crc kubenswrapper[4700]: I0227 18:01:00.372952 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwjdt\" (UniqueName: \"kubernetes.io/projected/167e1521-363c-4a1a-9ff0-67981b7ebf37-kube-api-access-fwjdt\") pod \"keystone-cron-29536921-kzv47\" (UID: \"167e1521-363c-4a1a-9ff0-67981b7ebf37\") " pod="openstack/keystone-cron-29536921-kzv47" Feb 27 18:01:00 crc kubenswrapper[4700]: I0227 18:01:00.512741 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29536921-kzv47" Feb 27 18:01:01 crc kubenswrapper[4700]: I0227 18:01:01.090969 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29536921-kzv47"] Feb 27 18:01:01 crc kubenswrapper[4700]: I0227 18:01:01.735059 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29536921-kzv47" event={"ID":"167e1521-363c-4a1a-9ff0-67981b7ebf37","Type":"ContainerStarted","Data":"ad2e844454ee2c3e5ce7852664fd7f68b4667aa11fb085079d292b53813f96d3"} Feb 27 18:01:01 crc kubenswrapper[4700]: I0227 18:01:01.735472 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29536921-kzv47" event={"ID":"167e1521-363c-4a1a-9ff0-67981b7ebf37","Type":"ContainerStarted","Data":"d46fb124d012602a39039d245625106759858497933717e784ce253569f91784"} Feb 27 18:01:01 crc kubenswrapper[4700]: I0227 18:01:01.766022 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29536921-kzv47" podStartSLOduration=1.765997789 podStartE2EDuration="1.765997789s" podCreationTimestamp="2026-02-27 18:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 18:01:01.762363875 +0000 UTC m=+3621.747676632" watchObservedRunningTime="2026-02-27 18:01:01.765997789 +0000 UTC m=+3621.751310536" Feb 27 18:01:04 crc kubenswrapper[4700]: I0227 18:01:04.317102 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rzfdh"] Feb 27 18:01:04 crc kubenswrapper[4700]: I0227 18:01:04.320794 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rzfdh" Feb 27 18:01:04 crc kubenswrapper[4700]: I0227 18:01:04.333368 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rzfdh"] Feb 27 18:01:04 crc kubenswrapper[4700]: I0227 18:01:04.398592 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a-catalog-content\") pod \"redhat-marketplace-rzfdh\" (UID: \"8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a\") " pod="openshift-marketplace/redhat-marketplace-rzfdh" Feb 27 18:01:04 crc kubenswrapper[4700]: I0227 18:01:04.398777 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a-utilities\") pod \"redhat-marketplace-rzfdh\" (UID: \"8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a\") " pod="openshift-marketplace/redhat-marketplace-rzfdh" Feb 27 18:01:04 crc kubenswrapper[4700]: I0227 18:01:04.398834 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7wdp\" (UniqueName: \"kubernetes.io/projected/8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a-kube-api-access-r7wdp\") pod \"redhat-marketplace-rzfdh\" (UID: \"8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a\") " pod="openshift-marketplace/redhat-marketplace-rzfdh" Feb 27 18:01:04 crc kubenswrapper[4700]: I0227 18:01:04.501505 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a-utilities\") pod \"redhat-marketplace-rzfdh\" (UID: \"8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a\") " pod="openshift-marketplace/redhat-marketplace-rzfdh" Feb 27 18:01:04 crc kubenswrapper[4700]: I0227 18:01:04.501546 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7wdp\" (UniqueName: \"kubernetes.io/projected/8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a-kube-api-access-r7wdp\") pod \"redhat-marketplace-rzfdh\" (UID: \"8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a\") " pod="openshift-marketplace/redhat-marketplace-rzfdh" Feb 27 18:01:04 crc kubenswrapper[4700]: I0227 18:01:04.501637 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a-catalog-content\") pod \"redhat-marketplace-rzfdh\" (UID: \"8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a\") " pod="openshift-marketplace/redhat-marketplace-rzfdh" Feb 27 18:01:04 crc kubenswrapper[4700]: I0227 18:01:04.502076 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a-catalog-content\") pod \"redhat-marketplace-rzfdh\" (UID: \"8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a\") " pod="openshift-marketplace/redhat-marketplace-rzfdh" Feb 27 18:01:04 crc kubenswrapper[4700]: I0227 18:01:04.502294 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a-utilities\") pod \"redhat-marketplace-rzfdh\" (UID: \"8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a\") " pod="openshift-marketplace/redhat-marketplace-rzfdh" Feb 27 18:01:04 crc kubenswrapper[4700]: I0227 18:01:04.530614 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7wdp\" (UniqueName: \"kubernetes.io/projected/8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a-kube-api-access-r7wdp\") pod \"redhat-marketplace-rzfdh\" (UID: \"8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a\") " pod="openshift-marketplace/redhat-marketplace-rzfdh" Feb 27 18:01:04 crc kubenswrapper[4700]: I0227 18:01:04.647662 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rzfdh" Feb 27 18:01:05 crc kubenswrapper[4700]: I0227 18:01:05.195783 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rzfdh"] Feb 27 18:01:05 crc kubenswrapper[4700]: I0227 18:01:05.805416 4700 generic.go:334] "Generic (PLEG): container finished" podID="8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a" containerID="287a4607689af65ee9be255fe3c8b3877b24e028c3c21a19eceb1649c3a3fd62" exitCode=0 Feb 27 18:01:05 crc kubenswrapper[4700]: I0227 18:01:05.805603 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rzfdh" event={"ID":"8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a","Type":"ContainerDied","Data":"287a4607689af65ee9be255fe3c8b3877b24e028c3c21a19eceb1649c3a3fd62"} Feb 27 18:01:05 crc kubenswrapper[4700]: I0227 18:01:05.805809 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rzfdh" event={"ID":"8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a","Type":"ContainerStarted","Data":"e45808127e2ef0f70858240603a802c83876db17dbda17feee3c2dc7e8a2f954"} Feb 27 18:01:05 crc kubenswrapper[4700]: I0227 18:01:05.808647 4700 generic.go:334] "Generic (PLEG): container finished" podID="167e1521-363c-4a1a-9ff0-67981b7ebf37" containerID="ad2e844454ee2c3e5ce7852664fd7f68b4667aa11fb085079d292b53813f96d3" exitCode=0 Feb 27 18:01:05 crc kubenswrapper[4700]: I0227 18:01:05.808720 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29536921-kzv47" event={"ID":"167e1521-363c-4a1a-9ff0-67981b7ebf37","Type":"ContainerDied","Data":"ad2e844454ee2c3e5ce7852664fd7f68b4667aa11fb085079d292b53813f96d3"} Feb 27 18:01:05 crc kubenswrapper[4700]: E0227 18:01:05.983394 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-2cflz" podUID="8e084226-907c-4278-8223-34f6fa78a9a4" Feb 27 18:01:06 crc kubenswrapper[4700]: E0227 18:01:06.529709 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 18:01:06 crc kubenswrapper[4700]: E0227 18:01:06.529869 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r7wdp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-rzfdh_openshift-marketplace(8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 18:01:06 crc kubenswrapper[4700]: E0227 18:01:06.531190 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-marketplace-rzfdh" podUID="8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a" Feb 27 18:01:06 crc kubenswrapper[4700]: E0227 18:01:06.820303 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-rzfdh" podUID="8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a" Feb 27 18:01:07 crc kubenswrapper[4700]: I0227 18:01:07.215018 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29536921-kzv47" Feb 27 18:01:07 crc kubenswrapper[4700]: I0227 18:01:07.273411 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/167e1521-363c-4a1a-9ff0-67981b7ebf37-config-data\") pod \"167e1521-363c-4a1a-9ff0-67981b7ebf37\" (UID: \"167e1521-363c-4a1a-9ff0-67981b7ebf37\") " Feb 27 18:01:07 crc kubenswrapper[4700]: I0227 18:01:07.273679 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/167e1521-363c-4a1a-9ff0-67981b7ebf37-fernet-keys\") pod \"167e1521-363c-4a1a-9ff0-67981b7ebf37\" (UID: \"167e1521-363c-4a1a-9ff0-67981b7ebf37\") " Feb 27 18:01:07 crc kubenswrapper[4700]: I0227 18:01:07.273721 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwjdt\" (UniqueName: \"kubernetes.io/projected/167e1521-363c-4a1a-9ff0-67981b7ebf37-kube-api-access-fwjdt\") pod \"167e1521-363c-4a1a-9ff0-67981b7ebf37\" (UID: \"167e1521-363c-4a1a-9ff0-67981b7ebf37\") " Feb 27 18:01:07 crc kubenswrapper[4700]: I0227 18:01:07.273880 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/167e1521-363c-4a1a-9ff0-67981b7ebf37-combined-ca-bundle\") pod \"167e1521-363c-4a1a-9ff0-67981b7ebf37\" (UID: \"167e1521-363c-4a1a-9ff0-67981b7ebf37\") " Feb 27 18:01:07 crc kubenswrapper[4700]: I0227 18:01:07.282581 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/167e1521-363c-4a1a-9ff0-67981b7ebf37-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "167e1521-363c-4a1a-9ff0-67981b7ebf37" (UID: "167e1521-363c-4a1a-9ff0-67981b7ebf37"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 18:01:07 crc kubenswrapper[4700]: I0227 18:01:07.290211 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/167e1521-363c-4a1a-9ff0-67981b7ebf37-kube-api-access-fwjdt" (OuterVolumeSpecName: "kube-api-access-fwjdt") pod "167e1521-363c-4a1a-9ff0-67981b7ebf37" (UID: "167e1521-363c-4a1a-9ff0-67981b7ebf37"). InnerVolumeSpecName "kube-api-access-fwjdt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:01:07 crc kubenswrapper[4700]: I0227 18:01:07.318404 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/167e1521-363c-4a1a-9ff0-67981b7ebf37-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "167e1521-363c-4a1a-9ff0-67981b7ebf37" (UID: "167e1521-363c-4a1a-9ff0-67981b7ebf37"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 18:01:07 crc kubenswrapper[4700]: I0227 18:01:07.376228 4700 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/167e1521-363c-4a1a-9ff0-67981b7ebf37-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 27 18:01:07 crc kubenswrapper[4700]: I0227 18:01:07.376260 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwjdt\" (UniqueName: \"kubernetes.io/projected/167e1521-363c-4a1a-9ff0-67981b7ebf37-kube-api-access-fwjdt\") on node \"crc\" DevicePath \"\"" Feb 27 18:01:07 crc kubenswrapper[4700]: I0227 18:01:07.376270 4700 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/167e1521-363c-4a1a-9ff0-67981b7ebf37-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 18:01:07 crc kubenswrapper[4700]: I0227 18:01:07.379831 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/167e1521-363c-4a1a-9ff0-67981b7ebf37-config-data" (OuterVolumeSpecName: "config-data") pod "167e1521-363c-4a1a-9ff0-67981b7ebf37" (UID: "167e1521-363c-4a1a-9ff0-67981b7ebf37"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 18:01:07 crc kubenswrapper[4700]: I0227 18:01:07.481070 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/167e1521-363c-4a1a-9ff0-67981b7ebf37-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 18:01:07 crc kubenswrapper[4700]: I0227 18:01:07.830072 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29536921-kzv47" event={"ID":"167e1521-363c-4a1a-9ff0-67981b7ebf37","Type":"ContainerDied","Data":"d46fb124d012602a39039d245625106759858497933717e784ce253569f91784"} Feb 27 18:01:07 crc kubenswrapper[4700]: I0227 18:01:07.830119 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d46fb124d012602a39039d245625106759858497933717e784ce253569f91784" Feb 27 18:01:07 crc kubenswrapper[4700]: I0227 18:01:07.830171 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29536921-kzv47" Feb 27 18:01:17 crc kubenswrapper[4700]: E0227 18:01:17.984851 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-2cflz" podUID="8e084226-907c-4278-8223-34f6fa78a9a4" Feb 27 18:01:28 crc kubenswrapper[4700]: E0227 18:01:28.985117 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-2cflz" podUID="8e084226-907c-4278-8223-34f6fa78a9a4" Feb 27 18:01:36 crc kubenswrapper[4700]: I0227 18:01:36.411261 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 18:01:36 crc kubenswrapper[4700]: I0227 18:01:36.412237 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 18:01:40 crc kubenswrapper[4700]: I0227 18:01:40.260192 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rzfdh" event={"ID":"8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a","Type":"ContainerStarted","Data":"a5c820da9a7386c22c5b0b3a458c50f0b75f94a377b3af61bc265cd7643f5454"} Feb 27 18:01:41 crc kubenswrapper[4700]: I0227 18:01:41.280844 4700 generic.go:334] "Generic (PLEG): container finished" podID="8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a" containerID="a5c820da9a7386c22c5b0b3a458c50f0b75f94a377b3af61bc265cd7643f5454" exitCode=0 Feb 27 18:01:41 crc kubenswrapper[4700]: I0227 18:01:41.280989 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rzfdh" event={"ID":"8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a","Type":"ContainerDied","Data":"a5c820da9a7386c22c5b0b3a458c50f0b75f94a377b3af61bc265cd7643f5454"} Feb 27 18:01:42 crc kubenswrapper[4700]: I0227 18:01:42.293703 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rzfdh" event={"ID":"8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a","Type":"ContainerStarted","Data":"d68fb1880b79c2d24f7d6775c3307dd2d5fd3b734ffe728b72d1646ce1a8adcd"} Feb 27 18:01:42 crc kubenswrapper[4700]: I0227 18:01:42.324166 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rzfdh" podStartSLOduration=2.416411604 podStartE2EDuration="38.324143539s" podCreationTimestamp="2026-02-27 18:01:04 +0000 UTC" firstStartedPulling="2026-02-27 18:01:05.808007065 +0000 UTC m=+3625.793319812" lastFinishedPulling="2026-02-27 18:01:41.715739 +0000 UTC m=+3661.701051747" observedRunningTime="2026-02-27 18:01:42.321612753 +0000 UTC m=+3662.306925570" watchObservedRunningTime="2026-02-27 18:01:42.324143539 +0000 UTC m=+3662.309456296" Feb 27 18:01:44 crc kubenswrapper[4700]: I0227 18:01:44.648646 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rzfdh" Feb 27 18:01:44 crc kubenswrapper[4700]: I0227 18:01:44.649246 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rzfdh" Feb 27 18:01:44 crc kubenswrapper[4700]: I0227 18:01:44.734033 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rzfdh" Feb 27 18:01:46 crc kubenswrapper[4700]: I0227 18:01:46.345133 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2cflz" event={"ID":"8e084226-907c-4278-8223-34f6fa78a9a4","Type":"ContainerStarted","Data":"169de72005f109b56d59b7748200069d120249fe2d134bc3d3582ca83d798441"} Feb 27 18:01:49 crc kubenswrapper[4700]: I0227 18:01:49.365408 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jks9x"] Feb 27 18:01:49 crc kubenswrapper[4700]: E0227 18:01:49.366556 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="167e1521-363c-4a1a-9ff0-67981b7ebf37" containerName="keystone-cron" Feb 27 18:01:49 crc kubenswrapper[4700]: I0227 18:01:49.366580 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="167e1521-363c-4a1a-9ff0-67981b7ebf37" containerName="keystone-cron" Feb 27 18:01:49 crc kubenswrapper[4700]: I0227 18:01:49.366974 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="167e1521-363c-4a1a-9ff0-67981b7ebf37" containerName="keystone-cron" Feb 27 18:01:49 crc kubenswrapper[4700]: I0227 18:01:49.377365 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jks9x" Feb 27 18:01:49 crc kubenswrapper[4700]: I0227 18:01:49.387430 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jks9x"] Feb 27 18:01:49 crc kubenswrapper[4700]: I0227 18:01:49.405132 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6jmm\" (UniqueName: \"kubernetes.io/projected/e14e41da-6e02-4d9f-9faf-68789c1dbb9e-kube-api-access-j6jmm\") pod \"community-operators-jks9x\" (UID: \"e14e41da-6e02-4d9f-9faf-68789c1dbb9e\") " pod="openshift-marketplace/community-operators-jks9x" Feb 27 18:01:49 crc kubenswrapper[4700]: I0227 18:01:49.405265 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e14e41da-6e02-4d9f-9faf-68789c1dbb9e-utilities\") pod \"community-operators-jks9x\" (UID: \"e14e41da-6e02-4d9f-9faf-68789c1dbb9e\") " pod="openshift-marketplace/community-operators-jks9x" Feb 27 18:01:49 crc kubenswrapper[4700]: I0227 18:01:49.405455 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e14e41da-6e02-4d9f-9faf-68789c1dbb9e-catalog-content\") pod \"community-operators-jks9x\" (UID: \"e14e41da-6e02-4d9f-9faf-68789c1dbb9e\") " pod="openshift-marketplace/community-operators-jks9x" Feb 27 18:01:49 crc kubenswrapper[4700]: I0227 18:01:49.508134 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e14e41da-6e02-4d9f-9faf-68789c1dbb9e-catalog-content\") pod \"community-operators-jks9x\" (UID: \"e14e41da-6e02-4d9f-9faf-68789c1dbb9e\") " pod="openshift-marketplace/community-operators-jks9x" Feb 27 18:01:49 crc kubenswrapper[4700]: I0227 18:01:49.508325 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6jmm\" (UniqueName: \"kubernetes.io/projected/e14e41da-6e02-4d9f-9faf-68789c1dbb9e-kube-api-access-j6jmm\") pod \"community-operators-jks9x\" (UID: \"e14e41da-6e02-4d9f-9faf-68789c1dbb9e\") " pod="openshift-marketplace/community-operators-jks9x" Feb 27 18:01:49 crc kubenswrapper[4700]: I0227 18:01:49.508364 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e14e41da-6e02-4d9f-9faf-68789c1dbb9e-utilities\") pod \"community-operators-jks9x\" (UID: \"e14e41da-6e02-4d9f-9faf-68789c1dbb9e\") " pod="openshift-marketplace/community-operators-jks9x" Feb 27 18:01:49 crc kubenswrapper[4700]: I0227 18:01:49.508538 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e14e41da-6e02-4d9f-9faf-68789c1dbb9e-catalog-content\") pod \"community-operators-jks9x\" (UID: \"e14e41da-6e02-4d9f-9faf-68789c1dbb9e\") " pod="openshift-marketplace/community-operators-jks9x" Feb 27 18:01:49 crc kubenswrapper[4700]: I0227 18:01:49.508804 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e14e41da-6e02-4d9f-9faf-68789c1dbb9e-utilities\") pod \"community-operators-jks9x\" (UID: \"e14e41da-6e02-4d9f-9faf-68789c1dbb9e\") " pod="openshift-marketplace/community-operators-jks9x" Feb 27 18:01:49 crc kubenswrapper[4700]: I0227 18:01:49.530643 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6jmm\" (UniqueName: \"kubernetes.io/projected/e14e41da-6e02-4d9f-9faf-68789c1dbb9e-kube-api-access-j6jmm\") pod \"community-operators-jks9x\" (UID: \"e14e41da-6e02-4d9f-9faf-68789c1dbb9e\") " pod="openshift-marketplace/community-operators-jks9x" Feb 27 18:01:49 crc kubenswrapper[4700]: I0227 18:01:49.751144 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jks9x" Feb 27 18:01:50 crc kubenswrapper[4700]: I0227 18:01:50.275640 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jks9x"] Feb 27 18:01:50 crc kubenswrapper[4700]: I0227 18:01:50.404728 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jks9x" event={"ID":"e14e41da-6e02-4d9f-9faf-68789c1dbb9e","Type":"ContainerStarted","Data":"a7a180c328babda376860297fb4df32e9cd6c984cf630d2992630c892fde6e4d"} Feb 27 18:01:50 crc kubenswrapper[4700]: I0227 18:01:50.408820 4700 generic.go:334] "Generic (PLEG): container finished" podID="8e084226-907c-4278-8223-34f6fa78a9a4" containerID="169de72005f109b56d59b7748200069d120249fe2d134bc3d3582ca83d798441" exitCode=0 Feb 27 18:01:50 crc kubenswrapper[4700]: I0227 18:01:50.408875 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2cflz" event={"ID":"8e084226-907c-4278-8223-34f6fa78a9a4","Type":"ContainerDied","Data":"169de72005f109b56d59b7748200069d120249fe2d134bc3d3582ca83d798441"} Feb 27 18:01:51 crc kubenswrapper[4700]: I0227 18:01:51.421341 4700 generic.go:334] "Generic (PLEG): container finished" podID="e14e41da-6e02-4d9f-9faf-68789c1dbb9e" containerID="6d782c33eeb65e1278cf57a8b3e2d70a65e05abfd37f8dc16f9f10a5b13d5060" exitCode=0 Feb 27 18:01:51 crc kubenswrapper[4700]: I0227 18:01:51.422029 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jks9x" event={"ID":"e14e41da-6e02-4d9f-9faf-68789c1dbb9e","Type":"ContainerDied","Data":"6d782c33eeb65e1278cf57a8b3e2d70a65e05abfd37f8dc16f9f10a5b13d5060"} Feb 27 18:01:51 crc kubenswrapper[4700]: I0227 18:01:51.427109 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2cflz" event={"ID":"8e084226-907c-4278-8223-34f6fa78a9a4","Type":"ContainerStarted","Data":"b0eff5189f682c698b28377e71ace37538c9f60a800799ed7c57fe0f5a4b5ab7"} Feb 27 18:01:51 crc kubenswrapper[4700]: I0227 18:01:51.474127 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2cflz" podStartSLOduration=2.735251457 podStartE2EDuration="1m41.474107741s" podCreationTimestamp="2026-02-27 18:00:10 +0000 UTC" firstStartedPulling="2026-02-27 18:00:12.125160365 +0000 UTC m=+3572.110473112" lastFinishedPulling="2026-02-27 18:01:50.864016639 +0000 UTC m=+3670.849329396" observedRunningTime="2026-02-27 18:01:51.464387948 +0000 UTC m=+3671.449700695" watchObservedRunningTime="2026-02-27 18:01:51.474107741 +0000 UTC m=+3671.459420488" Feb 27 18:01:53 crc kubenswrapper[4700]: I0227 18:01:53.451900 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jks9x" event={"ID":"e14e41da-6e02-4d9f-9faf-68789c1dbb9e","Type":"ContainerStarted","Data":"43618a3b775d62bd156d2ce6324815a7cb34457216f134b7947655e25c421cb6"} Feb 27 18:01:54 crc kubenswrapper[4700]: I0227 18:01:54.732982 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rzfdh" Feb 27 18:01:55 crc kubenswrapper[4700]: I0227 18:01:55.475685 4700 generic.go:334] "Generic (PLEG): container finished" podID="e14e41da-6e02-4d9f-9faf-68789c1dbb9e" containerID="43618a3b775d62bd156d2ce6324815a7cb34457216f134b7947655e25c421cb6" exitCode=0 Feb 27 18:01:55 crc kubenswrapper[4700]: I0227 18:01:55.475779 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jks9x" event={"ID":"e14e41da-6e02-4d9f-9faf-68789c1dbb9e","Type":"ContainerDied","Data":"43618a3b775d62bd156d2ce6324815a7cb34457216f134b7947655e25c421cb6"} Feb 27 18:01:56 crc kubenswrapper[4700]: I0227 18:01:56.487322 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jks9x" event={"ID":"e14e41da-6e02-4d9f-9faf-68789c1dbb9e","Type":"ContainerStarted","Data":"9f8636e8557d4162b97b500f76725cb8f53a7a8adbd47ad9e2c382b551517fae"} Feb 27 18:01:56 crc kubenswrapper[4700]: I0227 18:01:56.520219 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jks9x" podStartSLOduration=3.024381883 podStartE2EDuration="7.520194255s" podCreationTimestamp="2026-02-27 18:01:49 +0000 UTC" firstStartedPulling="2026-02-27 18:01:51.423723137 +0000 UTC m=+3671.409035884" lastFinishedPulling="2026-02-27 18:01:55.919535519 +0000 UTC m=+3675.904848256" observedRunningTime="2026-02-27 18:01:56.508950112 +0000 UTC m=+3676.494262869" watchObservedRunningTime="2026-02-27 18:01:56.520194255 +0000 UTC m=+3676.505507002" Feb 27 18:01:57 crc kubenswrapper[4700]: I0227 18:01:57.111850 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rzfdh"] Feb 27 18:01:57 crc kubenswrapper[4700]: I0227 18:01:57.112370 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rzfdh" podUID="8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a" containerName="registry-server" containerID="cri-o://d68fb1880b79c2d24f7d6775c3307dd2d5fd3b734ffe728b72d1646ce1a8adcd" gracePeriod=2 Feb 27 18:01:57 crc kubenswrapper[4700]: I0227 18:01:57.502329 4700 generic.go:334] "Generic (PLEG): container finished" podID="8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a" containerID="d68fb1880b79c2d24f7d6775c3307dd2d5fd3b734ffe728b72d1646ce1a8adcd" exitCode=0 Feb 27 18:01:57 crc kubenswrapper[4700]: I0227 18:01:57.502544 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rzfdh" event={"ID":"8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a","Type":"ContainerDied","Data":"d68fb1880b79c2d24f7d6775c3307dd2d5fd3b734ffe728b72d1646ce1a8adcd"} Feb 27 18:01:57 crc kubenswrapper[4700]: I0227 18:01:57.613767 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rzfdh" Feb 27 18:01:57 crc kubenswrapper[4700]: I0227 18:01:57.791589 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a-utilities\") pod \"8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a\" (UID: \"8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a\") " Feb 27 18:01:57 crc kubenswrapper[4700]: I0227 18:01:57.791694 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7wdp\" (UniqueName: \"kubernetes.io/projected/8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a-kube-api-access-r7wdp\") pod \"8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a\" (UID: \"8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a\") " Feb 27 18:01:57 crc kubenswrapper[4700]: I0227 18:01:57.791847 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a-catalog-content\") pod \"8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a\" (UID: \"8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a\") " Feb 27 18:01:57 crc kubenswrapper[4700]: I0227 18:01:57.794037 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a-utilities" (OuterVolumeSpecName: "utilities") pod "8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a" (UID: "8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 18:01:57 crc kubenswrapper[4700]: I0227 18:01:57.799004 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a-kube-api-access-r7wdp" (OuterVolumeSpecName: "kube-api-access-r7wdp") pod "8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a" (UID: "8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a"). InnerVolumeSpecName "kube-api-access-r7wdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:01:57 crc kubenswrapper[4700]: I0227 18:01:57.823784 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a" (UID: "8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 18:01:57 crc kubenswrapper[4700]: I0227 18:01:57.896264 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 18:01:57 crc kubenswrapper[4700]: I0227 18:01:57.896299 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7wdp\" (UniqueName: \"kubernetes.io/projected/8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a-kube-api-access-r7wdp\") on node \"crc\" DevicePath \"\"" Feb 27 18:01:57 crc kubenswrapper[4700]: I0227 18:01:57.896311 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 18:01:58 crc kubenswrapper[4700]: I0227 18:01:58.518601 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rzfdh" event={"ID":"8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a","Type":"ContainerDied","Data":"e45808127e2ef0f70858240603a802c83876db17dbda17feee3c2dc7e8a2f954"} Feb 27 18:01:58 crc kubenswrapper[4700]: I0227 18:01:58.518982 4700 scope.go:117] "RemoveContainer" containerID="d68fb1880b79c2d24f7d6775c3307dd2d5fd3b734ffe728b72d1646ce1a8adcd" Feb 27 18:01:58 crc kubenswrapper[4700]: I0227 18:01:58.518906 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rzfdh" Feb 27 18:01:58 crc kubenswrapper[4700]: I0227 18:01:58.546213 4700 scope.go:117] "RemoveContainer" containerID="a5c820da9a7386c22c5b0b3a458c50f0b75f94a377b3af61bc265cd7643f5454" Feb 27 18:01:58 crc kubenswrapper[4700]: I0227 18:01:58.572548 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rzfdh"] Feb 27 18:01:58 crc kubenswrapper[4700]: I0227 18:01:58.591812 4700 scope.go:117] "RemoveContainer" containerID="287a4607689af65ee9be255fe3c8b3877b24e028c3c21a19eceb1649c3a3fd62" Feb 27 18:01:58 crc kubenswrapper[4700]: I0227 18:01:58.597141 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rzfdh"] Feb 27 18:01:58 crc kubenswrapper[4700]: I0227 18:01:58.991196 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a" path="/var/lib/kubelet/pods/8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a/volumes" Feb 27 18:01:59 crc kubenswrapper[4700]: I0227 18:01:59.751285 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jks9x" Feb 27 18:01:59 crc kubenswrapper[4700]: I0227 18:01:59.751621 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jks9x" Feb 27 18:01:59 crc kubenswrapper[4700]: I0227 18:01:59.831326 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jks9x" Feb 27 18:02:00 crc kubenswrapper[4700]: I0227 18:02:00.173893 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536922-d7lzf"] Feb 27 18:02:00 crc kubenswrapper[4700]: E0227 18:02:00.175099 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a" containerName="extract-content" Feb 27 18:02:00 crc kubenswrapper[4700]: I0227 18:02:00.175138 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a" containerName="extract-content" Feb 27 18:02:00 crc kubenswrapper[4700]: E0227 18:02:00.175201 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a" containerName="extract-utilities" Feb 27 18:02:00 crc kubenswrapper[4700]: I0227 18:02:00.175216 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a" containerName="extract-utilities" Feb 27 18:02:00 crc kubenswrapper[4700]: E0227 18:02:00.175264 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a" containerName="registry-server" Feb 27 18:02:00 crc kubenswrapper[4700]: I0227 18:02:00.175285 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a" containerName="registry-server" Feb 27 18:02:00 crc kubenswrapper[4700]: I0227 18:02:00.180069 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d6c2a5a-4d7b-4ec3-81b9-1bea5b61dc2a" containerName="registry-server" Feb 27 18:02:00 crc kubenswrapper[4700]: I0227 18:02:00.182048 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536922-d7lzf" Feb 27 18:02:00 crc kubenswrapper[4700]: I0227 18:02:00.188035 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d62kc" Feb 27 18:02:00 crc kubenswrapper[4700]: I0227 18:02:00.188202 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 18:02:00 crc kubenswrapper[4700]: I0227 18:02:00.188274 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 18:02:00 crc kubenswrapper[4700]: I0227 18:02:00.193709 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536922-d7lzf"] Feb 27 18:02:00 crc kubenswrapper[4700]: I0227 18:02:00.349303 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7vxj\" (UniqueName: \"kubernetes.io/projected/793f7988-e7d0-4b3b-a35f-7d0445ebab61-kube-api-access-d7vxj\") pod \"auto-csr-approver-29536922-d7lzf\" (UID: \"793f7988-e7d0-4b3b-a35f-7d0445ebab61\") " pod="openshift-infra/auto-csr-approver-29536922-d7lzf" Feb 27 18:02:00 crc kubenswrapper[4700]: I0227 18:02:00.451915 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7vxj\" (UniqueName: \"kubernetes.io/projected/793f7988-e7d0-4b3b-a35f-7d0445ebab61-kube-api-access-d7vxj\") pod \"auto-csr-approver-29536922-d7lzf\" (UID: \"793f7988-e7d0-4b3b-a35f-7d0445ebab61\") " pod="openshift-infra/auto-csr-approver-29536922-d7lzf" Feb 27 18:02:00 crc kubenswrapper[4700]: I0227 18:02:00.477071 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7vxj\" (UniqueName: \"kubernetes.io/projected/793f7988-e7d0-4b3b-a35f-7d0445ebab61-kube-api-access-d7vxj\") pod \"auto-csr-approver-29536922-d7lzf\" (UID: \"793f7988-e7d0-4b3b-a35f-7d0445ebab61\") " pod="openshift-infra/auto-csr-approver-29536922-d7lzf" Feb 27 18:02:00 crc kubenswrapper[4700]: I0227 18:02:00.526148 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536922-d7lzf" Feb 27 18:02:00 crc kubenswrapper[4700]: I0227 18:02:00.571969 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2cflz" Feb 27 18:02:00 crc kubenswrapper[4700]: I0227 18:02:00.572753 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2cflz" Feb 27 18:02:01 crc kubenswrapper[4700]: W0227 18:02:01.017660 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod793f7988_e7d0_4b3b_a35f_7d0445ebab61.slice/crio-a5849d0e4ab42c85e16ad82c41942984f6c41e0309fb72fc6c786f02414e5030 WatchSource:0}: Error finding container a5849d0e4ab42c85e16ad82c41942984f6c41e0309fb72fc6c786f02414e5030: Status 404 returned error can't find the container with id a5849d0e4ab42c85e16ad82c41942984f6c41e0309fb72fc6c786f02414e5030 Feb 27 18:02:01 crc kubenswrapper[4700]: I0227 18:02:01.024909 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536922-d7lzf"] Feb 27 18:02:01 crc kubenswrapper[4700]: I0227 18:02:01.555109 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536922-d7lzf" event={"ID":"793f7988-e7d0-4b3b-a35f-7d0445ebab61","Type":"ContainerStarted","Data":"a5849d0e4ab42c85e16ad82c41942984f6c41e0309fb72fc6c786f02414e5030"} Feb 27 18:02:01 crc kubenswrapper[4700]: I0227 18:02:01.661176 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-2cflz" podUID="8e084226-907c-4278-8223-34f6fa78a9a4" containerName="registry-server" probeResult="failure" output=< Feb 27 18:02:01 crc kubenswrapper[4700]: timeout: failed to connect service ":50051" within 1s Feb 27 18:02:01 crc kubenswrapper[4700]: > Feb 27 18:02:02 crc kubenswrapper[4700]: E0227 18:02:02.303391 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 18:02:02 crc kubenswrapper[4700]: E0227 18:02:02.303780 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 18:02:02 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 18:02:02 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-d7vxj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536922-d7lzf_openshift-infra(793f7988-e7d0-4b3b-a35f-7d0445ebab61): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 18:02:02 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 18:02:02 crc kubenswrapper[4700]: E0227 18:02:02.304928 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536922-d7lzf" podUID="793f7988-e7d0-4b3b-a35f-7d0445ebab61" Feb 27 18:02:02 crc kubenswrapper[4700]: E0227 18:02:02.566239 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536922-d7lzf" podUID="793f7988-e7d0-4b3b-a35f-7d0445ebab61" Feb 27 18:02:06 crc kubenswrapper[4700]: I0227 18:02:06.410750 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 18:02:06 crc kubenswrapper[4700]: I0227 18:02:06.412902 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 18:02:09 crc kubenswrapper[4700]: I0227 18:02:09.833430 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jks9x" Feb 27 18:02:09 crc kubenswrapper[4700]: I0227 18:02:09.908098 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jks9x"] Feb 27 18:02:10 crc kubenswrapper[4700]: I0227 18:02:10.630796 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2cflz" Feb 27 18:02:10 crc kubenswrapper[4700]: I0227 18:02:10.652002 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jks9x" podUID="e14e41da-6e02-4d9f-9faf-68789c1dbb9e" containerName="registry-server" containerID="cri-o://9f8636e8557d4162b97b500f76725cb8f53a7a8adbd47ad9e2c382b551517fae" gracePeriod=2 Feb 27 18:02:10 crc kubenswrapper[4700]: I0227 18:02:10.697143 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2cflz" Feb 27 18:02:11 crc kubenswrapper[4700]: I0227 18:02:11.169879 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jks9x" Feb 27 18:02:11 crc kubenswrapper[4700]: I0227 18:02:11.323724 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j6jmm\" (UniqueName: \"kubernetes.io/projected/e14e41da-6e02-4d9f-9faf-68789c1dbb9e-kube-api-access-j6jmm\") pod \"e14e41da-6e02-4d9f-9faf-68789c1dbb9e\" (UID: \"e14e41da-6e02-4d9f-9faf-68789c1dbb9e\") " Feb 27 18:02:11 crc kubenswrapper[4700]: I0227 18:02:11.323908 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e14e41da-6e02-4d9f-9faf-68789c1dbb9e-catalog-content\") pod \"e14e41da-6e02-4d9f-9faf-68789c1dbb9e\" (UID: \"e14e41da-6e02-4d9f-9faf-68789c1dbb9e\") " Feb 27 18:02:11 crc kubenswrapper[4700]: I0227 18:02:11.324024 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e14e41da-6e02-4d9f-9faf-68789c1dbb9e-utilities\") pod \"e14e41da-6e02-4d9f-9faf-68789c1dbb9e\" (UID: \"e14e41da-6e02-4d9f-9faf-68789c1dbb9e\") " Feb 27 18:02:11 crc kubenswrapper[4700]: I0227 18:02:11.326395 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e14e41da-6e02-4d9f-9faf-68789c1dbb9e-utilities" (OuterVolumeSpecName: "utilities") pod "e14e41da-6e02-4d9f-9faf-68789c1dbb9e" (UID: "e14e41da-6e02-4d9f-9faf-68789c1dbb9e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 18:02:11 crc kubenswrapper[4700]: I0227 18:02:11.334795 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e14e41da-6e02-4d9f-9faf-68789c1dbb9e-kube-api-access-j6jmm" (OuterVolumeSpecName: "kube-api-access-j6jmm") pod "e14e41da-6e02-4d9f-9faf-68789c1dbb9e" (UID: "e14e41da-6e02-4d9f-9faf-68789c1dbb9e"). InnerVolumeSpecName "kube-api-access-j6jmm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:02:11 crc kubenswrapper[4700]: I0227 18:02:11.403525 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e14e41da-6e02-4d9f-9faf-68789c1dbb9e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e14e41da-6e02-4d9f-9faf-68789c1dbb9e" (UID: "e14e41da-6e02-4d9f-9faf-68789c1dbb9e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 18:02:11 crc kubenswrapper[4700]: I0227 18:02:11.426790 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e14e41da-6e02-4d9f-9faf-68789c1dbb9e-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 18:02:11 crc kubenswrapper[4700]: I0227 18:02:11.426851 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j6jmm\" (UniqueName: \"kubernetes.io/projected/e14e41da-6e02-4d9f-9faf-68789c1dbb9e-kube-api-access-j6jmm\") on node \"crc\" DevicePath \"\"" Feb 27 18:02:11 crc kubenswrapper[4700]: I0227 18:02:11.426867 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e14e41da-6e02-4d9f-9faf-68789c1dbb9e-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 18:02:11 crc kubenswrapper[4700]: I0227 18:02:11.665763 4700 generic.go:334] "Generic (PLEG): container finished" podID="e14e41da-6e02-4d9f-9faf-68789c1dbb9e" containerID="9f8636e8557d4162b97b500f76725cb8f53a7a8adbd47ad9e2c382b551517fae" exitCode=0 Feb 27 18:02:11 crc kubenswrapper[4700]: I0227 18:02:11.665805 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jks9x" event={"ID":"e14e41da-6e02-4d9f-9faf-68789c1dbb9e","Type":"ContainerDied","Data":"9f8636e8557d4162b97b500f76725cb8f53a7a8adbd47ad9e2c382b551517fae"} Feb 27 18:02:11 crc kubenswrapper[4700]: I0227 18:02:11.665837 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jks9x" event={"ID":"e14e41da-6e02-4d9f-9faf-68789c1dbb9e","Type":"ContainerDied","Data":"a7a180c328babda376860297fb4df32e9cd6c984cf630d2992630c892fde6e4d"} Feb 27 18:02:11 crc kubenswrapper[4700]: I0227 18:02:11.665841 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jks9x" Feb 27 18:02:11 crc kubenswrapper[4700]: I0227 18:02:11.665855 4700 scope.go:117] "RemoveContainer" containerID="9f8636e8557d4162b97b500f76725cb8f53a7a8adbd47ad9e2c382b551517fae" Feb 27 18:02:11 crc kubenswrapper[4700]: I0227 18:02:11.711142 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2cflz"] Feb 27 18:02:11 crc kubenswrapper[4700]: I0227 18:02:11.711369 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2cflz" podUID="8e084226-907c-4278-8223-34f6fa78a9a4" containerName="registry-server" containerID="cri-o://b0eff5189f682c698b28377e71ace37538c9f60a800799ed7c57fe0f5a4b5ab7" gracePeriod=2 Feb 27 18:02:11 crc kubenswrapper[4700]: I0227 18:02:11.713166 4700 scope.go:117] "RemoveContainer" containerID="43618a3b775d62bd156d2ce6324815a7cb34457216f134b7947655e25c421cb6" Feb 27 18:02:11 crc kubenswrapper[4700]: I0227 18:02:11.725612 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jks9x"] Feb 27 18:02:11 crc kubenswrapper[4700]: I0227 18:02:11.737868 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jks9x"] Feb 27 18:02:11 crc kubenswrapper[4700]: I0227 18:02:11.785773 4700 scope.go:117] "RemoveContainer" containerID="6d782c33eeb65e1278cf57a8b3e2d70a65e05abfd37f8dc16f9f10a5b13d5060" Feb 27 18:02:11 crc kubenswrapper[4700]: E0227 18:02:11.831206 4700 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode14e41da_6e02_4d9f_9faf_68789c1dbb9e.slice\": RecentStats: unable to find data in memory cache]" Feb 27 18:02:11 crc kubenswrapper[4700]: I0227 18:02:11.945190 4700 scope.go:117] "RemoveContainer" containerID="9f8636e8557d4162b97b500f76725cb8f53a7a8adbd47ad9e2c382b551517fae" Feb 27 18:02:11 crc kubenswrapper[4700]: E0227 18:02:11.947843 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f8636e8557d4162b97b500f76725cb8f53a7a8adbd47ad9e2c382b551517fae\": container with ID starting with 9f8636e8557d4162b97b500f76725cb8f53a7a8adbd47ad9e2c382b551517fae not found: ID does not exist" containerID="9f8636e8557d4162b97b500f76725cb8f53a7a8adbd47ad9e2c382b551517fae" Feb 27 18:02:11 crc kubenswrapper[4700]: I0227 18:02:11.947896 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f8636e8557d4162b97b500f76725cb8f53a7a8adbd47ad9e2c382b551517fae"} err="failed to get container status \"9f8636e8557d4162b97b500f76725cb8f53a7a8adbd47ad9e2c382b551517fae\": rpc error: code = NotFound desc = could not find container \"9f8636e8557d4162b97b500f76725cb8f53a7a8adbd47ad9e2c382b551517fae\": container with ID starting with 9f8636e8557d4162b97b500f76725cb8f53a7a8adbd47ad9e2c382b551517fae not found: ID does not exist" Feb 27 18:02:11 crc kubenswrapper[4700]: I0227 18:02:11.947920 4700 scope.go:117] "RemoveContainer" containerID="43618a3b775d62bd156d2ce6324815a7cb34457216f134b7947655e25c421cb6" Feb 27 18:02:11 crc kubenswrapper[4700]: E0227 18:02:11.948406 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43618a3b775d62bd156d2ce6324815a7cb34457216f134b7947655e25c421cb6\": container with ID starting with 43618a3b775d62bd156d2ce6324815a7cb34457216f134b7947655e25c421cb6 not found: ID does not exist" containerID="43618a3b775d62bd156d2ce6324815a7cb34457216f134b7947655e25c421cb6" Feb 27 18:02:11 crc kubenswrapper[4700]: I0227 18:02:11.948484 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43618a3b775d62bd156d2ce6324815a7cb34457216f134b7947655e25c421cb6"} err="failed to get container status \"43618a3b775d62bd156d2ce6324815a7cb34457216f134b7947655e25c421cb6\": rpc error: code = NotFound desc = could not find container \"43618a3b775d62bd156d2ce6324815a7cb34457216f134b7947655e25c421cb6\": container with ID starting with 43618a3b775d62bd156d2ce6324815a7cb34457216f134b7947655e25c421cb6 not found: ID does not exist" Feb 27 18:02:11 crc kubenswrapper[4700]: I0227 18:02:11.948519 4700 scope.go:117] "RemoveContainer" containerID="6d782c33eeb65e1278cf57a8b3e2d70a65e05abfd37f8dc16f9f10a5b13d5060" Feb 27 18:02:11 crc kubenswrapper[4700]: E0227 18:02:11.948949 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d782c33eeb65e1278cf57a8b3e2d70a65e05abfd37f8dc16f9f10a5b13d5060\": container with ID starting with 6d782c33eeb65e1278cf57a8b3e2d70a65e05abfd37f8dc16f9f10a5b13d5060 not found: ID does not exist" containerID="6d782c33eeb65e1278cf57a8b3e2d70a65e05abfd37f8dc16f9f10a5b13d5060" Feb 27 18:02:11 crc kubenswrapper[4700]: I0227 18:02:11.949003 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d782c33eeb65e1278cf57a8b3e2d70a65e05abfd37f8dc16f9f10a5b13d5060"} err="failed to get container status \"6d782c33eeb65e1278cf57a8b3e2d70a65e05abfd37f8dc16f9f10a5b13d5060\": rpc error: code = NotFound desc = could not find container \"6d782c33eeb65e1278cf57a8b3e2d70a65e05abfd37f8dc16f9f10a5b13d5060\": container with ID starting with 6d782c33eeb65e1278cf57a8b3e2d70a65e05abfd37f8dc16f9f10a5b13d5060 not found: ID does not exist" Feb 27 18:02:12 crc kubenswrapper[4700]: I0227 18:02:12.240117 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2cflz" Feb 27 18:02:12 crc kubenswrapper[4700]: I0227 18:02:12.343439 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e084226-907c-4278-8223-34f6fa78a9a4-catalog-content\") pod \"8e084226-907c-4278-8223-34f6fa78a9a4\" (UID: \"8e084226-907c-4278-8223-34f6fa78a9a4\") " Feb 27 18:02:12 crc kubenswrapper[4700]: I0227 18:02:12.343560 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9n8w\" (UniqueName: \"kubernetes.io/projected/8e084226-907c-4278-8223-34f6fa78a9a4-kube-api-access-x9n8w\") pod \"8e084226-907c-4278-8223-34f6fa78a9a4\" (UID: \"8e084226-907c-4278-8223-34f6fa78a9a4\") " Feb 27 18:02:12 crc kubenswrapper[4700]: I0227 18:02:12.343597 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e084226-907c-4278-8223-34f6fa78a9a4-utilities\") pod \"8e084226-907c-4278-8223-34f6fa78a9a4\" (UID: \"8e084226-907c-4278-8223-34f6fa78a9a4\") " Feb 27 18:02:12 crc kubenswrapper[4700]: I0227 18:02:12.344765 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e084226-907c-4278-8223-34f6fa78a9a4-utilities" (OuterVolumeSpecName: "utilities") pod "8e084226-907c-4278-8223-34f6fa78a9a4" (UID: "8e084226-907c-4278-8223-34f6fa78a9a4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 18:02:12 crc kubenswrapper[4700]: I0227 18:02:12.366300 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e084226-907c-4278-8223-34f6fa78a9a4-kube-api-access-x9n8w" (OuterVolumeSpecName: "kube-api-access-x9n8w") pod "8e084226-907c-4278-8223-34f6fa78a9a4" (UID: "8e084226-907c-4278-8223-34f6fa78a9a4"). InnerVolumeSpecName "kube-api-access-x9n8w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:02:12 crc kubenswrapper[4700]: I0227 18:02:12.418146 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e084226-907c-4278-8223-34f6fa78a9a4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8e084226-907c-4278-8223-34f6fa78a9a4" (UID: "8e084226-907c-4278-8223-34f6fa78a9a4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 18:02:12 crc kubenswrapper[4700]: I0227 18:02:12.446653 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e084226-907c-4278-8223-34f6fa78a9a4-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 18:02:12 crc kubenswrapper[4700]: I0227 18:02:12.446695 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9n8w\" (UniqueName: \"kubernetes.io/projected/8e084226-907c-4278-8223-34f6fa78a9a4-kube-api-access-x9n8w\") on node \"crc\" DevicePath \"\"" Feb 27 18:02:12 crc kubenswrapper[4700]: I0227 18:02:12.446709 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e084226-907c-4278-8223-34f6fa78a9a4-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 18:02:12 crc kubenswrapper[4700]: I0227 18:02:12.684845 4700 generic.go:334] "Generic (PLEG): container finished" podID="8e084226-907c-4278-8223-34f6fa78a9a4" containerID="b0eff5189f682c698b28377e71ace37538c9f60a800799ed7c57fe0f5a4b5ab7" exitCode=0 Feb 27 18:02:12 crc kubenswrapper[4700]: I0227 18:02:12.684888 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2cflz" event={"ID":"8e084226-907c-4278-8223-34f6fa78a9a4","Type":"ContainerDied","Data":"b0eff5189f682c698b28377e71ace37538c9f60a800799ed7c57fe0f5a4b5ab7"} Feb 27 18:02:12 crc kubenswrapper[4700]: I0227 18:02:12.684912 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2cflz" event={"ID":"8e084226-907c-4278-8223-34f6fa78a9a4","Type":"ContainerDied","Data":"ffe84c0e3d835a0c0d373caf392b80b4ed1a617e3f109e293e535841ff269329"} Feb 27 18:02:12 crc kubenswrapper[4700]: I0227 18:02:12.684932 4700 scope.go:117] "RemoveContainer" containerID="b0eff5189f682c698b28377e71ace37538c9f60a800799ed7c57fe0f5a4b5ab7" Feb 27 18:02:12 crc kubenswrapper[4700]: I0227 18:02:12.685050 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2cflz" Feb 27 18:02:12 crc kubenswrapper[4700]: I0227 18:02:12.739071 4700 scope.go:117] "RemoveContainer" containerID="169de72005f109b56d59b7748200069d120249fe2d134bc3d3582ca83d798441" Feb 27 18:02:12 crc kubenswrapper[4700]: I0227 18:02:12.745575 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2cflz"] Feb 27 18:02:12 crc kubenswrapper[4700]: I0227 18:02:12.761898 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2cflz"] Feb 27 18:02:12 crc kubenswrapper[4700]: I0227 18:02:12.767489 4700 scope.go:117] "RemoveContainer" containerID="58779ac75cbde384a3aa88df750835504c26ee66a6901a033019925eba11d358" Feb 27 18:02:12 crc kubenswrapper[4700]: I0227 18:02:12.813885 4700 scope.go:117] "RemoveContainer" containerID="b0eff5189f682c698b28377e71ace37538c9f60a800799ed7c57fe0f5a4b5ab7" Feb 27 18:02:12 crc kubenswrapper[4700]: E0227 18:02:12.814698 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0eff5189f682c698b28377e71ace37538c9f60a800799ed7c57fe0f5a4b5ab7\": container with ID starting with b0eff5189f682c698b28377e71ace37538c9f60a800799ed7c57fe0f5a4b5ab7 not found: ID does not exist" containerID="b0eff5189f682c698b28377e71ace37538c9f60a800799ed7c57fe0f5a4b5ab7" Feb 27 18:02:12 crc kubenswrapper[4700]: I0227 18:02:12.814733 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0eff5189f682c698b28377e71ace37538c9f60a800799ed7c57fe0f5a4b5ab7"} err="failed to get container status \"b0eff5189f682c698b28377e71ace37538c9f60a800799ed7c57fe0f5a4b5ab7\": rpc error: code = NotFound desc = could not find container \"b0eff5189f682c698b28377e71ace37538c9f60a800799ed7c57fe0f5a4b5ab7\": container with ID starting with b0eff5189f682c698b28377e71ace37538c9f60a800799ed7c57fe0f5a4b5ab7 not found: ID does not exist" Feb 27 18:02:12 crc kubenswrapper[4700]: I0227 18:02:12.814762 4700 scope.go:117] "RemoveContainer" containerID="169de72005f109b56d59b7748200069d120249fe2d134bc3d3582ca83d798441" Feb 27 18:02:12 crc kubenswrapper[4700]: E0227 18:02:12.815146 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"169de72005f109b56d59b7748200069d120249fe2d134bc3d3582ca83d798441\": container with ID starting with 169de72005f109b56d59b7748200069d120249fe2d134bc3d3582ca83d798441 not found: ID does not exist" containerID="169de72005f109b56d59b7748200069d120249fe2d134bc3d3582ca83d798441" Feb 27 18:02:12 crc kubenswrapper[4700]: I0227 18:02:12.815168 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"169de72005f109b56d59b7748200069d120249fe2d134bc3d3582ca83d798441"} err="failed to get container status \"169de72005f109b56d59b7748200069d120249fe2d134bc3d3582ca83d798441\": rpc error: code = NotFound desc = could not find container \"169de72005f109b56d59b7748200069d120249fe2d134bc3d3582ca83d798441\": container with ID starting with 169de72005f109b56d59b7748200069d120249fe2d134bc3d3582ca83d798441 not found: ID does not exist" Feb 27 18:02:12 crc kubenswrapper[4700]: I0227 18:02:12.815186 4700 scope.go:117] "RemoveContainer" containerID="58779ac75cbde384a3aa88df750835504c26ee66a6901a033019925eba11d358" Feb 27 18:02:12 crc kubenswrapper[4700]: E0227 18:02:12.815670 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58779ac75cbde384a3aa88df750835504c26ee66a6901a033019925eba11d358\": container with ID starting with 58779ac75cbde384a3aa88df750835504c26ee66a6901a033019925eba11d358 not found: ID does not exist" containerID="58779ac75cbde384a3aa88df750835504c26ee66a6901a033019925eba11d358" Feb 27 18:02:12 crc kubenswrapper[4700]: I0227 18:02:12.815694 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58779ac75cbde384a3aa88df750835504c26ee66a6901a033019925eba11d358"} err="failed to get container status \"58779ac75cbde384a3aa88df750835504c26ee66a6901a033019925eba11d358\": rpc error: code = NotFound desc = could not find container \"58779ac75cbde384a3aa88df750835504c26ee66a6901a033019925eba11d358\": container with ID starting with 58779ac75cbde384a3aa88df750835504c26ee66a6901a033019925eba11d358 not found: ID does not exist" Feb 27 18:02:12 crc kubenswrapper[4700]: I0227 18:02:12.998743 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e084226-907c-4278-8223-34f6fa78a9a4" path="/var/lib/kubelet/pods/8e084226-907c-4278-8223-34f6fa78a9a4/volumes" Feb 27 18:02:13 crc kubenswrapper[4700]: I0227 18:02:13.001606 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e14e41da-6e02-4d9f-9faf-68789c1dbb9e" path="/var/lib/kubelet/pods/e14e41da-6e02-4d9f-9faf-68789c1dbb9e/volumes" Feb 27 18:02:17 crc kubenswrapper[4700]: E0227 18:02:17.067687 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 18:02:17 crc kubenswrapper[4700]: E0227 18:02:17.069731 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 18:02:17 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 18:02:17 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-d7vxj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536922-d7lzf_openshift-infra(793f7988-e7d0-4b3b-a35f-7d0445ebab61): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 18:02:17 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 18:02:17 crc kubenswrapper[4700]: E0227 18:02:17.071292 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536922-d7lzf" podUID="793f7988-e7d0-4b3b-a35f-7d0445ebab61" Feb 27 18:02:27 crc kubenswrapper[4700]: E0227 18:02:27.986149 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536922-d7lzf" podUID="793f7988-e7d0-4b3b-a35f-7d0445ebab61" Feb 27 18:02:36 crc kubenswrapper[4700]: I0227 18:02:36.410747 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 18:02:36 crc kubenswrapper[4700]: I0227 18:02:36.411549 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 18:02:36 crc kubenswrapper[4700]: I0227 18:02:36.411628 4700 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" Feb 27 18:02:36 crc kubenswrapper[4700]: I0227 18:02:36.412932 4700 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5496fef1f08034cab735e8cc9a87daee6229da7616d276a174410ba8eacb94a6"} pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 18:02:36 crc kubenswrapper[4700]: I0227 18:02:36.413018 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" containerID="cri-o://5496fef1f08034cab735e8cc9a87daee6229da7616d276a174410ba8eacb94a6" gracePeriod=600 Feb 27 18:02:36 crc kubenswrapper[4700]: E0227 18:02:36.542552 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:02:36 crc kubenswrapper[4700]: I0227 18:02:36.992154 4700 generic.go:334] "Generic (PLEG): container finished" podID="84b45511-b94c-479f-98db-fd2c4eceec46" containerID="5496fef1f08034cab735e8cc9a87daee6229da7616d276a174410ba8eacb94a6" exitCode=0 Feb 27 18:02:36 crc kubenswrapper[4700]: I0227 18:02:36.996969 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerDied","Data":"5496fef1f08034cab735e8cc9a87daee6229da7616d276a174410ba8eacb94a6"} Feb 27 18:02:36 crc kubenswrapper[4700]: I0227 18:02:36.997057 4700 scope.go:117] "RemoveContainer" containerID="85becbca57d9b923c2d58f8b905a4fa4ab2c81b51b8ed23006735a99d384929c" Feb 27 18:02:36 crc kubenswrapper[4700]: I0227 18:02:36.998261 4700 scope.go:117] "RemoveContainer" containerID="5496fef1f08034cab735e8cc9a87daee6229da7616d276a174410ba8eacb94a6" Feb 27 18:02:36 crc kubenswrapper[4700]: E0227 18:02:36.999035 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:02:44 crc kubenswrapper[4700]: I0227 18:02:44.075513 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536922-d7lzf" event={"ID":"793f7988-e7d0-4b3b-a35f-7d0445ebab61","Type":"ContainerStarted","Data":"4a3e7ec9b3459eb0394533a5ebdf6e05a419a0ff52f797f7120460349a9a59cb"} Feb 27 18:02:44 crc kubenswrapper[4700]: I0227 18:02:44.103930 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536922-d7lzf" podStartSLOduration=1.6261918610000001 podStartE2EDuration="44.103906618s" podCreationTimestamp="2026-02-27 18:02:00 +0000 UTC" firstStartedPulling="2026-02-27 18:02:01.019656101 +0000 UTC m=+3681.004968858" lastFinishedPulling="2026-02-27 18:02:43.497370828 +0000 UTC m=+3723.482683615" observedRunningTime="2026-02-27 18:02:44.096955657 +0000 UTC m=+3724.082268444" watchObservedRunningTime="2026-02-27 18:02:44.103906618 +0000 UTC m=+3724.089219385" Feb 27 18:02:45 crc kubenswrapper[4700]: I0227 18:02:45.090812 4700 generic.go:334] "Generic (PLEG): container finished" podID="793f7988-e7d0-4b3b-a35f-7d0445ebab61" containerID="4a3e7ec9b3459eb0394533a5ebdf6e05a419a0ff52f797f7120460349a9a59cb" exitCode=0 Feb 27 18:02:45 crc kubenswrapper[4700]: I0227 18:02:45.090954 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536922-d7lzf" event={"ID":"793f7988-e7d0-4b3b-a35f-7d0445ebab61","Type":"ContainerDied","Data":"4a3e7ec9b3459eb0394533a5ebdf6e05a419a0ff52f797f7120460349a9a59cb"} Feb 27 18:02:46 crc kubenswrapper[4700]: I0227 18:02:46.536404 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536922-d7lzf" Feb 27 18:02:46 crc kubenswrapper[4700]: I0227 18:02:46.623126 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d7vxj\" (UniqueName: \"kubernetes.io/projected/793f7988-e7d0-4b3b-a35f-7d0445ebab61-kube-api-access-d7vxj\") pod \"793f7988-e7d0-4b3b-a35f-7d0445ebab61\" (UID: \"793f7988-e7d0-4b3b-a35f-7d0445ebab61\") " Feb 27 18:02:46 crc kubenswrapper[4700]: I0227 18:02:46.630736 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/793f7988-e7d0-4b3b-a35f-7d0445ebab61-kube-api-access-d7vxj" (OuterVolumeSpecName: "kube-api-access-d7vxj") pod "793f7988-e7d0-4b3b-a35f-7d0445ebab61" (UID: "793f7988-e7d0-4b3b-a35f-7d0445ebab61"). InnerVolumeSpecName "kube-api-access-d7vxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:02:46 crc kubenswrapper[4700]: I0227 18:02:46.725633 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d7vxj\" (UniqueName: \"kubernetes.io/projected/793f7988-e7d0-4b3b-a35f-7d0445ebab61-kube-api-access-d7vxj\") on node \"crc\" DevicePath \"\"" Feb 27 18:02:47 crc kubenswrapper[4700]: I0227 18:02:47.114568 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536922-d7lzf" event={"ID":"793f7988-e7d0-4b3b-a35f-7d0445ebab61","Type":"ContainerDied","Data":"a5849d0e4ab42c85e16ad82c41942984f6c41e0309fb72fc6c786f02414e5030"} Feb 27 18:02:47 crc kubenswrapper[4700]: I0227 18:02:47.114609 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536922-d7lzf" Feb 27 18:02:47 crc kubenswrapper[4700]: I0227 18:02:47.114614 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a5849d0e4ab42c85e16ad82c41942984f6c41e0309fb72fc6c786f02414e5030" Feb 27 18:02:47 crc kubenswrapper[4700]: I0227 18:02:47.174702 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536916-l7sxp"] Feb 27 18:02:47 crc kubenswrapper[4700]: I0227 18:02:47.184138 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536916-l7sxp"] Feb 27 18:02:48 crc kubenswrapper[4700]: I0227 18:02:48.997955 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd4abc15-d07f-46ee-9f1c-98164a4b2783" path="/var/lib/kubelet/pods/cd4abc15-d07f-46ee-9f1c-98164a4b2783/volumes" Feb 27 18:02:50 crc kubenswrapper[4700]: I0227 18:02:50.988508 4700 scope.go:117] "RemoveContainer" containerID="5496fef1f08034cab735e8cc9a87daee6229da7616d276a174410ba8eacb94a6" Feb 27 18:02:50 crc kubenswrapper[4700]: E0227 18:02:50.988836 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:03:02 crc kubenswrapper[4700]: I0227 18:03:02.981247 4700 scope.go:117] "RemoveContainer" containerID="5496fef1f08034cab735e8cc9a87daee6229da7616d276a174410ba8eacb94a6" Feb 27 18:03:02 crc kubenswrapper[4700]: E0227 18:03:02.982058 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:03:11 crc kubenswrapper[4700]: I0227 18:03:11.537755 4700 scope.go:117] "RemoveContainer" containerID="c8775f85417900834955f825c0a0230a38074772d1cf1c592d6c5dfd1c92d9fc" Feb 27 18:03:13 crc kubenswrapper[4700]: I0227 18:03:13.982727 4700 scope.go:117] "RemoveContainer" containerID="5496fef1f08034cab735e8cc9a87daee6229da7616d276a174410ba8eacb94a6" Feb 27 18:03:13 crc kubenswrapper[4700]: E0227 18:03:13.984309 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:03:26 crc kubenswrapper[4700]: I0227 18:03:26.980759 4700 scope.go:117] "RemoveContainer" containerID="5496fef1f08034cab735e8cc9a87daee6229da7616d276a174410ba8eacb94a6" Feb 27 18:03:26 crc kubenswrapper[4700]: E0227 18:03:26.982366 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:03:27 crc kubenswrapper[4700]: E0227 18:03:27.058449 4700 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.222:41668->38.102.83.222:41369: write tcp 38.102.83.222:41668->38.102.83.222:41369: write: broken pipe Feb 27 18:03:38 crc kubenswrapper[4700]: I0227 18:03:38.981375 4700 scope.go:117] "RemoveContainer" containerID="5496fef1f08034cab735e8cc9a87daee6229da7616d276a174410ba8eacb94a6" Feb 27 18:03:38 crc kubenswrapper[4700]: E0227 18:03:38.982569 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:03:50 crc kubenswrapper[4700]: I0227 18:03:50.990065 4700 scope.go:117] "RemoveContainer" containerID="5496fef1f08034cab735e8cc9a87daee6229da7616d276a174410ba8eacb94a6" Feb 27 18:03:50 crc kubenswrapper[4700]: E0227 18:03:50.991021 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:04:00 crc kubenswrapper[4700]: I0227 18:04:00.167794 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536924-7tpz9"] Feb 27 18:04:00 crc kubenswrapper[4700]: E0227 18:04:00.169704 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e14e41da-6e02-4d9f-9faf-68789c1dbb9e" containerName="extract-content" Feb 27 18:04:00 crc kubenswrapper[4700]: I0227 18:04:00.169745 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="e14e41da-6e02-4d9f-9faf-68789c1dbb9e" containerName="extract-content" Feb 27 18:04:00 crc kubenswrapper[4700]: E0227 18:04:00.169805 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e14e41da-6e02-4d9f-9faf-68789c1dbb9e" containerName="registry-server" Feb 27 18:04:00 crc kubenswrapper[4700]: I0227 18:04:00.169822 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="e14e41da-6e02-4d9f-9faf-68789c1dbb9e" containerName="registry-server" Feb 27 18:04:00 crc kubenswrapper[4700]: E0227 18:04:00.169856 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e084226-907c-4278-8223-34f6fa78a9a4" containerName="extract-utilities" Feb 27 18:04:00 crc kubenswrapper[4700]: I0227 18:04:00.169906 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e084226-907c-4278-8223-34f6fa78a9a4" containerName="extract-utilities" Feb 27 18:04:00 crc kubenswrapper[4700]: E0227 18:04:00.169931 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e084226-907c-4278-8223-34f6fa78a9a4" containerName="extract-content" Feb 27 18:04:00 crc kubenswrapper[4700]: I0227 18:04:00.169947 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e084226-907c-4278-8223-34f6fa78a9a4" containerName="extract-content" Feb 27 18:04:00 crc kubenswrapper[4700]: E0227 18:04:00.169986 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e084226-907c-4278-8223-34f6fa78a9a4" containerName="registry-server" Feb 27 18:04:00 crc kubenswrapper[4700]: I0227 18:04:00.170005 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e084226-907c-4278-8223-34f6fa78a9a4" containerName="registry-server" Feb 27 18:04:00 crc kubenswrapper[4700]: E0227 18:04:00.170026 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="793f7988-e7d0-4b3b-a35f-7d0445ebab61" containerName="oc" Feb 27 18:04:00 crc kubenswrapper[4700]: I0227 18:04:00.170043 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="793f7988-e7d0-4b3b-a35f-7d0445ebab61" containerName="oc" Feb 27 18:04:00 crc kubenswrapper[4700]: E0227 18:04:00.170084 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e14e41da-6e02-4d9f-9faf-68789c1dbb9e" containerName="extract-utilities" Feb 27 18:04:00 crc kubenswrapper[4700]: I0227 18:04:00.170100 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="e14e41da-6e02-4d9f-9faf-68789c1dbb9e" containerName="extract-utilities" Feb 27 18:04:00 crc kubenswrapper[4700]: I0227 18:04:00.170625 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="e14e41da-6e02-4d9f-9faf-68789c1dbb9e" containerName="registry-server" Feb 27 18:04:00 crc kubenswrapper[4700]: I0227 18:04:00.170685 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="793f7988-e7d0-4b3b-a35f-7d0445ebab61" containerName="oc" Feb 27 18:04:00 crc kubenswrapper[4700]: I0227 18:04:00.170713 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e084226-907c-4278-8223-34f6fa78a9a4" containerName="registry-server" Feb 27 18:04:00 crc kubenswrapper[4700]: I0227 18:04:00.172379 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536924-7tpz9" Feb 27 18:04:00 crc kubenswrapper[4700]: I0227 18:04:00.175607 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d62kc" Feb 27 18:04:00 crc kubenswrapper[4700]: I0227 18:04:00.178004 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 18:04:00 crc kubenswrapper[4700]: I0227 18:04:00.178100 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 18:04:00 crc kubenswrapper[4700]: I0227 18:04:00.184418 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536924-7tpz9"] Feb 27 18:04:00 crc kubenswrapper[4700]: I0227 18:04:00.223272 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25dtx\" (UniqueName: \"kubernetes.io/projected/860b240c-50d6-4d7d-88d6-d1ad82c7eece-kube-api-access-25dtx\") pod \"auto-csr-approver-29536924-7tpz9\" (UID: \"860b240c-50d6-4d7d-88d6-d1ad82c7eece\") " pod="openshift-infra/auto-csr-approver-29536924-7tpz9" Feb 27 18:04:00 crc kubenswrapper[4700]: I0227 18:04:00.326046 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25dtx\" (UniqueName: \"kubernetes.io/projected/860b240c-50d6-4d7d-88d6-d1ad82c7eece-kube-api-access-25dtx\") pod \"auto-csr-approver-29536924-7tpz9\" (UID: \"860b240c-50d6-4d7d-88d6-d1ad82c7eece\") " pod="openshift-infra/auto-csr-approver-29536924-7tpz9" Feb 27 18:04:00 crc kubenswrapper[4700]: I0227 18:04:00.367908 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25dtx\" (UniqueName: \"kubernetes.io/projected/860b240c-50d6-4d7d-88d6-d1ad82c7eece-kube-api-access-25dtx\") pod \"auto-csr-approver-29536924-7tpz9\" (UID: \"860b240c-50d6-4d7d-88d6-d1ad82c7eece\") " pod="openshift-infra/auto-csr-approver-29536924-7tpz9" Feb 27 18:04:00 crc kubenswrapper[4700]: I0227 18:04:00.505797 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536924-7tpz9" Feb 27 18:04:01 crc kubenswrapper[4700]: I0227 18:04:01.055127 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536924-7tpz9"] Feb 27 18:04:02 crc kubenswrapper[4700]: I0227 18:04:02.024617 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536924-7tpz9" event={"ID":"860b240c-50d6-4d7d-88d6-d1ad82c7eece","Type":"ContainerStarted","Data":"8fb078ed33f606150c701cf415face6f71c705228a64836d3b97e43c2fbb33c5"} Feb 27 18:04:02 crc kubenswrapper[4700]: I0227 18:04:02.982521 4700 scope.go:117] "RemoveContainer" containerID="5496fef1f08034cab735e8cc9a87daee6229da7616d276a174410ba8eacb94a6" Feb 27 18:04:02 crc kubenswrapper[4700]: E0227 18:04:02.982990 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:04:04 crc kubenswrapper[4700]: I0227 18:04:04.045263 4700 generic.go:334] "Generic (PLEG): container finished" podID="860b240c-50d6-4d7d-88d6-d1ad82c7eece" containerID="96f0a92db10c87c54ea74c7765c35e5bff7413044cf9c77d45a486e2b89f4ccc" exitCode=0 Feb 27 18:04:04 crc kubenswrapper[4700]: I0227 18:04:04.045340 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536924-7tpz9" event={"ID":"860b240c-50d6-4d7d-88d6-d1ad82c7eece","Type":"ContainerDied","Data":"96f0a92db10c87c54ea74c7765c35e5bff7413044cf9c77d45a486e2b89f4ccc"} Feb 27 18:04:05 crc kubenswrapper[4700]: I0227 18:04:05.421217 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536924-7tpz9" Feb 27 18:04:05 crc kubenswrapper[4700]: I0227 18:04:05.545129 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25dtx\" (UniqueName: \"kubernetes.io/projected/860b240c-50d6-4d7d-88d6-d1ad82c7eece-kube-api-access-25dtx\") pod \"860b240c-50d6-4d7d-88d6-d1ad82c7eece\" (UID: \"860b240c-50d6-4d7d-88d6-d1ad82c7eece\") " Feb 27 18:04:05 crc kubenswrapper[4700]: I0227 18:04:05.556730 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/860b240c-50d6-4d7d-88d6-d1ad82c7eece-kube-api-access-25dtx" (OuterVolumeSpecName: "kube-api-access-25dtx") pod "860b240c-50d6-4d7d-88d6-d1ad82c7eece" (UID: "860b240c-50d6-4d7d-88d6-d1ad82c7eece"). InnerVolumeSpecName "kube-api-access-25dtx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:04:05 crc kubenswrapper[4700]: I0227 18:04:05.647554 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25dtx\" (UniqueName: \"kubernetes.io/projected/860b240c-50d6-4d7d-88d6-d1ad82c7eece-kube-api-access-25dtx\") on node \"crc\" DevicePath \"\"" Feb 27 18:04:06 crc kubenswrapper[4700]: I0227 18:04:06.064505 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536924-7tpz9" event={"ID":"860b240c-50d6-4d7d-88d6-d1ad82c7eece","Type":"ContainerDied","Data":"8fb078ed33f606150c701cf415face6f71c705228a64836d3b97e43c2fbb33c5"} Feb 27 18:04:06 crc kubenswrapper[4700]: I0227 18:04:06.064831 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8fb078ed33f606150c701cf415face6f71c705228a64836d3b97e43c2fbb33c5" Feb 27 18:04:06 crc kubenswrapper[4700]: I0227 18:04:06.064595 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536924-7tpz9" Feb 27 18:04:06 crc kubenswrapper[4700]: I0227 18:04:06.504172 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536918-njfw6"] Feb 27 18:04:06 crc kubenswrapper[4700]: I0227 18:04:06.516286 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536918-njfw6"] Feb 27 18:04:06 crc kubenswrapper[4700]: I0227 18:04:06.998964 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fe3904b-bcc3-473c-ba68-94922e3915cb" path="/var/lib/kubelet/pods/8fe3904b-bcc3-473c-ba68-94922e3915cb/volumes" Feb 27 18:04:11 crc kubenswrapper[4700]: I0227 18:04:11.669982 4700 scope.go:117] "RemoveContainer" containerID="144b2eabdc01738aa139e79f03eb42021d0d81450267e2c35e9c082b88d79cb3" Feb 27 18:04:17 crc kubenswrapper[4700]: I0227 18:04:17.981870 4700 scope.go:117] "RemoveContainer" containerID="5496fef1f08034cab735e8cc9a87daee6229da7616d276a174410ba8eacb94a6" Feb 27 18:04:17 crc kubenswrapper[4700]: E0227 18:04:17.983126 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:04:32 crc kubenswrapper[4700]: I0227 18:04:32.981918 4700 scope.go:117] "RemoveContainer" containerID="5496fef1f08034cab735e8cc9a87daee6229da7616d276a174410ba8eacb94a6" Feb 27 18:04:32 crc kubenswrapper[4700]: E0227 18:04:32.984053 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:04:47 crc kubenswrapper[4700]: I0227 18:04:47.981913 4700 scope.go:117] "RemoveContainer" containerID="5496fef1f08034cab735e8cc9a87daee6229da7616d276a174410ba8eacb94a6" Feb 27 18:04:47 crc kubenswrapper[4700]: E0227 18:04:47.982785 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:04:58 crc kubenswrapper[4700]: I0227 18:04:58.981618 4700 scope.go:117] "RemoveContainer" containerID="5496fef1f08034cab735e8cc9a87daee6229da7616d276a174410ba8eacb94a6" Feb 27 18:04:58 crc kubenswrapper[4700]: E0227 18:04:58.983175 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:05:09 crc kubenswrapper[4700]: I0227 18:05:09.982323 4700 scope.go:117] "RemoveContainer" containerID="5496fef1f08034cab735e8cc9a87daee6229da7616d276a174410ba8eacb94a6" Feb 27 18:05:09 crc kubenswrapper[4700]: E0227 18:05:09.983617 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:05:22 crc kubenswrapper[4700]: I0227 18:05:22.981922 4700 scope.go:117] "RemoveContainer" containerID="5496fef1f08034cab735e8cc9a87daee6229da7616d276a174410ba8eacb94a6" Feb 27 18:05:22 crc kubenswrapper[4700]: E0227 18:05:22.983960 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:05:34 crc kubenswrapper[4700]: I0227 18:05:34.982160 4700 scope.go:117] "RemoveContainer" containerID="5496fef1f08034cab735e8cc9a87daee6229da7616d276a174410ba8eacb94a6" Feb 27 18:05:34 crc kubenswrapper[4700]: E0227 18:05:34.983327 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:05:45 crc kubenswrapper[4700]: I0227 18:05:45.982994 4700 scope.go:117] "RemoveContainer" containerID="5496fef1f08034cab735e8cc9a87daee6229da7616d276a174410ba8eacb94a6" Feb 27 18:05:45 crc kubenswrapper[4700]: E0227 18:05:45.985247 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:05:56 crc kubenswrapper[4700]: I0227 18:05:56.982453 4700 scope.go:117] "RemoveContainer" containerID="5496fef1f08034cab735e8cc9a87daee6229da7616d276a174410ba8eacb94a6" Feb 27 18:05:56 crc kubenswrapper[4700]: E0227 18:05:56.983818 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:06:00 crc kubenswrapper[4700]: I0227 18:06:00.161286 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536926-s6ft8"] Feb 27 18:06:00 crc kubenswrapper[4700]: E0227 18:06:00.162450 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="860b240c-50d6-4d7d-88d6-d1ad82c7eece" containerName="oc" Feb 27 18:06:00 crc kubenswrapper[4700]: I0227 18:06:00.162489 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="860b240c-50d6-4d7d-88d6-d1ad82c7eece" containerName="oc" Feb 27 18:06:00 crc kubenswrapper[4700]: I0227 18:06:00.162780 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="860b240c-50d6-4d7d-88d6-d1ad82c7eece" containerName="oc" Feb 27 18:06:00 crc kubenswrapper[4700]: I0227 18:06:00.163733 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" Feb 27 18:06:00 crc kubenswrapper[4700]: I0227 18:06:00.166860 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 18:06:00 crc kubenswrapper[4700]: I0227 18:06:00.166882 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 18:06:00 crc kubenswrapper[4700]: I0227 18:06:00.167786 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d62kc" Feb 27 18:06:00 crc kubenswrapper[4700]: I0227 18:06:00.178346 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536926-s6ft8"] Feb 27 18:06:00 crc kubenswrapper[4700]: I0227 18:06:00.240578 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjvdh\" (UniqueName: \"kubernetes.io/projected/d37cbaaf-1dee-4efd-9327-7bee80986a35-kube-api-access-hjvdh\") pod \"auto-csr-approver-29536926-s6ft8\" (UID: \"d37cbaaf-1dee-4efd-9327-7bee80986a35\") " pod="openshift-infra/auto-csr-approver-29536926-s6ft8" Feb 27 18:06:00 crc kubenswrapper[4700]: I0227 18:06:00.343341 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjvdh\" (UniqueName: \"kubernetes.io/projected/d37cbaaf-1dee-4efd-9327-7bee80986a35-kube-api-access-hjvdh\") pod \"auto-csr-approver-29536926-s6ft8\" (UID: \"d37cbaaf-1dee-4efd-9327-7bee80986a35\") " pod="openshift-infra/auto-csr-approver-29536926-s6ft8" Feb 27 18:06:00 crc kubenswrapper[4700]: I0227 18:06:00.363389 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjvdh\" (UniqueName: \"kubernetes.io/projected/d37cbaaf-1dee-4efd-9327-7bee80986a35-kube-api-access-hjvdh\") pod \"auto-csr-approver-29536926-s6ft8\" (UID: \"d37cbaaf-1dee-4efd-9327-7bee80986a35\") " pod="openshift-infra/auto-csr-approver-29536926-s6ft8" Feb 27 18:06:00 crc kubenswrapper[4700]: I0227 18:06:00.495243 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" Feb 27 18:06:01 crc kubenswrapper[4700]: I0227 18:06:01.015856 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536926-s6ft8"] Feb 27 18:06:01 crc kubenswrapper[4700]: I0227 18:06:01.030456 4700 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 18:06:01 crc kubenswrapper[4700]: I0227 18:06:01.609779 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" event={"ID":"d37cbaaf-1dee-4efd-9327-7bee80986a35","Type":"ContainerStarted","Data":"b52f8a05c71006850ae39fde28800be4e5a2a2fa515d775f7cac159f03db22dd"} Feb 27 18:06:01 crc kubenswrapper[4700]: E0227 18:06:01.946939 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 18:06:01 crc kubenswrapper[4700]: E0227 18:06:01.947111 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 18:06:01 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 18:06:01 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hjvdh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536926-s6ft8_openshift-infra(d37cbaaf-1dee-4efd-9327-7bee80986a35): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 18:06:01 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 18:06:01 crc kubenswrapper[4700]: E0227 18:06:01.948281 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:06:02 crc kubenswrapper[4700]: E0227 18:06:02.630015 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:06:08 crc kubenswrapper[4700]: I0227 18:06:08.982035 4700 scope.go:117] "RemoveContainer" containerID="5496fef1f08034cab735e8cc9a87daee6229da7616d276a174410ba8eacb94a6" Feb 27 18:06:08 crc kubenswrapper[4700]: E0227 18:06:08.983168 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:06:17 crc kubenswrapper[4700]: E0227 18:06:17.450294 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 18:06:17 crc kubenswrapper[4700]: E0227 18:06:17.451218 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 18:06:17 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 18:06:17 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hjvdh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536926-s6ft8_openshift-infra(d37cbaaf-1dee-4efd-9327-7bee80986a35): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 18:06:17 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 18:06:17 crc kubenswrapper[4700]: E0227 18:06:17.452619 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:06:21 crc kubenswrapper[4700]: I0227 18:06:21.982650 4700 scope.go:117] "RemoveContainer" containerID="5496fef1f08034cab735e8cc9a87daee6229da7616d276a174410ba8eacb94a6" Feb 27 18:06:21 crc kubenswrapper[4700]: E0227 18:06:21.984059 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:06:28 crc kubenswrapper[4700]: E0227 18:06:28.985203 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:06:35 crc kubenswrapper[4700]: I0227 18:06:35.982248 4700 scope.go:117] "RemoveContainer" containerID="5496fef1f08034cab735e8cc9a87daee6229da7616d276a174410ba8eacb94a6" Feb 27 18:06:35 crc kubenswrapper[4700]: E0227 18:06:35.983863 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:06:45 crc kubenswrapper[4700]: E0227 18:06:45.555161 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 18:06:45 crc kubenswrapper[4700]: E0227 18:06:45.556003 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 18:06:45 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 18:06:45 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hjvdh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536926-s6ft8_openshift-infra(d37cbaaf-1dee-4efd-9327-7bee80986a35): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 18:06:45 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 18:06:45 crc kubenswrapper[4700]: E0227 18:06:45.557297 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:06:47 crc kubenswrapper[4700]: I0227 18:06:47.981815 4700 scope.go:117] "RemoveContainer" containerID="5496fef1f08034cab735e8cc9a87daee6229da7616d276a174410ba8eacb94a6" Feb 27 18:06:47 crc kubenswrapper[4700]: E0227 18:06:47.982826 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:06:56 crc kubenswrapper[4700]: E0227 18:06:56.985953 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:06:57 crc kubenswrapper[4700]: I0227 18:06:57.906738 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-65457d8799-dclxm" podUID="e43c5d5e-63ef-45b9-af4e-627a035b376e" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Feb 27 18:06:59 crc kubenswrapper[4700]: I0227 18:06:59.981132 4700 scope.go:117] "RemoveContainer" containerID="5496fef1f08034cab735e8cc9a87daee6229da7616d276a174410ba8eacb94a6" Feb 27 18:06:59 crc kubenswrapper[4700]: E0227 18:06:59.981832 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:07:11 crc kubenswrapper[4700]: E0227 18:07:11.000435 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:07:12 crc kubenswrapper[4700]: I0227 18:07:12.981242 4700 scope.go:117] "RemoveContainer" containerID="5496fef1f08034cab735e8cc9a87daee6229da7616d276a174410ba8eacb94a6" Feb 27 18:07:12 crc kubenswrapper[4700]: E0227 18:07:12.982135 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:07:22 crc kubenswrapper[4700]: E0227 18:07:22.985364 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:07:27 crc kubenswrapper[4700]: I0227 18:07:27.981746 4700 scope.go:117] "RemoveContainer" containerID="5496fef1f08034cab735e8cc9a87daee6229da7616d276a174410ba8eacb94a6" Feb 27 18:07:27 crc kubenswrapper[4700]: E0227 18:07:27.983143 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:07:38 crc kubenswrapper[4700]: E0227 18:07:38.793851 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 18:07:38 crc kubenswrapper[4700]: E0227 18:07:38.794709 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 18:07:38 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 18:07:38 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hjvdh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536926-s6ft8_openshift-infra(d37cbaaf-1dee-4efd-9327-7bee80986a35): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 18:07:38 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 18:07:38 crc kubenswrapper[4700]: E0227 18:07:38.795981 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:07:42 crc kubenswrapper[4700]: I0227 18:07:42.981860 4700 scope.go:117] "RemoveContainer" containerID="5496fef1f08034cab735e8cc9a87daee6229da7616d276a174410ba8eacb94a6" Feb 27 18:07:44 crc kubenswrapper[4700]: I0227 18:07:44.068362 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerStarted","Data":"a22d07b860150538bb9582eddfae339a5c84757100045eb131b89e75822dfe43"} Feb 27 18:07:49 crc kubenswrapper[4700]: E0227 18:07:49.985445 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:08:00 crc kubenswrapper[4700]: I0227 18:08:00.176498 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536928-hphz9"] Feb 27 18:08:00 crc kubenswrapper[4700]: I0227 18:08:00.178635 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536928-hphz9" Feb 27 18:08:00 crc kubenswrapper[4700]: I0227 18:08:00.190858 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536928-hphz9"] Feb 27 18:08:00 crc kubenswrapper[4700]: I0227 18:08:00.192457 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2qgr\" (UniqueName: \"kubernetes.io/projected/e89023d8-dd8a-49ad-8749-5e54c469007d-kube-api-access-k2qgr\") pod \"auto-csr-approver-29536928-hphz9\" (UID: \"e89023d8-dd8a-49ad-8749-5e54c469007d\") " pod="openshift-infra/auto-csr-approver-29536928-hphz9" Feb 27 18:08:00 crc kubenswrapper[4700]: I0227 18:08:00.294723 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2qgr\" (UniqueName: \"kubernetes.io/projected/e89023d8-dd8a-49ad-8749-5e54c469007d-kube-api-access-k2qgr\") pod \"auto-csr-approver-29536928-hphz9\" (UID: \"e89023d8-dd8a-49ad-8749-5e54c469007d\") " pod="openshift-infra/auto-csr-approver-29536928-hphz9" Feb 27 18:08:00 crc kubenswrapper[4700]: I0227 18:08:00.330227 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2qgr\" (UniqueName: \"kubernetes.io/projected/e89023d8-dd8a-49ad-8749-5e54c469007d-kube-api-access-k2qgr\") pod \"auto-csr-approver-29536928-hphz9\" (UID: \"e89023d8-dd8a-49ad-8749-5e54c469007d\") " pod="openshift-infra/auto-csr-approver-29536928-hphz9" Feb 27 18:08:00 crc kubenswrapper[4700]: I0227 18:08:00.522826 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536928-hphz9" Feb 27 18:08:01 crc kubenswrapper[4700]: I0227 18:08:01.049908 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536928-hphz9"] Feb 27 18:08:01 crc kubenswrapper[4700]: I0227 18:08:01.249983 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536928-hphz9" event={"ID":"e89023d8-dd8a-49ad-8749-5e54c469007d","Type":"ContainerStarted","Data":"a7c04a5bde176c99f9a44c6b4693adcd72736473e83586b5f526fa5f3d5a3818"} Feb 27 18:08:03 crc kubenswrapper[4700]: E0227 18:08:03.017363 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 18:08:03 crc kubenswrapper[4700]: E0227 18:08:03.019027 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 18:08:03 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 18:08:03 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-k2qgr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536928-hphz9_openshift-infra(e89023d8-dd8a-49ad-8749-5e54c469007d): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 18:08:03 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 18:08:03 crc kubenswrapper[4700]: E0227 18:08:03.020815 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536928-hphz9" podUID="e89023d8-dd8a-49ad-8749-5e54c469007d" Feb 27 18:08:03 crc kubenswrapper[4700]: E0227 18:08:03.279211 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536928-hphz9" podUID="e89023d8-dd8a-49ad-8749-5e54c469007d" Feb 27 18:08:03 crc kubenswrapper[4700]: E0227 18:08:03.983207 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:08:14 crc kubenswrapper[4700]: E0227 18:08:14.984734 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:08:17 crc kubenswrapper[4700]: I0227 18:08:17.455105 4700 generic.go:334] "Generic (PLEG): container finished" podID="e89023d8-dd8a-49ad-8749-5e54c469007d" containerID="6011bbefc46aa294432da75f46bd16d87194a34721c8efb015a8d11d76420056" exitCode=0 Feb 27 18:08:17 crc kubenswrapper[4700]: I0227 18:08:17.455218 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536928-hphz9" event={"ID":"e89023d8-dd8a-49ad-8749-5e54c469007d","Type":"ContainerDied","Data":"6011bbefc46aa294432da75f46bd16d87194a34721c8efb015a8d11d76420056"} Feb 27 18:08:18 crc kubenswrapper[4700]: I0227 18:08:18.994050 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536928-hphz9" Feb 27 18:08:19 crc kubenswrapper[4700]: I0227 18:08:19.133549 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k2qgr\" (UniqueName: \"kubernetes.io/projected/e89023d8-dd8a-49ad-8749-5e54c469007d-kube-api-access-k2qgr\") pod \"e89023d8-dd8a-49ad-8749-5e54c469007d\" (UID: \"e89023d8-dd8a-49ad-8749-5e54c469007d\") " Feb 27 18:08:19 crc kubenswrapper[4700]: I0227 18:08:19.151663 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e89023d8-dd8a-49ad-8749-5e54c469007d-kube-api-access-k2qgr" (OuterVolumeSpecName: "kube-api-access-k2qgr") pod "e89023d8-dd8a-49ad-8749-5e54c469007d" (UID: "e89023d8-dd8a-49ad-8749-5e54c469007d"). InnerVolumeSpecName "kube-api-access-k2qgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:08:19 crc kubenswrapper[4700]: I0227 18:08:19.237050 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k2qgr\" (UniqueName: \"kubernetes.io/projected/e89023d8-dd8a-49ad-8749-5e54c469007d-kube-api-access-k2qgr\") on node \"crc\" DevicePath \"\"" Feb 27 18:08:19 crc kubenswrapper[4700]: I0227 18:08:19.487091 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536928-hphz9" event={"ID":"e89023d8-dd8a-49ad-8749-5e54c469007d","Type":"ContainerDied","Data":"a7c04a5bde176c99f9a44c6b4693adcd72736473e83586b5f526fa5f3d5a3818"} Feb 27 18:08:19 crc kubenswrapper[4700]: I0227 18:08:19.487549 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7c04a5bde176c99f9a44c6b4693adcd72736473e83586b5f526fa5f3d5a3818" Feb 27 18:08:19 crc kubenswrapper[4700]: I0227 18:08:19.487622 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536928-hphz9" Feb 27 18:08:20 crc kubenswrapper[4700]: I0227 18:08:20.105336 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536920-s86bn"] Feb 27 18:08:20 crc kubenswrapper[4700]: I0227 18:08:20.115781 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536920-s86bn"] Feb 27 18:08:21 crc kubenswrapper[4700]: I0227 18:08:21.003510 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1a1cead-0efc-4a4c-9c7e-5ebdfb14086a" path="/var/lib/kubelet/pods/b1a1cead-0efc-4a4c-9c7e-5ebdfb14086a/volumes" Feb 27 18:08:26 crc kubenswrapper[4700]: E0227 18:08:26.990096 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:08:41 crc kubenswrapper[4700]: E0227 18:08:40.999669 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:08:51 crc kubenswrapper[4700]: I0227 18:08:51.703515 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6hvnd"] Feb 27 18:08:51 crc kubenswrapper[4700]: E0227 18:08:51.704811 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e89023d8-dd8a-49ad-8749-5e54c469007d" containerName="oc" Feb 27 18:08:51 crc kubenswrapper[4700]: I0227 18:08:51.704830 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="e89023d8-dd8a-49ad-8749-5e54c469007d" containerName="oc" Feb 27 18:08:51 crc kubenswrapper[4700]: I0227 18:08:51.705121 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="e89023d8-dd8a-49ad-8749-5e54c469007d" containerName="oc" Feb 27 18:08:51 crc kubenswrapper[4700]: I0227 18:08:51.707146 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6hvnd" Feb 27 18:08:51 crc kubenswrapper[4700]: I0227 18:08:51.725660 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6hvnd"] Feb 27 18:08:51 crc kubenswrapper[4700]: I0227 18:08:51.785776 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea15dc00-097e-4356-8178-ac93ddb9ebea-utilities\") pod \"redhat-operators-6hvnd\" (UID: \"ea15dc00-097e-4356-8178-ac93ddb9ebea\") " pod="openshift-marketplace/redhat-operators-6hvnd" Feb 27 18:08:51 crc kubenswrapper[4700]: I0227 18:08:51.785952 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea15dc00-097e-4356-8178-ac93ddb9ebea-catalog-content\") pod \"redhat-operators-6hvnd\" (UID: \"ea15dc00-097e-4356-8178-ac93ddb9ebea\") " pod="openshift-marketplace/redhat-operators-6hvnd" Feb 27 18:08:51 crc kubenswrapper[4700]: I0227 18:08:51.786021 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdxw2\" (UniqueName: \"kubernetes.io/projected/ea15dc00-097e-4356-8178-ac93ddb9ebea-kube-api-access-rdxw2\") pod \"redhat-operators-6hvnd\" (UID: \"ea15dc00-097e-4356-8178-ac93ddb9ebea\") " pod="openshift-marketplace/redhat-operators-6hvnd" Feb 27 18:08:51 crc kubenswrapper[4700]: I0227 18:08:51.889125 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea15dc00-097e-4356-8178-ac93ddb9ebea-utilities\") pod \"redhat-operators-6hvnd\" (UID: \"ea15dc00-097e-4356-8178-ac93ddb9ebea\") " pod="openshift-marketplace/redhat-operators-6hvnd" Feb 27 18:08:51 crc kubenswrapper[4700]: I0227 18:08:51.889295 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea15dc00-097e-4356-8178-ac93ddb9ebea-catalog-content\") pod \"redhat-operators-6hvnd\" (UID: \"ea15dc00-097e-4356-8178-ac93ddb9ebea\") " pod="openshift-marketplace/redhat-operators-6hvnd" Feb 27 18:08:51 crc kubenswrapper[4700]: I0227 18:08:51.889352 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdxw2\" (UniqueName: \"kubernetes.io/projected/ea15dc00-097e-4356-8178-ac93ddb9ebea-kube-api-access-rdxw2\") pod \"redhat-operators-6hvnd\" (UID: \"ea15dc00-097e-4356-8178-ac93ddb9ebea\") " pod="openshift-marketplace/redhat-operators-6hvnd" Feb 27 18:08:51 crc kubenswrapper[4700]: I0227 18:08:51.889737 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea15dc00-097e-4356-8178-ac93ddb9ebea-utilities\") pod \"redhat-operators-6hvnd\" (UID: \"ea15dc00-097e-4356-8178-ac93ddb9ebea\") " pod="openshift-marketplace/redhat-operators-6hvnd" Feb 27 18:08:51 crc kubenswrapper[4700]: I0227 18:08:51.890574 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea15dc00-097e-4356-8178-ac93ddb9ebea-catalog-content\") pod \"redhat-operators-6hvnd\" (UID: \"ea15dc00-097e-4356-8178-ac93ddb9ebea\") " pod="openshift-marketplace/redhat-operators-6hvnd" Feb 27 18:08:51 crc kubenswrapper[4700]: I0227 18:08:51.929254 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdxw2\" (UniqueName: \"kubernetes.io/projected/ea15dc00-097e-4356-8178-ac93ddb9ebea-kube-api-access-rdxw2\") pod \"redhat-operators-6hvnd\" (UID: \"ea15dc00-097e-4356-8178-ac93ddb9ebea\") " pod="openshift-marketplace/redhat-operators-6hvnd" Feb 27 18:08:52 crc kubenswrapper[4700]: I0227 18:08:52.070224 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6hvnd" Feb 27 18:08:52 crc kubenswrapper[4700]: I0227 18:08:52.823267 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6hvnd"] Feb 27 18:08:52 crc kubenswrapper[4700]: I0227 18:08:52.976915 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6hvnd" event={"ID":"ea15dc00-097e-4356-8178-ac93ddb9ebea","Type":"ContainerStarted","Data":"fa207fd5b22c11b8eacede81b339fc65760c65b4f593a50a0ad6b67fd1f36ac7"} Feb 27 18:08:53 crc kubenswrapper[4700]: E0227 18:08:53.983948 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:08:53 crc kubenswrapper[4700]: I0227 18:08:53.990990 4700 generic.go:334] "Generic (PLEG): container finished" podID="ea15dc00-097e-4356-8178-ac93ddb9ebea" containerID="2275bd601d299407c983290efeaa194993070f46db036dc7b24d94606395c99f" exitCode=0 Feb 27 18:08:53 crc kubenswrapper[4700]: I0227 18:08:53.991073 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6hvnd" event={"ID":"ea15dc00-097e-4356-8178-ac93ddb9ebea","Type":"ContainerDied","Data":"2275bd601d299407c983290efeaa194993070f46db036dc7b24d94606395c99f"} Feb 27 18:08:54 crc kubenswrapper[4700]: E0227 18:08:54.793150 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 27 18:08:54 crc kubenswrapper[4700]: E0227 18:08:54.794584 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdxw2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-6hvnd_openshift-marketplace(ea15dc00-097e-4356-8178-ac93ddb9ebea): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 18:08:54 crc kubenswrapper[4700]: E0227 18:08:54.796080 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-operators-6hvnd" podUID="ea15dc00-097e-4356-8178-ac93ddb9ebea" Feb 27 18:08:55 crc kubenswrapper[4700]: E0227 18:08:55.008722 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-6hvnd" podUID="ea15dc00-097e-4356-8178-ac93ddb9ebea" Feb 27 18:09:06 crc kubenswrapper[4700]: E0227 18:09:06.038104 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 18:09:06 crc kubenswrapper[4700]: E0227 18:09:06.039010 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 18:09:06 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 18:09:06 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hjvdh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536926-s6ft8_openshift-infra(d37cbaaf-1dee-4efd-9327-7bee80986a35): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 18:09:06 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 18:09:06 crc kubenswrapper[4700]: E0227 18:09:06.040428 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:09:08 crc kubenswrapper[4700]: I0227 18:09:08.172677 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6hvnd" event={"ID":"ea15dc00-097e-4356-8178-ac93ddb9ebea","Type":"ContainerStarted","Data":"0d1769ac82fd0dae453ce862c137b761e1484d90b8c0e56ded1ac857f6c122e1"} Feb 27 18:09:11 crc kubenswrapper[4700]: I0227 18:09:11.857420 4700 scope.go:117] "RemoveContainer" containerID="30ba2e2521752f4b2c379c95364dca6e343fd4dbbcfe85d1673eb48311cf395e" Feb 27 18:09:17 crc kubenswrapper[4700]: E0227 18:09:17.984412 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:09:20 crc kubenswrapper[4700]: I0227 18:09:20.313186 4700 generic.go:334] "Generic (PLEG): container finished" podID="ea15dc00-097e-4356-8178-ac93ddb9ebea" containerID="0d1769ac82fd0dae453ce862c137b761e1484d90b8c0e56ded1ac857f6c122e1" exitCode=0 Feb 27 18:09:20 crc kubenswrapper[4700]: I0227 18:09:20.313248 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6hvnd" event={"ID":"ea15dc00-097e-4356-8178-ac93ddb9ebea","Type":"ContainerDied","Data":"0d1769ac82fd0dae453ce862c137b761e1484d90b8c0e56ded1ac857f6c122e1"} Feb 27 18:09:21 crc kubenswrapper[4700]: I0227 18:09:21.325140 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6hvnd" event={"ID":"ea15dc00-097e-4356-8178-ac93ddb9ebea","Type":"ContainerStarted","Data":"7ce54e47f84b8e705c23cb198be3d58a5bc2d8f4c1e6194df6e1a3007edfac88"} Feb 27 18:09:21 crc kubenswrapper[4700]: I0227 18:09:21.366584 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6hvnd" podStartSLOduration=3.443657935 podStartE2EDuration="30.366561101s" podCreationTimestamp="2026-02-27 18:08:51 +0000 UTC" firstStartedPulling="2026-02-27 18:08:53.993125099 +0000 UTC m=+4093.978437846" lastFinishedPulling="2026-02-27 18:09:20.916028265 +0000 UTC m=+4120.901341012" observedRunningTime="2026-02-27 18:09:21.354332952 +0000 UTC m=+4121.339645689" watchObservedRunningTime="2026-02-27 18:09:21.366561101 +0000 UTC m=+4121.351873858" Feb 27 18:09:22 crc kubenswrapper[4700]: I0227 18:09:22.071253 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6hvnd" Feb 27 18:09:22 crc kubenswrapper[4700]: I0227 18:09:22.071317 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6hvnd" Feb 27 18:09:23 crc kubenswrapper[4700]: I0227 18:09:23.138487 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6hvnd" podUID="ea15dc00-097e-4356-8178-ac93ddb9ebea" containerName="registry-server" probeResult="failure" output=< Feb 27 18:09:23 crc kubenswrapper[4700]: timeout: failed to connect service ":50051" within 1s Feb 27 18:09:23 crc kubenswrapper[4700]: > Feb 27 18:09:32 crc kubenswrapper[4700]: E0227 18:09:32.983049 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:09:33 crc kubenswrapper[4700]: I0227 18:09:33.143024 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6hvnd" podUID="ea15dc00-097e-4356-8178-ac93ddb9ebea" containerName="registry-server" probeResult="failure" output=< Feb 27 18:09:33 crc kubenswrapper[4700]: timeout: failed to connect service ":50051" within 1s Feb 27 18:09:33 crc kubenswrapper[4700]: > Feb 27 18:09:42 crc kubenswrapper[4700]: I0227 18:09:42.206370 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6hvnd" Feb 27 18:09:42 crc kubenswrapper[4700]: I0227 18:09:42.298175 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6hvnd" Feb 27 18:09:42 crc kubenswrapper[4700]: I0227 18:09:42.458702 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6hvnd"] Feb 27 18:09:43 crc kubenswrapper[4700]: I0227 18:09:43.603539 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6hvnd" podUID="ea15dc00-097e-4356-8178-ac93ddb9ebea" containerName="registry-server" containerID="cri-o://7ce54e47f84b8e705c23cb198be3d58a5bc2d8f4c1e6194df6e1a3007edfac88" gracePeriod=2 Feb 27 18:09:44 crc kubenswrapper[4700]: I0227 18:09:44.152985 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6hvnd" Feb 27 18:09:44 crc kubenswrapper[4700]: I0227 18:09:44.185274 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea15dc00-097e-4356-8178-ac93ddb9ebea-utilities\") pod \"ea15dc00-097e-4356-8178-ac93ddb9ebea\" (UID: \"ea15dc00-097e-4356-8178-ac93ddb9ebea\") " Feb 27 18:09:44 crc kubenswrapper[4700]: I0227 18:09:44.185482 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rdxw2\" (UniqueName: \"kubernetes.io/projected/ea15dc00-097e-4356-8178-ac93ddb9ebea-kube-api-access-rdxw2\") pod \"ea15dc00-097e-4356-8178-ac93ddb9ebea\" (UID: \"ea15dc00-097e-4356-8178-ac93ddb9ebea\") " Feb 27 18:09:44 crc kubenswrapper[4700]: I0227 18:09:44.185709 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea15dc00-097e-4356-8178-ac93ddb9ebea-catalog-content\") pod \"ea15dc00-097e-4356-8178-ac93ddb9ebea\" (UID: \"ea15dc00-097e-4356-8178-ac93ddb9ebea\") " Feb 27 18:09:44 crc kubenswrapper[4700]: I0227 18:09:44.187707 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea15dc00-097e-4356-8178-ac93ddb9ebea-utilities" (OuterVolumeSpecName: "utilities") pod "ea15dc00-097e-4356-8178-ac93ddb9ebea" (UID: "ea15dc00-097e-4356-8178-ac93ddb9ebea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 18:09:44 crc kubenswrapper[4700]: I0227 18:09:44.193848 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea15dc00-097e-4356-8178-ac93ddb9ebea-kube-api-access-rdxw2" (OuterVolumeSpecName: "kube-api-access-rdxw2") pod "ea15dc00-097e-4356-8178-ac93ddb9ebea" (UID: "ea15dc00-097e-4356-8178-ac93ddb9ebea"). InnerVolumeSpecName "kube-api-access-rdxw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:09:44 crc kubenswrapper[4700]: I0227 18:09:44.348703 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea15dc00-097e-4356-8178-ac93ddb9ebea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ea15dc00-097e-4356-8178-ac93ddb9ebea" (UID: "ea15dc00-097e-4356-8178-ac93ddb9ebea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 18:09:44 crc kubenswrapper[4700]: I0227 18:09:44.351294 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rdxw2\" (UniqueName: \"kubernetes.io/projected/ea15dc00-097e-4356-8178-ac93ddb9ebea-kube-api-access-rdxw2\") on node \"crc\" DevicePath \"\"" Feb 27 18:09:44 crc kubenswrapper[4700]: I0227 18:09:44.351313 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea15dc00-097e-4356-8178-ac93ddb9ebea-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 18:09:44 crc kubenswrapper[4700]: I0227 18:09:44.351321 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea15dc00-097e-4356-8178-ac93ddb9ebea-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 18:09:44 crc kubenswrapper[4700]: I0227 18:09:44.621759 4700 generic.go:334] "Generic (PLEG): container finished" podID="ea15dc00-097e-4356-8178-ac93ddb9ebea" containerID="7ce54e47f84b8e705c23cb198be3d58a5bc2d8f4c1e6194df6e1a3007edfac88" exitCode=0 Feb 27 18:09:44 crc kubenswrapper[4700]: I0227 18:09:44.621890 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6hvnd" event={"ID":"ea15dc00-097e-4356-8178-ac93ddb9ebea","Type":"ContainerDied","Data":"7ce54e47f84b8e705c23cb198be3d58a5bc2d8f4c1e6194df6e1a3007edfac88"} Feb 27 18:09:44 crc kubenswrapper[4700]: I0227 18:09:44.622281 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6hvnd" event={"ID":"ea15dc00-097e-4356-8178-ac93ddb9ebea","Type":"ContainerDied","Data":"fa207fd5b22c11b8eacede81b339fc65760c65b4f593a50a0ad6b67fd1f36ac7"} Feb 27 18:09:44 crc kubenswrapper[4700]: I0227 18:09:44.621912 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6hvnd" Feb 27 18:09:44 crc kubenswrapper[4700]: I0227 18:09:44.622347 4700 scope.go:117] "RemoveContainer" containerID="7ce54e47f84b8e705c23cb198be3d58a5bc2d8f4c1e6194df6e1a3007edfac88" Feb 27 18:09:44 crc kubenswrapper[4700]: I0227 18:09:44.679678 4700 scope.go:117] "RemoveContainer" containerID="0d1769ac82fd0dae453ce862c137b761e1484d90b8c0e56ded1ac857f6c122e1" Feb 27 18:09:44 crc kubenswrapper[4700]: I0227 18:09:44.695694 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6hvnd"] Feb 27 18:09:44 crc kubenswrapper[4700]: I0227 18:09:44.717587 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6hvnd"] Feb 27 18:09:44 crc kubenswrapper[4700]: I0227 18:09:44.720349 4700 scope.go:117] "RemoveContainer" containerID="2275bd601d299407c983290efeaa194993070f46db036dc7b24d94606395c99f" Feb 27 18:09:44 crc kubenswrapper[4700]: I0227 18:09:44.784143 4700 scope.go:117] "RemoveContainer" containerID="7ce54e47f84b8e705c23cb198be3d58a5bc2d8f4c1e6194df6e1a3007edfac88" Feb 27 18:09:44 crc kubenswrapper[4700]: E0227 18:09:44.784863 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ce54e47f84b8e705c23cb198be3d58a5bc2d8f4c1e6194df6e1a3007edfac88\": container with ID starting with 7ce54e47f84b8e705c23cb198be3d58a5bc2d8f4c1e6194df6e1a3007edfac88 not found: ID does not exist" containerID="7ce54e47f84b8e705c23cb198be3d58a5bc2d8f4c1e6194df6e1a3007edfac88" Feb 27 18:09:44 crc kubenswrapper[4700]: I0227 18:09:44.784937 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ce54e47f84b8e705c23cb198be3d58a5bc2d8f4c1e6194df6e1a3007edfac88"} err="failed to get container status \"7ce54e47f84b8e705c23cb198be3d58a5bc2d8f4c1e6194df6e1a3007edfac88\": rpc error: code = NotFound desc = could not find container \"7ce54e47f84b8e705c23cb198be3d58a5bc2d8f4c1e6194df6e1a3007edfac88\": container with ID starting with 7ce54e47f84b8e705c23cb198be3d58a5bc2d8f4c1e6194df6e1a3007edfac88 not found: ID does not exist" Feb 27 18:09:44 crc kubenswrapper[4700]: I0227 18:09:44.784983 4700 scope.go:117] "RemoveContainer" containerID="0d1769ac82fd0dae453ce862c137b761e1484d90b8c0e56ded1ac857f6c122e1" Feb 27 18:09:44 crc kubenswrapper[4700]: E0227 18:09:44.785771 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d1769ac82fd0dae453ce862c137b761e1484d90b8c0e56ded1ac857f6c122e1\": container with ID starting with 0d1769ac82fd0dae453ce862c137b761e1484d90b8c0e56ded1ac857f6c122e1 not found: ID does not exist" containerID="0d1769ac82fd0dae453ce862c137b761e1484d90b8c0e56ded1ac857f6c122e1" Feb 27 18:09:44 crc kubenswrapper[4700]: I0227 18:09:44.785831 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d1769ac82fd0dae453ce862c137b761e1484d90b8c0e56ded1ac857f6c122e1"} err="failed to get container status \"0d1769ac82fd0dae453ce862c137b761e1484d90b8c0e56ded1ac857f6c122e1\": rpc error: code = NotFound desc = could not find container \"0d1769ac82fd0dae453ce862c137b761e1484d90b8c0e56ded1ac857f6c122e1\": container with ID starting with 0d1769ac82fd0dae453ce862c137b761e1484d90b8c0e56ded1ac857f6c122e1 not found: ID does not exist" Feb 27 18:09:44 crc kubenswrapper[4700]: I0227 18:09:44.785881 4700 scope.go:117] "RemoveContainer" containerID="2275bd601d299407c983290efeaa194993070f46db036dc7b24d94606395c99f" Feb 27 18:09:44 crc kubenswrapper[4700]: E0227 18:09:44.786488 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2275bd601d299407c983290efeaa194993070f46db036dc7b24d94606395c99f\": container with ID starting with 2275bd601d299407c983290efeaa194993070f46db036dc7b24d94606395c99f not found: ID does not exist" containerID="2275bd601d299407c983290efeaa194993070f46db036dc7b24d94606395c99f" Feb 27 18:09:44 crc kubenswrapper[4700]: I0227 18:09:44.786683 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2275bd601d299407c983290efeaa194993070f46db036dc7b24d94606395c99f"} err="failed to get container status \"2275bd601d299407c983290efeaa194993070f46db036dc7b24d94606395c99f\": rpc error: code = NotFound desc = could not find container \"2275bd601d299407c983290efeaa194993070f46db036dc7b24d94606395c99f\": container with ID starting with 2275bd601d299407c983290efeaa194993070f46db036dc7b24d94606395c99f not found: ID does not exist" Feb 27 18:09:45 crc kubenswrapper[4700]: I0227 18:09:45.000744 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea15dc00-097e-4356-8178-ac93ddb9ebea" path="/var/lib/kubelet/pods/ea15dc00-097e-4356-8178-ac93ddb9ebea/volumes" Feb 27 18:09:46 crc kubenswrapper[4700]: E0227 18:09:46.985217 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:10:00 crc kubenswrapper[4700]: I0227 18:10:00.180608 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536930-xbg7f"] Feb 27 18:10:00 crc kubenswrapper[4700]: E0227 18:10:00.182801 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea15dc00-097e-4356-8178-ac93ddb9ebea" containerName="extract-content" Feb 27 18:10:00 crc kubenswrapper[4700]: I0227 18:10:00.182856 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea15dc00-097e-4356-8178-ac93ddb9ebea" containerName="extract-content" Feb 27 18:10:00 crc kubenswrapper[4700]: E0227 18:10:00.182909 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea15dc00-097e-4356-8178-ac93ddb9ebea" containerName="registry-server" Feb 27 18:10:00 crc kubenswrapper[4700]: I0227 18:10:00.182923 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea15dc00-097e-4356-8178-ac93ddb9ebea" containerName="registry-server" Feb 27 18:10:00 crc kubenswrapper[4700]: E0227 18:10:00.182989 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea15dc00-097e-4356-8178-ac93ddb9ebea" containerName="extract-utilities" Feb 27 18:10:00 crc kubenswrapper[4700]: I0227 18:10:00.183005 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea15dc00-097e-4356-8178-ac93ddb9ebea" containerName="extract-utilities" Feb 27 18:10:00 crc kubenswrapper[4700]: I0227 18:10:00.183433 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea15dc00-097e-4356-8178-ac93ddb9ebea" containerName="registry-server" Feb 27 18:10:00 crc kubenswrapper[4700]: I0227 18:10:00.185131 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" Feb 27 18:10:00 crc kubenswrapper[4700]: I0227 18:10:00.192555 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536930-xbg7f"] Feb 27 18:10:00 crc kubenswrapper[4700]: I0227 18:10:00.228078 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtm8j\" (UniqueName: \"kubernetes.io/projected/3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5-kube-api-access-gtm8j\") pod \"auto-csr-approver-29536930-xbg7f\" (UID: \"3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5\") " pod="openshift-infra/auto-csr-approver-29536930-xbg7f" Feb 27 18:10:00 crc kubenswrapper[4700]: I0227 18:10:00.331808 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtm8j\" (UniqueName: \"kubernetes.io/projected/3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5-kube-api-access-gtm8j\") pod \"auto-csr-approver-29536930-xbg7f\" (UID: \"3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5\") " pod="openshift-infra/auto-csr-approver-29536930-xbg7f" Feb 27 18:10:00 crc kubenswrapper[4700]: I0227 18:10:00.368869 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtm8j\" (UniqueName: \"kubernetes.io/projected/3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5-kube-api-access-gtm8j\") pod \"auto-csr-approver-29536930-xbg7f\" (UID: \"3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5\") " pod="openshift-infra/auto-csr-approver-29536930-xbg7f" Feb 27 18:10:00 crc kubenswrapper[4700]: I0227 18:10:00.515407 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" Feb 27 18:10:01 crc kubenswrapper[4700]: I0227 18:10:01.061038 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536930-xbg7f"] Feb 27 18:10:01 crc kubenswrapper[4700]: I0227 18:10:01.886176 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" event={"ID":"3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5","Type":"ContainerStarted","Data":"e20847a166818207490b96d0764bc0a4028117728ae13315b61ff1d7d2895c27"} Feb 27 18:10:01 crc kubenswrapper[4700]: E0227 18:10:01.988097 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:10:02 crc kubenswrapper[4700]: E0227 18:10:02.093131 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 18:10:02 crc kubenswrapper[4700]: E0227 18:10:02.093876 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 18:10:02 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 18:10:02 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gtm8j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536930-xbg7f_openshift-infra(3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 18:10:02 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 18:10:02 crc kubenswrapper[4700]: E0227 18:10:02.097440 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:10:02 crc kubenswrapper[4700]: E0227 18:10:02.904629 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:10:06 crc kubenswrapper[4700]: I0227 18:10:06.410530 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 18:10:06 crc kubenswrapper[4700]: I0227 18:10:06.412034 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 18:10:12 crc kubenswrapper[4700]: E0227 18:10:12.988510 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:10:20 crc kubenswrapper[4700]: E0227 18:10:20.005211 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 18:10:20 crc kubenswrapper[4700]: E0227 18:10:20.006114 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 18:10:20 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 18:10:20 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gtm8j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536930-xbg7f_openshift-infra(3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 18:10:20 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 18:10:20 crc kubenswrapper[4700]: E0227 18:10:20.007640 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:10:23 crc kubenswrapper[4700]: E0227 18:10:23.984709 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:10:32 crc kubenswrapper[4700]: E0227 18:10:32.985031 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:10:34 crc kubenswrapper[4700]: E0227 18:10:34.984610 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:10:36 crc kubenswrapper[4700]: I0227 18:10:36.410300 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 18:10:36 crc kubenswrapper[4700]: I0227 18:10:36.410811 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 18:10:44 crc kubenswrapper[4700]: E0227 18:10:44.921609 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 18:10:44 crc kubenswrapper[4700]: E0227 18:10:44.922713 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 18:10:44 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 18:10:44 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gtm8j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536930-xbg7f_openshift-infra(3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 18:10:44 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 18:10:44 crc kubenswrapper[4700]: E0227 18:10:44.923981 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:10:49 crc kubenswrapper[4700]: E0227 18:10:49.986119 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:10:57 crc kubenswrapper[4700]: E0227 18:10:57.990423 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:11:02 crc kubenswrapper[4700]: E0227 18:11:02.984548 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:11:06 crc kubenswrapper[4700]: I0227 18:11:06.411056 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 18:11:06 crc kubenswrapper[4700]: I0227 18:11:06.411588 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 18:11:06 crc kubenswrapper[4700]: I0227 18:11:06.411657 4700 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" Feb 27 18:11:06 crc kubenswrapper[4700]: I0227 18:11:06.412907 4700 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a22d07b860150538bb9582eddfae339a5c84757100045eb131b89e75822dfe43"} pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 18:11:06 crc kubenswrapper[4700]: I0227 18:11:06.413013 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" containerID="cri-o://a22d07b860150538bb9582eddfae339a5c84757100045eb131b89e75822dfe43" gracePeriod=600 Feb 27 18:11:06 crc kubenswrapper[4700]: I0227 18:11:06.769592 4700 generic.go:334] "Generic (PLEG): container finished" podID="84b45511-b94c-479f-98db-fd2c4eceec46" containerID="a22d07b860150538bb9582eddfae339a5c84757100045eb131b89e75822dfe43" exitCode=0 Feb 27 18:11:06 crc kubenswrapper[4700]: I0227 18:11:06.769664 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerDied","Data":"a22d07b860150538bb9582eddfae339a5c84757100045eb131b89e75822dfe43"} Feb 27 18:11:06 crc kubenswrapper[4700]: I0227 18:11:06.769719 4700 scope.go:117] "RemoveContainer" containerID="5496fef1f08034cab735e8cc9a87daee6229da7616d276a174410ba8eacb94a6" Feb 27 18:11:07 crc kubenswrapper[4700]: I0227 18:11:07.790055 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerStarted","Data":"454da4538d721f5f5da4ce85346adc3fbf77394a4e27f80b9bc979aa3bc37aa2"} Feb 27 18:11:11 crc kubenswrapper[4700]: E0227 18:11:11.983720 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:11:12 crc kubenswrapper[4700]: I0227 18:11:12.873499 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9q6kw"] Feb 27 18:11:12 crc kubenswrapper[4700]: I0227 18:11:12.877724 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9q6kw" Feb 27 18:11:12 crc kubenswrapper[4700]: I0227 18:11:12.893692 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9q6kw"] Feb 27 18:11:12 crc kubenswrapper[4700]: I0227 18:11:12.916615 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxzcv\" (UniqueName: \"kubernetes.io/projected/44c7aa20-a94a-4ba6-b812-02664ab56dd7-kube-api-access-xxzcv\") pod \"redhat-marketplace-9q6kw\" (UID: \"44c7aa20-a94a-4ba6-b812-02664ab56dd7\") " pod="openshift-marketplace/redhat-marketplace-9q6kw" Feb 27 18:11:12 crc kubenswrapper[4700]: I0227 18:11:12.916711 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44c7aa20-a94a-4ba6-b812-02664ab56dd7-utilities\") pod \"redhat-marketplace-9q6kw\" (UID: \"44c7aa20-a94a-4ba6-b812-02664ab56dd7\") " pod="openshift-marketplace/redhat-marketplace-9q6kw" Feb 27 18:11:12 crc kubenswrapper[4700]: I0227 18:11:12.916771 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44c7aa20-a94a-4ba6-b812-02664ab56dd7-catalog-content\") pod \"redhat-marketplace-9q6kw\" (UID: \"44c7aa20-a94a-4ba6-b812-02664ab56dd7\") " pod="openshift-marketplace/redhat-marketplace-9q6kw" Feb 27 18:11:13 crc kubenswrapper[4700]: I0227 18:11:13.018403 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxzcv\" (UniqueName: \"kubernetes.io/projected/44c7aa20-a94a-4ba6-b812-02664ab56dd7-kube-api-access-xxzcv\") pod \"redhat-marketplace-9q6kw\" (UID: \"44c7aa20-a94a-4ba6-b812-02664ab56dd7\") " pod="openshift-marketplace/redhat-marketplace-9q6kw" Feb 27 18:11:13 crc kubenswrapper[4700]: I0227 18:11:13.018537 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44c7aa20-a94a-4ba6-b812-02664ab56dd7-utilities\") pod \"redhat-marketplace-9q6kw\" (UID: \"44c7aa20-a94a-4ba6-b812-02664ab56dd7\") " pod="openshift-marketplace/redhat-marketplace-9q6kw" Feb 27 18:11:13 crc kubenswrapper[4700]: I0227 18:11:13.018596 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44c7aa20-a94a-4ba6-b812-02664ab56dd7-catalog-content\") pod \"redhat-marketplace-9q6kw\" (UID: \"44c7aa20-a94a-4ba6-b812-02664ab56dd7\") " pod="openshift-marketplace/redhat-marketplace-9q6kw" Feb 27 18:11:13 crc kubenswrapper[4700]: I0227 18:11:13.019170 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44c7aa20-a94a-4ba6-b812-02664ab56dd7-catalog-content\") pod \"redhat-marketplace-9q6kw\" (UID: \"44c7aa20-a94a-4ba6-b812-02664ab56dd7\") " pod="openshift-marketplace/redhat-marketplace-9q6kw" Feb 27 18:11:13 crc kubenswrapper[4700]: I0227 18:11:13.019293 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44c7aa20-a94a-4ba6-b812-02664ab56dd7-utilities\") pod \"redhat-marketplace-9q6kw\" (UID: \"44c7aa20-a94a-4ba6-b812-02664ab56dd7\") " pod="openshift-marketplace/redhat-marketplace-9q6kw" Feb 27 18:11:13 crc kubenswrapper[4700]: I0227 18:11:13.040829 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxzcv\" (UniqueName: \"kubernetes.io/projected/44c7aa20-a94a-4ba6-b812-02664ab56dd7-kube-api-access-xxzcv\") pod \"redhat-marketplace-9q6kw\" (UID: \"44c7aa20-a94a-4ba6-b812-02664ab56dd7\") " pod="openshift-marketplace/redhat-marketplace-9q6kw" Feb 27 18:11:13 crc kubenswrapper[4700]: I0227 18:11:13.208203 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9q6kw" Feb 27 18:11:13 crc kubenswrapper[4700]: I0227 18:11:13.725748 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9q6kw"] Feb 27 18:11:13 crc kubenswrapper[4700]: I0227 18:11:13.862923 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9q6kw" event={"ID":"44c7aa20-a94a-4ba6-b812-02664ab56dd7","Type":"ContainerStarted","Data":"7fb25313ad910b37ac9698925f739845da39df2deb7ce70c85b512e464349086"} Feb 27 18:11:14 crc kubenswrapper[4700]: I0227 18:11:14.878739 4700 generic.go:334] "Generic (PLEG): container finished" podID="44c7aa20-a94a-4ba6-b812-02664ab56dd7" containerID="e560726ec2e795250dce312c50ae5623ba52cf10ad9795e8c65f463fe9f34b92" exitCode=0 Feb 27 18:11:14 crc kubenswrapper[4700]: I0227 18:11:14.878879 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9q6kw" event={"ID":"44c7aa20-a94a-4ba6-b812-02664ab56dd7","Type":"ContainerDied","Data":"e560726ec2e795250dce312c50ae5623ba52cf10ad9795e8c65f463fe9f34b92"} Feb 27 18:11:14 crc kubenswrapper[4700]: I0227 18:11:14.882084 4700 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 18:11:15 crc kubenswrapper[4700]: E0227 18:11:15.716236 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 18:11:15 crc kubenswrapper[4700]: E0227 18:11:15.717677 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xxzcv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-9q6kw_openshift-marketplace(44c7aa20-a94a-4ba6-b812-02664ab56dd7): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 18:11:15 crc kubenswrapper[4700]: E0227 18:11:15.719126 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-marketplace-9q6kw" podUID="44c7aa20-a94a-4ba6-b812-02664ab56dd7" Feb 27 18:11:15 crc kubenswrapper[4700]: E0227 18:11:15.893799 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-9q6kw" podUID="44c7aa20-a94a-4ba6-b812-02664ab56dd7" Feb 27 18:11:17 crc kubenswrapper[4700]: E0227 18:11:17.984192 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:11:27 crc kubenswrapper[4700]: E0227 18:11:27.190777 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 18:11:27 crc kubenswrapper[4700]: E0227 18:11:27.191640 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 18:11:27 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 18:11:27 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gtm8j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536930-xbg7f_openshift-infra(3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 18:11:27 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 18:11:27 crc kubenswrapper[4700]: E0227 18:11:27.192858 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:11:28 crc kubenswrapper[4700]: E0227 18:11:28.605255 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 18:11:28 crc kubenswrapper[4700]: E0227 18:11:28.605855 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xxzcv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-9q6kw_openshift-marketplace(44c7aa20-a94a-4ba6-b812-02664ab56dd7): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 18:11:28 crc kubenswrapper[4700]: E0227 18:11:28.607092 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-marketplace-9q6kw" podUID="44c7aa20-a94a-4ba6-b812-02664ab56dd7" Feb 27 18:11:31 crc kubenswrapper[4700]: E0227 18:11:31.986093 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:11:33 crc kubenswrapper[4700]: I0227 18:11:33.020883 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-d86g8"] Feb 27 18:11:33 crc kubenswrapper[4700]: I0227 18:11:33.023842 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d86g8" Feb 27 18:11:33 crc kubenswrapper[4700]: I0227 18:11:33.055352 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d86g8"] Feb 27 18:11:33 crc kubenswrapper[4700]: I0227 18:11:33.174206 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/243afaab-17c1-4631-974c-f7b693236b08-catalog-content\") pod \"certified-operators-d86g8\" (UID: \"243afaab-17c1-4631-974c-f7b693236b08\") " pod="openshift-marketplace/certified-operators-d86g8" Feb 27 18:11:33 crc kubenswrapper[4700]: I0227 18:11:33.174625 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkwzw\" (UniqueName: \"kubernetes.io/projected/243afaab-17c1-4631-974c-f7b693236b08-kube-api-access-vkwzw\") pod \"certified-operators-d86g8\" (UID: \"243afaab-17c1-4631-974c-f7b693236b08\") " pod="openshift-marketplace/certified-operators-d86g8" Feb 27 18:11:33 crc kubenswrapper[4700]: I0227 18:11:33.174774 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/243afaab-17c1-4631-974c-f7b693236b08-utilities\") pod \"certified-operators-d86g8\" (UID: \"243afaab-17c1-4631-974c-f7b693236b08\") " pod="openshift-marketplace/certified-operators-d86g8" Feb 27 18:11:33 crc kubenswrapper[4700]: I0227 18:11:33.277497 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkwzw\" (UniqueName: \"kubernetes.io/projected/243afaab-17c1-4631-974c-f7b693236b08-kube-api-access-vkwzw\") pod \"certified-operators-d86g8\" (UID: \"243afaab-17c1-4631-974c-f7b693236b08\") " pod="openshift-marketplace/certified-operators-d86g8" Feb 27 18:11:33 crc kubenswrapper[4700]: I0227 18:11:33.277554 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/243afaab-17c1-4631-974c-f7b693236b08-utilities\") pod \"certified-operators-d86g8\" (UID: \"243afaab-17c1-4631-974c-f7b693236b08\") " pod="openshift-marketplace/certified-operators-d86g8" Feb 27 18:11:33 crc kubenswrapper[4700]: I0227 18:11:33.277746 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/243afaab-17c1-4631-974c-f7b693236b08-catalog-content\") pod \"certified-operators-d86g8\" (UID: \"243afaab-17c1-4631-974c-f7b693236b08\") " pod="openshift-marketplace/certified-operators-d86g8" Feb 27 18:11:33 crc kubenswrapper[4700]: I0227 18:11:33.278271 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/243afaab-17c1-4631-974c-f7b693236b08-catalog-content\") pod \"certified-operators-d86g8\" (UID: \"243afaab-17c1-4631-974c-f7b693236b08\") " pod="openshift-marketplace/certified-operators-d86g8" Feb 27 18:11:33 crc kubenswrapper[4700]: I0227 18:11:33.279221 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/243afaab-17c1-4631-974c-f7b693236b08-utilities\") pod \"certified-operators-d86g8\" (UID: \"243afaab-17c1-4631-974c-f7b693236b08\") " pod="openshift-marketplace/certified-operators-d86g8" Feb 27 18:11:33 crc kubenswrapper[4700]: I0227 18:11:33.320493 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkwzw\" (UniqueName: \"kubernetes.io/projected/243afaab-17c1-4631-974c-f7b693236b08-kube-api-access-vkwzw\") pod \"certified-operators-d86g8\" (UID: \"243afaab-17c1-4631-974c-f7b693236b08\") " pod="openshift-marketplace/certified-operators-d86g8" Feb 27 18:11:33 crc kubenswrapper[4700]: I0227 18:11:33.360549 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d86g8" Feb 27 18:11:33 crc kubenswrapper[4700]: I0227 18:11:33.862927 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d86g8"] Feb 27 18:11:33 crc kubenswrapper[4700]: W0227 18:11:33.876663 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod243afaab_17c1_4631_974c_f7b693236b08.slice/crio-16257cbd86f69bef2cc6b3c9a37f67f6da2159e9766be7be5d802dee309affd4 WatchSource:0}: Error finding container 16257cbd86f69bef2cc6b3c9a37f67f6da2159e9766be7be5d802dee309affd4: Status 404 returned error can't find the container with id 16257cbd86f69bef2cc6b3c9a37f67f6da2159e9766be7be5d802dee309affd4 Feb 27 18:11:34 crc kubenswrapper[4700]: I0227 18:11:34.168522 4700 generic.go:334] "Generic (PLEG): container finished" podID="243afaab-17c1-4631-974c-f7b693236b08" containerID="3b8324dc261d79c510a5a6fe4498764a5ccfedc2a55e39d88d50dce728b771f4" exitCode=0 Feb 27 18:11:34 crc kubenswrapper[4700]: I0227 18:11:34.168602 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d86g8" event={"ID":"243afaab-17c1-4631-974c-f7b693236b08","Type":"ContainerDied","Data":"3b8324dc261d79c510a5a6fe4498764a5ccfedc2a55e39d88d50dce728b771f4"} Feb 27 18:11:34 crc kubenswrapper[4700]: I0227 18:11:34.168853 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d86g8" event={"ID":"243afaab-17c1-4631-974c-f7b693236b08","Type":"ContainerStarted","Data":"16257cbd86f69bef2cc6b3c9a37f67f6da2159e9766be7be5d802dee309affd4"} Feb 27 18:11:35 crc kubenswrapper[4700]: E0227 18:11:35.082171 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 27 18:11:35 crc kubenswrapper[4700]: E0227 18:11:35.083261 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vkwzw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-d86g8_openshift-marketplace(243afaab-17c1-4631-974c-f7b693236b08): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 18:11:35 crc kubenswrapper[4700]: E0227 18:11:35.084795 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:11:35 crc kubenswrapper[4700]: E0227 18:11:35.185764 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:11:39 crc kubenswrapper[4700]: E0227 18:11:39.984532 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:11:42 crc kubenswrapper[4700]: E0227 18:11:42.984760 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-9q6kw" podUID="44c7aa20-a94a-4ba6-b812-02664ab56dd7" Feb 27 18:11:44 crc kubenswrapper[4700]: E0227 18:11:44.988969 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:11:47 crc kubenswrapper[4700]: E0227 18:11:47.912789 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 27 18:11:47 crc kubenswrapper[4700]: E0227 18:11:47.913778 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vkwzw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-d86g8_openshift-marketplace(243afaab-17c1-4631-974c-f7b693236b08): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 18:11:47 crc kubenswrapper[4700]: E0227 18:11:47.915045 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:11:51 crc kubenswrapper[4700]: E0227 18:11:51.985389 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:11:55 crc kubenswrapper[4700]: E0227 18:11:55.575614 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 18:11:55 crc kubenswrapper[4700]: E0227 18:11:55.576382 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xxzcv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-9q6kw_openshift-marketplace(44c7aa20-a94a-4ba6-b812-02664ab56dd7): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 18:11:55 crc kubenswrapper[4700]: E0227 18:11:55.577612 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-marketplace-9q6kw" podUID="44c7aa20-a94a-4ba6-b812-02664ab56dd7" Feb 27 18:12:00 crc kubenswrapper[4700]: E0227 18:12:00.008910 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 18:12:00 crc kubenswrapper[4700]: E0227 18:12:00.010667 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 18:12:00 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 18:12:00 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hjvdh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536926-s6ft8_openshift-infra(d37cbaaf-1dee-4efd-9327-7bee80986a35): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 18:12:00 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 18:12:00 crc kubenswrapper[4700]: E0227 18:12:00.011820 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:12:00 crc kubenswrapper[4700]: I0227 18:12:00.167891 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536932-l4s4p"] Feb 27 18:12:00 crc kubenswrapper[4700]: I0227 18:12:00.169786 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536932-l4s4p" Feb 27 18:12:00 crc kubenswrapper[4700]: I0227 18:12:00.185019 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536932-l4s4p"] Feb 27 18:12:00 crc kubenswrapper[4700]: I0227 18:12:00.238226 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fbxd\" (UniqueName: \"kubernetes.io/projected/3d8e8a24-2862-4552-82c1-3b445fb6628d-kube-api-access-4fbxd\") pod \"auto-csr-approver-29536932-l4s4p\" (UID: \"3d8e8a24-2862-4552-82c1-3b445fb6628d\") " pod="openshift-infra/auto-csr-approver-29536932-l4s4p" Feb 27 18:12:00 crc kubenswrapper[4700]: I0227 18:12:00.340164 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fbxd\" (UniqueName: \"kubernetes.io/projected/3d8e8a24-2862-4552-82c1-3b445fb6628d-kube-api-access-4fbxd\") pod \"auto-csr-approver-29536932-l4s4p\" (UID: \"3d8e8a24-2862-4552-82c1-3b445fb6628d\") " pod="openshift-infra/auto-csr-approver-29536932-l4s4p" Feb 27 18:12:00 crc kubenswrapper[4700]: I0227 18:12:00.399773 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fbxd\" (UniqueName: \"kubernetes.io/projected/3d8e8a24-2862-4552-82c1-3b445fb6628d-kube-api-access-4fbxd\") pod \"auto-csr-approver-29536932-l4s4p\" (UID: \"3d8e8a24-2862-4552-82c1-3b445fb6628d\") " pod="openshift-infra/auto-csr-approver-29536932-l4s4p" Feb 27 18:12:00 crc kubenswrapper[4700]: I0227 18:12:00.508956 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536932-l4s4p" Feb 27 18:12:00 crc kubenswrapper[4700]: E0227 18:12:00.995182 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:12:01 crc kubenswrapper[4700]: I0227 18:12:01.050422 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536932-l4s4p"] Feb 27 18:12:01 crc kubenswrapper[4700]: I0227 18:12:01.511218 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536932-l4s4p" event={"ID":"3d8e8a24-2862-4552-82c1-3b445fb6628d","Type":"ContainerStarted","Data":"7eb597ce10446d6ecc60323b0198b8c9ef4b0a79a8cddea61be7ea7af248cf90"} Feb 27 18:12:02 crc kubenswrapper[4700]: E0227 18:12:02.077417 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 18:12:02 crc kubenswrapper[4700]: E0227 18:12:02.077656 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 18:12:02 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 18:12:02 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4fbxd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536932-l4s4p_openshift-infra(3d8e8a24-2862-4552-82c1-3b445fb6628d): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 18:12:02 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 18:12:02 crc kubenswrapper[4700]: E0227 18:12:02.078961 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536932-l4s4p" podUID="3d8e8a24-2862-4552-82c1-3b445fb6628d" Feb 27 18:12:02 crc kubenswrapper[4700]: E0227 18:12:02.525741 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536932-l4s4p" podUID="3d8e8a24-2862-4552-82c1-3b445fb6628d" Feb 27 18:12:02 crc kubenswrapper[4700]: E0227 18:12:02.983982 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:12:07 crc kubenswrapper[4700]: E0227 18:12:07.985566 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-9q6kw" podUID="44c7aa20-a94a-4ba6-b812-02664ab56dd7" Feb 27 18:12:13 crc kubenswrapper[4700]: E0227 18:12:13.984295 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:12:14 crc kubenswrapper[4700]: E0227 18:12:14.982865 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:12:15 crc kubenswrapper[4700]: E0227 18:12:15.050961 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 18:12:15 crc kubenswrapper[4700]: E0227 18:12:15.051188 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 18:12:15 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 18:12:15 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4fbxd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536932-l4s4p_openshift-infra(3d8e8a24-2862-4552-82c1-3b445fb6628d): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 18:12:15 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 18:12:15 crc kubenswrapper[4700]: E0227 18:12:15.052370 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536932-l4s4p" podUID="3d8e8a24-2862-4552-82c1-3b445fb6628d" Feb 27 18:12:16 crc kubenswrapper[4700]: E0227 18:12:16.593066 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 27 18:12:16 crc kubenswrapper[4700]: E0227 18:12:16.593837 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vkwzw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-d86g8_openshift-marketplace(243afaab-17c1-4631-974c-f7b693236b08): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 18:12:16 crc kubenswrapper[4700]: E0227 18:12:16.595202 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:12:18 crc kubenswrapper[4700]: E0227 18:12:18.986379 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-9q6kw" podUID="44c7aa20-a94a-4ba6-b812-02664ab56dd7" Feb 27 18:12:27 crc kubenswrapper[4700]: E0227 18:12:27.985640 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:12:28 crc kubenswrapper[4700]: E0227 18:12:28.990678 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536932-l4s4p" podUID="3d8e8a24-2862-4552-82c1-3b445fb6628d" Feb 27 18:12:28 crc kubenswrapper[4700]: E0227 18:12:28.992354 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:12:29 crc kubenswrapper[4700]: E0227 18:12:29.983820 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:12:31 crc kubenswrapper[4700]: E0227 18:12:31.986067 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-9q6kw" podUID="44c7aa20-a94a-4ba6-b812-02664ab56dd7" Feb 27 18:12:39 crc kubenswrapper[4700]: E0227 18:12:39.989072 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:12:41 crc kubenswrapper[4700]: E0227 18:12:41.008025 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:12:42 crc kubenswrapper[4700]: E0227 18:12:42.992217 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:12:45 crc kubenswrapper[4700]: E0227 18:12:45.561673 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 18:12:45 crc kubenswrapper[4700]: E0227 18:12:45.562736 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xxzcv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-9q6kw_openshift-marketplace(44c7aa20-a94a-4ba6-b812-02664ab56dd7): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 18:12:45 crc kubenswrapper[4700]: E0227 18:12:45.564021 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-marketplace-9q6kw" podUID="44c7aa20-a94a-4ba6-b812-02664ab56dd7" Feb 27 18:12:46 crc kubenswrapper[4700]: I0227 18:12:46.075388 4700 generic.go:334] "Generic (PLEG): container finished" podID="3d8e8a24-2862-4552-82c1-3b445fb6628d" containerID="d81845e950d957edbe8fade1a8eb3a28a30ca213b35a6ec2f6568db90f5b452c" exitCode=0 Feb 27 18:12:46 crc kubenswrapper[4700]: I0227 18:12:46.075537 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536932-l4s4p" event={"ID":"3d8e8a24-2862-4552-82c1-3b445fb6628d","Type":"ContainerDied","Data":"d81845e950d957edbe8fade1a8eb3a28a30ca213b35a6ec2f6568db90f5b452c"} Feb 27 18:12:47 crc kubenswrapper[4700]: I0227 18:12:47.536339 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536932-l4s4p" Feb 27 18:12:47 crc kubenswrapper[4700]: I0227 18:12:47.617959 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fbxd\" (UniqueName: \"kubernetes.io/projected/3d8e8a24-2862-4552-82c1-3b445fb6628d-kube-api-access-4fbxd\") pod \"3d8e8a24-2862-4552-82c1-3b445fb6628d\" (UID: \"3d8e8a24-2862-4552-82c1-3b445fb6628d\") " Feb 27 18:12:47 crc kubenswrapper[4700]: I0227 18:12:47.625351 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d8e8a24-2862-4552-82c1-3b445fb6628d-kube-api-access-4fbxd" (OuterVolumeSpecName: "kube-api-access-4fbxd") pod "3d8e8a24-2862-4552-82c1-3b445fb6628d" (UID: "3d8e8a24-2862-4552-82c1-3b445fb6628d"). InnerVolumeSpecName "kube-api-access-4fbxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:12:47 crc kubenswrapper[4700]: I0227 18:12:47.720906 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fbxd\" (UniqueName: \"kubernetes.io/projected/3d8e8a24-2862-4552-82c1-3b445fb6628d-kube-api-access-4fbxd\") on node \"crc\" DevicePath \"\"" Feb 27 18:12:48 crc kubenswrapper[4700]: I0227 18:12:48.115156 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536932-l4s4p" event={"ID":"3d8e8a24-2862-4552-82c1-3b445fb6628d","Type":"ContainerDied","Data":"7eb597ce10446d6ecc60323b0198b8c9ef4b0a79a8cddea61be7ea7af248cf90"} Feb 27 18:12:48 crc kubenswrapper[4700]: I0227 18:12:48.115221 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7eb597ce10446d6ecc60323b0198b8c9ef4b0a79a8cddea61be7ea7af248cf90" Feb 27 18:12:48 crc kubenswrapper[4700]: I0227 18:12:48.115315 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536932-l4s4p" Feb 27 18:12:48 crc kubenswrapper[4700]: I0227 18:12:48.658414 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536922-d7lzf"] Feb 27 18:12:48 crc kubenswrapper[4700]: I0227 18:12:48.674033 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536922-d7lzf"] Feb 27 18:12:49 crc kubenswrapper[4700]: I0227 18:12:49.003257 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="793f7988-e7d0-4b3b-a35f-7d0445ebab61" path="/var/lib/kubelet/pods/793f7988-e7d0-4b3b-a35f-7d0445ebab61/volumes" Feb 27 18:12:52 crc kubenswrapper[4700]: E0227 18:12:52.985012 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:12:54 crc kubenswrapper[4700]: E0227 18:12:54.984541 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:12:58 crc kubenswrapper[4700]: E0227 18:12:58.924842 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 18:12:58 crc kubenswrapper[4700]: E0227 18:12:58.925632 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 18:12:58 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 18:12:58 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gtm8j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536930-xbg7f_openshift-infra(3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 18:12:58 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 18:12:58 crc kubenswrapper[4700]: E0227 18:12:58.926859 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:12:58 crc kubenswrapper[4700]: E0227 18:12:58.984436 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-9q6kw" podUID="44c7aa20-a94a-4ba6-b812-02664ab56dd7" Feb 27 18:13:06 crc kubenswrapper[4700]: I0227 18:13:06.410800 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 18:13:06 crc kubenswrapper[4700]: I0227 18:13:06.411811 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 18:13:08 crc kubenswrapper[4700]: E0227 18:13:08.652894 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 27 18:13:08 crc kubenswrapper[4700]: E0227 18:13:08.653407 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vkwzw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-d86g8_openshift-marketplace(243afaab-17c1-4631-974c-f7b693236b08): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 18:13:08 crc kubenswrapper[4700]: E0227 18:13:08.655096 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:13:08 crc kubenswrapper[4700]: E0227 18:13:08.984015 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:13:09 crc kubenswrapper[4700]: E0227 18:13:09.987546 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:13:12 crc kubenswrapper[4700]: I0227 18:13:12.061241 4700 scope.go:117] "RemoveContainer" containerID="4a3e7ec9b3459eb0394533a5ebdf6e05a419a0ff52f797f7120460349a9a59cb" Feb 27 18:13:12 crc kubenswrapper[4700]: E0227 18:13:12.983982 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-9q6kw" podUID="44c7aa20-a94a-4ba6-b812-02664ab56dd7" Feb 27 18:13:19 crc kubenswrapper[4700]: E0227 18:13:19.984296 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:13:20 crc kubenswrapper[4700]: E0227 18:13:20.996801 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:13:23 crc kubenswrapper[4700]: E0227 18:13:23.984864 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:13:25 crc kubenswrapper[4700]: E0227 18:13:25.984912 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-9q6kw" podUID="44c7aa20-a94a-4ba6-b812-02664ab56dd7" Feb 27 18:13:31 crc kubenswrapper[4700]: E0227 18:13:31.005344 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:13:33 crc kubenswrapper[4700]: E0227 18:13:33.983923 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:13:35 crc kubenswrapper[4700]: E0227 18:13:35.984518 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:13:36 crc kubenswrapper[4700]: I0227 18:13:36.410809 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 18:13:36 crc kubenswrapper[4700]: I0227 18:13:36.410921 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 18:13:37 crc kubenswrapper[4700]: E0227 18:13:37.984396 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-9q6kw" podUID="44c7aa20-a94a-4ba6-b812-02664ab56dd7" Feb 27 18:13:44 crc kubenswrapper[4700]: E0227 18:13:44.986752 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:13:47 crc kubenswrapper[4700]: E0227 18:13:47.984083 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:13:48 crc kubenswrapper[4700]: E0227 18:13:48.983781 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-9q6kw" podUID="44c7aa20-a94a-4ba6-b812-02664ab56dd7" Feb 27 18:13:51 crc kubenswrapper[4700]: E0227 18:13:50.999362 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:13:58 crc kubenswrapper[4700]: E0227 18:13:58.984044 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:14:00 crc kubenswrapper[4700]: I0227 18:14:00.175609 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536934-rw77f"] Feb 27 18:14:00 crc kubenswrapper[4700]: E0227 18:14:00.176946 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d8e8a24-2862-4552-82c1-3b445fb6628d" containerName="oc" Feb 27 18:14:00 crc kubenswrapper[4700]: I0227 18:14:00.176980 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d8e8a24-2862-4552-82c1-3b445fb6628d" containerName="oc" Feb 27 18:14:00 crc kubenswrapper[4700]: I0227 18:14:00.177568 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d8e8a24-2862-4552-82c1-3b445fb6628d" containerName="oc" Feb 27 18:14:00 crc kubenswrapper[4700]: I0227 18:14:00.179177 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536934-rw77f" Feb 27 18:14:00 crc kubenswrapper[4700]: I0227 18:14:00.194195 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536934-rw77f"] Feb 27 18:14:00 crc kubenswrapper[4700]: I0227 18:14:00.293583 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkjfz\" (UniqueName: \"kubernetes.io/projected/caa4e57d-b02d-4d65-944d-451b8294f884-kube-api-access-qkjfz\") pod \"auto-csr-approver-29536934-rw77f\" (UID: \"caa4e57d-b02d-4d65-944d-451b8294f884\") " pod="openshift-infra/auto-csr-approver-29536934-rw77f" Feb 27 18:14:00 crc kubenswrapper[4700]: I0227 18:14:00.396102 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkjfz\" (UniqueName: \"kubernetes.io/projected/caa4e57d-b02d-4d65-944d-451b8294f884-kube-api-access-qkjfz\") pod \"auto-csr-approver-29536934-rw77f\" (UID: \"caa4e57d-b02d-4d65-944d-451b8294f884\") " pod="openshift-infra/auto-csr-approver-29536934-rw77f" Feb 27 18:14:00 crc kubenswrapper[4700]: I0227 18:14:00.443661 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkjfz\" (UniqueName: \"kubernetes.io/projected/caa4e57d-b02d-4d65-944d-451b8294f884-kube-api-access-qkjfz\") pod \"auto-csr-approver-29536934-rw77f\" (UID: \"caa4e57d-b02d-4d65-944d-451b8294f884\") " pod="openshift-infra/auto-csr-approver-29536934-rw77f" Feb 27 18:14:00 crc kubenswrapper[4700]: I0227 18:14:00.517413 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536934-rw77f" Feb 27 18:14:01 crc kubenswrapper[4700]: E0227 18:14:01.003908 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-9q6kw" podUID="44c7aa20-a94a-4ba6-b812-02664ab56dd7" Feb 27 18:14:01 crc kubenswrapper[4700]: I0227 18:14:01.037592 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536934-rw77f"] Feb 27 18:14:01 crc kubenswrapper[4700]: E0227 18:14:01.838802 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 18:14:01 crc kubenswrapper[4700]: E0227 18:14:01.838995 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 18:14:01 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 18:14:01 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qkjfz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536934-rw77f_openshift-infra(caa4e57d-b02d-4d65-944d-451b8294f884): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 18:14:01 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 18:14:01 crc kubenswrapper[4700]: E0227 18:14:01.840254 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:14:02 crc kubenswrapper[4700]: I0227 18:14:02.053264 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536934-rw77f" event={"ID":"caa4e57d-b02d-4d65-944d-451b8294f884","Type":"ContainerStarted","Data":"8462629da14118f9f26d64718724c34cf1a6253798bba5deabaac00477a43029"} Feb 27 18:14:02 crc kubenswrapper[4700]: E0227 18:14:02.056024 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:14:02 crc kubenswrapper[4700]: E0227 18:14:02.984569 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:14:03 crc kubenswrapper[4700]: E0227 18:14:03.069012 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:14:03 crc kubenswrapper[4700]: E0227 18:14:03.985542 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:14:06 crc kubenswrapper[4700]: I0227 18:14:06.411099 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 18:14:06 crc kubenswrapper[4700]: I0227 18:14:06.411393 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 18:14:06 crc kubenswrapper[4700]: I0227 18:14:06.411452 4700 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" Feb 27 18:14:06 crc kubenswrapper[4700]: I0227 18:14:06.412319 4700 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"454da4538d721f5f5da4ce85346adc3fbf77394a4e27f80b9bc979aa3bc37aa2"} pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 18:14:06 crc kubenswrapper[4700]: I0227 18:14:06.412385 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" containerID="cri-o://454da4538d721f5f5da4ce85346adc3fbf77394a4e27f80b9bc979aa3bc37aa2" gracePeriod=600 Feb 27 18:14:06 crc kubenswrapper[4700]: E0227 18:14:06.536944 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:14:07 crc kubenswrapper[4700]: I0227 18:14:07.128323 4700 generic.go:334] "Generic (PLEG): container finished" podID="84b45511-b94c-479f-98db-fd2c4eceec46" containerID="454da4538d721f5f5da4ce85346adc3fbf77394a4e27f80b9bc979aa3bc37aa2" exitCode=0 Feb 27 18:14:07 crc kubenswrapper[4700]: I0227 18:14:07.128414 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerDied","Data":"454da4538d721f5f5da4ce85346adc3fbf77394a4e27f80b9bc979aa3bc37aa2"} Feb 27 18:14:07 crc kubenswrapper[4700]: I0227 18:14:07.128972 4700 scope.go:117] "RemoveContainer" containerID="a22d07b860150538bb9582eddfae339a5c84757100045eb131b89e75822dfe43" Feb 27 18:14:07 crc kubenswrapper[4700]: I0227 18:14:07.130704 4700 scope.go:117] "RemoveContainer" containerID="454da4538d721f5f5da4ce85346adc3fbf77394a4e27f80b9bc979aa3bc37aa2" Feb 27 18:14:07 crc kubenswrapper[4700]: E0227 18:14:07.132210 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:14:09 crc kubenswrapper[4700]: E0227 18:14:09.985307 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:14:13 crc kubenswrapper[4700]: I0227 18:14:13.213714 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9q6kw" event={"ID":"44c7aa20-a94a-4ba6-b812-02664ab56dd7","Type":"ContainerStarted","Data":"06f61cd576fc8fc27bddf971a4d8b1114ef4c6c2d8a0fd0d5dc20fd986dad4a0"} Feb 27 18:14:14 crc kubenswrapper[4700]: I0227 18:14:14.226769 4700 generic.go:334] "Generic (PLEG): container finished" podID="44c7aa20-a94a-4ba6-b812-02664ab56dd7" containerID="06f61cd576fc8fc27bddf971a4d8b1114ef4c6c2d8a0fd0d5dc20fd986dad4a0" exitCode=0 Feb 27 18:14:14 crc kubenswrapper[4700]: I0227 18:14:14.226860 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9q6kw" event={"ID":"44c7aa20-a94a-4ba6-b812-02664ab56dd7","Type":"ContainerDied","Data":"06f61cd576fc8fc27bddf971a4d8b1114ef4c6c2d8a0fd0d5dc20fd986dad4a0"} Feb 27 18:14:15 crc kubenswrapper[4700]: I0227 18:14:15.245412 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9q6kw" event={"ID":"44c7aa20-a94a-4ba6-b812-02664ab56dd7","Type":"ContainerStarted","Data":"dda8398936162799828b0e421212db8b85a7c7e19e955e57a1117b5743a58f0b"} Feb 27 18:14:15 crc kubenswrapper[4700]: I0227 18:14:15.272344 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9q6kw" podStartSLOduration=3.484978729 podStartE2EDuration="3m3.272326566s" podCreationTimestamp="2026-02-27 18:11:12 +0000 UTC" firstStartedPulling="2026-02-27 18:11:14.881765757 +0000 UTC m=+4234.867078504" lastFinishedPulling="2026-02-27 18:14:14.669113554 +0000 UTC m=+4414.654426341" observedRunningTime="2026-02-27 18:14:15.27016161 +0000 UTC m=+4415.255474357" watchObservedRunningTime="2026-02-27 18:14:15.272326566 +0000 UTC m=+4415.257639313" Feb 27 18:14:15 crc kubenswrapper[4700]: E0227 18:14:15.986281 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:14:16 crc kubenswrapper[4700]: E0227 18:14:16.980140 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 18:14:16 crc kubenswrapper[4700]: E0227 18:14:16.980342 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 18:14:16 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 18:14:16 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qkjfz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536934-rw77f_openshift-infra(caa4e57d-b02d-4d65-944d-451b8294f884): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 18:14:16 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 18:14:16 crc kubenswrapper[4700]: E0227 18:14:16.981696 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:14:17 crc kubenswrapper[4700]: E0227 18:14:17.984094 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:14:19 crc kubenswrapper[4700]: I0227 18:14:19.982273 4700 scope.go:117] "RemoveContainer" containerID="454da4538d721f5f5da4ce85346adc3fbf77394a4e27f80b9bc979aa3bc37aa2" Feb 27 18:14:19 crc kubenswrapper[4700]: E0227 18:14:19.983159 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:14:21 crc kubenswrapper[4700]: E0227 18:14:21.985370 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:14:23 crc kubenswrapper[4700]: I0227 18:14:23.208436 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9q6kw" Feb 27 18:14:23 crc kubenswrapper[4700]: I0227 18:14:23.208559 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9q6kw" Feb 27 18:14:23 crc kubenswrapper[4700]: I0227 18:14:23.288227 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9q6kw" Feb 27 18:14:23 crc kubenswrapper[4700]: I0227 18:14:23.402635 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9q6kw" Feb 27 18:14:23 crc kubenswrapper[4700]: I0227 18:14:23.530065 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9q6kw"] Feb 27 18:14:25 crc kubenswrapper[4700]: I0227 18:14:25.379065 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9q6kw" podUID="44c7aa20-a94a-4ba6-b812-02664ab56dd7" containerName="registry-server" containerID="cri-o://dda8398936162799828b0e421212db8b85a7c7e19e955e57a1117b5743a58f0b" gracePeriod=2 Feb 27 18:14:25 crc kubenswrapper[4700]: I0227 18:14:25.927436 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9q6kw" Feb 27 18:14:25 crc kubenswrapper[4700]: I0227 18:14:25.987042 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44c7aa20-a94a-4ba6-b812-02664ab56dd7-catalog-content\") pod \"44c7aa20-a94a-4ba6-b812-02664ab56dd7\" (UID: \"44c7aa20-a94a-4ba6-b812-02664ab56dd7\") " Feb 27 18:14:25 crc kubenswrapper[4700]: I0227 18:14:25.987086 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxzcv\" (UniqueName: \"kubernetes.io/projected/44c7aa20-a94a-4ba6-b812-02664ab56dd7-kube-api-access-xxzcv\") pod \"44c7aa20-a94a-4ba6-b812-02664ab56dd7\" (UID: \"44c7aa20-a94a-4ba6-b812-02664ab56dd7\") " Feb 27 18:14:25 crc kubenswrapper[4700]: I0227 18:14:25.987252 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44c7aa20-a94a-4ba6-b812-02664ab56dd7-utilities\") pod \"44c7aa20-a94a-4ba6-b812-02664ab56dd7\" (UID: \"44c7aa20-a94a-4ba6-b812-02664ab56dd7\") " Feb 27 18:14:25 crc kubenswrapper[4700]: I0227 18:14:25.989085 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44c7aa20-a94a-4ba6-b812-02664ab56dd7-utilities" (OuterVolumeSpecName: "utilities") pod "44c7aa20-a94a-4ba6-b812-02664ab56dd7" (UID: "44c7aa20-a94a-4ba6-b812-02664ab56dd7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 18:14:26 crc kubenswrapper[4700]: I0227 18:14:26.011509 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44c7aa20-a94a-4ba6-b812-02664ab56dd7-kube-api-access-xxzcv" (OuterVolumeSpecName: "kube-api-access-xxzcv") pod "44c7aa20-a94a-4ba6-b812-02664ab56dd7" (UID: "44c7aa20-a94a-4ba6-b812-02664ab56dd7"). InnerVolumeSpecName "kube-api-access-xxzcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:14:26 crc kubenswrapper[4700]: I0227 18:14:26.037948 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44c7aa20-a94a-4ba6-b812-02664ab56dd7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "44c7aa20-a94a-4ba6-b812-02664ab56dd7" (UID: "44c7aa20-a94a-4ba6-b812-02664ab56dd7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 18:14:26 crc kubenswrapper[4700]: I0227 18:14:26.089622 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44c7aa20-a94a-4ba6-b812-02664ab56dd7-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 18:14:26 crc kubenswrapper[4700]: I0227 18:14:26.091342 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44c7aa20-a94a-4ba6-b812-02664ab56dd7-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 18:14:26 crc kubenswrapper[4700]: I0227 18:14:26.091358 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxzcv\" (UniqueName: \"kubernetes.io/projected/44c7aa20-a94a-4ba6-b812-02664ab56dd7-kube-api-access-xxzcv\") on node \"crc\" DevicePath \"\"" Feb 27 18:14:26 crc kubenswrapper[4700]: I0227 18:14:26.391731 4700 generic.go:334] "Generic (PLEG): container finished" podID="44c7aa20-a94a-4ba6-b812-02664ab56dd7" containerID="dda8398936162799828b0e421212db8b85a7c7e19e955e57a1117b5743a58f0b" exitCode=0 Feb 27 18:14:26 crc kubenswrapper[4700]: I0227 18:14:26.391789 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9q6kw" event={"ID":"44c7aa20-a94a-4ba6-b812-02664ab56dd7","Type":"ContainerDied","Data":"dda8398936162799828b0e421212db8b85a7c7e19e955e57a1117b5743a58f0b"} Feb 27 18:14:26 crc kubenswrapper[4700]: I0227 18:14:26.391824 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9q6kw" event={"ID":"44c7aa20-a94a-4ba6-b812-02664ab56dd7","Type":"ContainerDied","Data":"7fb25313ad910b37ac9698925f739845da39df2deb7ce70c85b512e464349086"} Feb 27 18:14:26 crc kubenswrapper[4700]: I0227 18:14:26.391850 4700 scope.go:117] "RemoveContainer" containerID="dda8398936162799828b0e421212db8b85a7c7e19e955e57a1117b5743a58f0b" Feb 27 18:14:26 crc kubenswrapper[4700]: I0227 18:14:26.391845 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9q6kw" Feb 27 18:14:26 crc kubenswrapper[4700]: I0227 18:14:26.421040 4700 scope.go:117] "RemoveContainer" containerID="06f61cd576fc8fc27bddf971a4d8b1114ef4c6c2d8a0fd0d5dc20fd986dad4a0" Feb 27 18:14:26 crc kubenswrapper[4700]: I0227 18:14:26.432877 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9q6kw"] Feb 27 18:14:26 crc kubenswrapper[4700]: I0227 18:14:26.453051 4700 scope.go:117] "RemoveContainer" containerID="e560726ec2e795250dce312c50ae5623ba52cf10ad9795e8c65f463fe9f34b92" Feb 27 18:14:26 crc kubenswrapper[4700]: I0227 18:14:26.456303 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9q6kw"] Feb 27 18:14:26 crc kubenswrapper[4700]: I0227 18:14:26.519560 4700 scope.go:117] "RemoveContainer" containerID="dda8398936162799828b0e421212db8b85a7c7e19e955e57a1117b5743a58f0b" Feb 27 18:14:26 crc kubenswrapper[4700]: E0227 18:14:26.520064 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dda8398936162799828b0e421212db8b85a7c7e19e955e57a1117b5743a58f0b\": container with ID starting with dda8398936162799828b0e421212db8b85a7c7e19e955e57a1117b5743a58f0b not found: ID does not exist" containerID="dda8398936162799828b0e421212db8b85a7c7e19e955e57a1117b5743a58f0b" Feb 27 18:14:26 crc kubenswrapper[4700]: I0227 18:14:26.520121 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dda8398936162799828b0e421212db8b85a7c7e19e955e57a1117b5743a58f0b"} err="failed to get container status \"dda8398936162799828b0e421212db8b85a7c7e19e955e57a1117b5743a58f0b\": rpc error: code = NotFound desc = could not find container \"dda8398936162799828b0e421212db8b85a7c7e19e955e57a1117b5743a58f0b\": container with ID starting with dda8398936162799828b0e421212db8b85a7c7e19e955e57a1117b5743a58f0b not found: ID does not exist" Feb 27 18:14:26 crc kubenswrapper[4700]: I0227 18:14:26.520153 4700 scope.go:117] "RemoveContainer" containerID="06f61cd576fc8fc27bddf971a4d8b1114ef4c6c2d8a0fd0d5dc20fd986dad4a0" Feb 27 18:14:26 crc kubenswrapper[4700]: E0227 18:14:26.520790 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06f61cd576fc8fc27bddf971a4d8b1114ef4c6c2d8a0fd0d5dc20fd986dad4a0\": container with ID starting with 06f61cd576fc8fc27bddf971a4d8b1114ef4c6c2d8a0fd0d5dc20fd986dad4a0 not found: ID does not exist" containerID="06f61cd576fc8fc27bddf971a4d8b1114ef4c6c2d8a0fd0d5dc20fd986dad4a0" Feb 27 18:14:26 crc kubenswrapper[4700]: I0227 18:14:26.520827 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06f61cd576fc8fc27bddf971a4d8b1114ef4c6c2d8a0fd0d5dc20fd986dad4a0"} err="failed to get container status \"06f61cd576fc8fc27bddf971a4d8b1114ef4c6c2d8a0fd0d5dc20fd986dad4a0\": rpc error: code = NotFound desc = could not find container \"06f61cd576fc8fc27bddf971a4d8b1114ef4c6c2d8a0fd0d5dc20fd986dad4a0\": container with ID starting with 06f61cd576fc8fc27bddf971a4d8b1114ef4c6c2d8a0fd0d5dc20fd986dad4a0 not found: ID does not exist" Feb 27 18:14:26 crc kubenswrapper[4700]: I0227 18:14:26.520849 4700 scope.go:117] "RemoveContainer" containerID="e560726ec2e795250dce312c50ae5623ba52cf10ad9795e8c65f463fe9f34b92" Feb 27 18:14:26 crc kubenswrapper[4700]: E0227 18:14:26.521135 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e560726ec2e795250dce312c50ae5623ba52cf10ad9795e8c65f463fe9f34b92\": container with ID starting with e560726ec2e795250dce312c50ae5623ba52cf10ad9795e8c65f463fe9f34b92 not found: ID does not exist" containerID="e560726ec2e795250dce312c50ae5623ba52cf10ad9795e8c65f463fe9f34b92" Feb 27 18:14:26 crc kubenswrapper[4700]: I0227 18:14:26.521160 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e560726ec2e795250dce312c50ae5623ba52cf10ad9795e8c65f463fe9f34b92"} err="failed to get container status \"e560726ec2e795250dce312c50ae5623ba52cf10ad9795e8c65f463fe9f34b92\": rpc error: code = NotFound desc = could not find container \"e560726ec2e795250dce312c50ae5623ba52cf10ad9795e8c65f463fe9f34b92\": container with ID starting with e560726ec2e795250dce312c50ae5623ba52cf10ad9795e8c65f463fe9f34b92 not found: ID does not exist" Feb 27 18:14:27 crc kubenswrapper[4700]: I0227 18:14:27.005082 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44c7aa20-a94a-4ba6-b812-02664ab56dd7" path="/var/lib/kubelet/pods/44c7aa20-a94a-4ba6-b812-02664ab56dd7/volumes" Feb 27 18:14:27 crc kubenswrapper[4700]: E0227 18:14:27.984956 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:14:31 crc kubenswrapper[4700]: E0227 18:14:30.998747 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:14:31 crc kubenswrapper[4700]: E0227 18:14:30.999895 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:14:33 crc kubenswrapper[4700]: I0227 18:14:33.982581 4700 scope.go:117] "RemoveContainer" containerID="454da4538d721f5f5da4ce85346adc3fbf77394a4e27f80b9bc979aa3bc37aa2" Feb 27 18:14:33 crc kubenswrapper[4700]: E0227 18:14:33.983633 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:14:36 crc kubenswrapper[4700]: E0227 18:14:36.773455 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 27 18:14:36 crc kubenswrapper[4700]: E0227 18:14:36.774229 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vkwzw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-d86g8_openshift-marketplace(243afaab-17c1-4631-974c-f7b693236b08): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 18:14:36 crc kubenswrapper[4700]: E0227 18:14:36.775964 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:14:38 crc kubenswrapper[4700]: E0227 18:14:38.986922 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:14:44 crc kubenswrapper[4700]: I0227 18:14:44.403200 4700 patch_prober.go:28] interesting pod/oauth-openshift-7d9bbcf4d4-6g7cs container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.67:6443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 18:14:44 crc kubenswrapper[4700]: I0227 18:14:44.403990 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-6g7cs" podUID="04191b5b-0e24-4490-a6aa-b9cda6b6c8d3" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.67:6443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 27 18:14:45 crc kubenswrapper[4700]: I0227 18:14:45.983535 4700 scope.go:117] "RemoveContainer" containerID="454da4538d721f5f5da4ce85346adc3fbf77394a4e27f80b9bc979aa3bc37aa2" Feb 27 18:14:45 crc kubenswrapper[4700]: E0227 18:14:45.983914 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:14:45 crc kubenswrapper[4700]: E0227 18:14:45.995225 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:14:47 crc kubenswrapper[4700]: E0227 18:14:47.019866 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 18:14:47 crc kubenswrapper[4700]: E0227 18:14:47.020099 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 18:14:47 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 18:14:47 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qkjfz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536934-rw77f_openshift-infra(caa4e57d-b02d-4d65-944d-451b8294f884): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 18:14:47 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 18:14:47 crc kubenswrapper[4700]: E0227 18:14:47.021356 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:14:49 crc kubenswrapper[4700]: E0227 18:14:49.985390 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:14:51 crc kubenswrapper[4700]: E0227 18:14:51.984984 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:14:56 crc kubenswrapper[4700]: I0227 18:14:56.981824 4700 scope.go:117] "RemoveContainer" containerID="454da4538d721f5f5da4ce85346adc3fbf77394a4e27f80b9bc979aa3bc37aa2" Feb 27 18:14:56 crc kubenswrapper[4700]: E0227 18:14:56.983055 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:14:57 crc kubenswrapper[4700]: E0227 18:14:57.986268 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:14:58 crc kubenswrapper[4700]: E0227 18:14:58.985589 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:15:00 crc kubenswrapper[4700]: I0227 18:15:00.157091 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536935-lk8rn"] Feb 27 18:15:00 crc kubenswrapper[4700]: E0227 18:15:00.158100 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44c7aa20-a94a-4ba6-b812-02664ab56dd7" containerName="registry-server" Feb 27 18:15:00 crc kubenswrapper[4700]: I0227 18:15:00.158122 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="44c7aa20-a94a-4ba6-b812-02664ab56dd7" containerName="registry-server" Feb 27 18:15:00 crc kubenswrapper[4700]: E0227 18:15:00.158142 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44c7aa20-a94a-4ba6-b812-02664ab56dd7" containerName="extract-content" Feb 27 18:15:00 crc kubenswrapper[4700]: I0227 18:15:00.158149 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="44c7aa20-a94a-4ba6-b812-02664ab56dd7" containerName="extract-content" Feb 27 18:15:00 crc kubenswrapper[4700]: E0227 18:15:00.158167 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44c7aa20-a94a-4ba6-b812-02664ab56dd7" containerName="extract-utilities" Feb 27 18:15:00 crc kubenswrapper[4700]: I0227 18:15:00.158176 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="44c7aa20-a94a-4ba6-b812-02664ab56dd7" containerName="extract-utilities" Feb 27 18:15:00 crc kubenswrapper[4700]: I0227 18:15:00.158401 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="44c7aa20-a94a-4ba6-b812-02664ab56dd7" containerName="registry-server" Feb 27 18:15:00 crc kubenswrapper[4700]: I0227 18:15:00.159374 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536935-lk8rn" Feb 27 18:15:00 crc kubenswrapper[4700]: I0227 18:15:00.162527 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 27 18:15:00 crc kubenswrapper[4700]: I0227 18:15:00.175789 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 27 18:15:00 crc kubenswrapper[4700]: I0227 18:15:00.175843 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536935-lk8rn"] Feb 27 18:15:00 crc kubenswrapper[4700]: I0227 18:15:00.256748 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb-secret-volume\") pod \"collect-profiles-29536935-lk8rn\" (UID: \"e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536935-lk8rn" Feb 27 18:15:00 crc kubenswrapper[4700]: I0227 18:15:00.256939 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4sm7\" (UniqueName: \"kubernetes.io/projected/e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb-kube-api-access-p4sm7\") pod \"collect-profiles-29536935-lk8rn\" (UID: \"e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536935-lk8rn" Feb 27 18:15:00 crc kubenswrapper[4700]: I0227 18:15:00.257434 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb-config-volume\") pod \"collect-profiles-29536935-lk8rn\" (UID: \"e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536935-lk8rn" Feb 27 18:15:00 crc kubenswrapper[4700]: I0227 18:15:00.360639 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb-config-volume\") pod \"collect-profiles-29536935-lk8rn\" (UID: \"e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536935-lk8rn" Feb 27 18:15:00 crc kubenswrapper[4700]: I0227 18:15:00.360757 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb-secret-volume\") pod \"collect-profiles-29536935-lk8rn\" (UID: \"e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536935-lk8rn" Feb 27 18:15:00 crc kubenswrapper[4700]: I0227 18:15:00.360937 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4sm7\" (UniqueName: \"kubernetes.io/projected/e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb-kube-api-access-p4sm7\") pod \"collect-profiles-29536935-lk8rn\" (UID: \"e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536935-lk8rn" Feb 27 18:15:00 crc kubenswrapper[4700]: I0227 18:15:00.362573 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb-config-volume\") pod \"collect-profiles-29536935-lk8rn\" (UID: \"e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536935-lk8rn" Feb 27 18:15:00 crc kubenswrapper[4700]: I0227 18:15:00.381898 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb-secret-volume\") pod \"collect-profiles-29536935-lk8rn\" (UID: \"e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536935-lk8rn" Feb 27 18:15:00 crc kubenswrapper[4700]: I0227 18:15:00.387772 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4sm7\" (UniqueName: \"kubernetes.io/projected/e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb-kube-api-access-p4sm7\") pod \"collect-profiles-29536935-lk8rn\" (UID: \"e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536935-lk8rn" Feb 27 18:15:00 crc kubenswrapper[4700]: I0227 18:15:00.487342 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536935-lk8rn" Feb 27 18:15:01 crc kubenswrapper[4700]: W0227 18:15:01.047868 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode2b1edd3_d2e0_4f31_876f_0c7e11f4cfcb.slice/crio-cedf8dc8b44625ff12294a925cc29b109097c1cdbe98c073ad65c071c9f79976 WatchSource:0}: Error finding container cedf8dc8b44625ff12294a925cc29b109097c1cdbe98c073ad65c071c9f79976: Status 404 returned error can't find the container with id cedf8dc8b44625ff12294a925cc29b109097c1cdbe98c073ad65c071c9f79976 Feb 27 18:15:01 crc kubenswrapper[4700]: I0227 18:15:01.052266 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536935-lk8rn"] Feb 27 18:15:01 crc kubenswrapper[4700]: I0227 18:15:01.870588 4700 generic.go:334] "Generic (PLEG): container finished" podID="e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb" containerID="143c9dd982d09d833f5a75b487e7a70cd409e26b547337a52dd4d5e4d56c6e2d" exitCode=0 Feb 27 18:15:01 crc kubenswrapper[4700]: I0227 18:15:01.870704 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536935-lk8rn" event={"ID":"e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb","Type":"ContainerDied","Data":"143c9dd982d09d833f5a75b487e7a70cd409e26b547337a52dd4d5e4d56c6e2d"} Feb 27 18:15:01 crc kubenswrapper[4700]: I0227 18:15:01.871103 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536935-lk8rn" event={"ID":"e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb","Type":"ContainerStarted","Data":"cedf8dc8b44625ff12294a925cc29b109097c1cdbe98c073ad65c071c9f79976"} Feb 27 18:15:02 crc kubenswrapper[4700]: E0227 18:15:02.985476 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:15:03 crc kubenswrapper[4700]: I0227 18:15:03.384258 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536935-lk8rn" Feb 27 18:15:03 crc kubenswrapper[4700]: I0227 18:15:03.576028 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4sm7\" (UniqueName: \"kubernetes.io/projected/e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb-kube-api-access-p4sm7\") pod \"e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb\" (UID: \"e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb\") " Feb 27 18:15:03 crc kubenswrapper[4700]: I0227 18:15:03.576147 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb-config-volume\") pod \"e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb\" (UID: \"e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb\") " Feb 27 18:15:03 crc kubenswrapper[4700]: I0227 18:15:03.576401 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb-secret-volume\") pod \"e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb\" (UID: \"e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb\") " Feb 27 18:15:03 crc kubenswrapper[4700]: I0227 18:15:03.576935 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb-config-volume" (OuterVolumeSpecName: "config-volume") pod "e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb" (UID: "e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 18:15:03 crc kubenswrapper[4700]: I0227 18:15:03.577359 4700 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb-config-volume\") on node \"crc\" DevicePath \"\"" Feb 27 18:15:03 crc kubenswrapper[4700]: I0227 18:15:03.583392 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb" (UID: "e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 18:15:03 crc kubenswrapper[4700]: I0227 18:15:03.585658 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb-kube-api-access-p4sm7" (OuterVolumeSpecName: "kube-api-access-p4sm7") pod "e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb" (UID: "e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb"). InnerVolumeSpecName "kube-api-access-p4sm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:15:03 crc kubenswrapper[4700]: I0227 18:15:03.679600 4700 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 27 18:15:03 crc kubenswrapper[4700]: I0227 18:15:03.679636 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4sm7\" (UniqueName: \"kubernetes.io/projected/e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb-kube-api-access-p4sm7\") on node \"crc\" DevicePath \"\"" Feb 27 18:15:03 crc kubenswrapper[4700]: I0227 18:15:03.902083 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536935-lk8rn" event={"ID":"e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb","Type":"ContainerDied","Data":"cedf8dc8b44625ff12294a925cc29b109097c1cdbe98c073ad65c071c9f79976"} Feb 27 18:15:03 crc kubenswrapper[4700]: I0227 18:15:03.902159 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cedf8dc8b44625ff12294a925cc29b109097c1cdbe98c073ad65c071c9f79976" Feb 27 18:15:03 crc kubenswrapper[4700]: I0227 18:15:03.902161 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536935-lk8rn" Feb 27 18:15:03 crc kubenswrapper[4700]: E0227 18:15:03.985180 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:15:04 crc kubenswrapper[4700]: I0227 18:15:04.505557 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536890-7xlw5"] Feb 27 18:15:04 crc kubenswrapper[4700]: I0227 18:15:04.516878 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536890-7xlw5"] Feb 27 18:15:05 crc kubenswrapper[4700]: I0227 18:15:05.009920 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7207687-fb96-4b11-8c3b-c9f9467e0373" path="/var/lib/kubelet/pods/a7207687-fb96-4b11-8c3b-c9f9467e0373/volumes" Feb 27 18:15:07 crc kubenswrapper[4700]: I0227 18:15:07.982761 4700 scope.go:117] "RemoveContainer" containerID="454da4538d721f5f5da4ce85346adc3fbf77394a4e27f80b9bc979aa3bc37aa2" Feb 27 18:15:07 crc kubenswrapper[4700]: E0227 18:15:07.984139 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:15:11 crc kubenswrapper[4700]: E0227 18:15:10.999006 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:15:11 crc kubenswrapper[4700]: E0227 18:15:11.000706 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:15:12 crc kubenswrapper[4700]: I0227 18:15:12.219593 4700 scope.go:117] "RemoveContainer" containerID="512b48dbe4a24e5db07e29406ee97abf45ae51a6445f58b4122ad3375c0a48a9" Feb 27 18:15:13 crc kubenswrapper[4700]: E0227 18:15:13.984251 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:15:17 crc kubenswrapper[4700]: E0227 18:15:17.985845 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:15:19 crc kubenswrapper[4700]: I0227 18:15:19.982111 4700 scope.go:117] "RemoveContainer" containerID="454da4538d721f5f5da4ce85346adc3fbf77394a4e27f80b9bc979aa3bc37aa2" Feb 27 18:15:19 crc kubenswrapper[4700]: E0227 18:15:19.982929 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:15:22 crc kubenswrapper[4700]: E0227 18:15:22.985913 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:15:23 crc kubenswrapper[4700]: E0227 18:15:23.984237 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:15:28 crc kubenswrapper[4700]: E0227 18:15:28.987413 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:15:28 crc kubenswrapper[4700]: E0227 18:15:28.987491 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:15:31 crc kubenswrapper[4700]: I0227 18:15:31.981257 4700 scope.go:117] "RemoveContainer" containerID="454da4538d721f5f5da4ce85346adc3fbf77394a4e27f80b9bc979aa3bc37aa2" Feb 27 18:15:31 crc kubenswrapper[4700]: E0227 18:15:31.982548 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:15:34 crc kubenswrapper[4700]: E0227 18:15:34.986485 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:15:39 crc kubenswrapper[4700]: E0227 18:15:39.005203 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 18:15:39 crc kubenswrapper[4700]: E0227 18:15:39.006227 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 18:15:39 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 18:15:39 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qkjfz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536934-rw77f_openshift-infra(caa4e57d-b02d-4d65-944d-451b8294f884): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 18:15:39 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 18:15:39 crc kubenswrapper[4700]: E0227 18:15:39.007756 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:15:39 crc kubenswrapper[4700]: I0227 18:15:39.749297 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="f2325e29-9f4c-438f-9400-e643e2494808" containerName="galera" probeResult="failure" output="command timed out" Feb 27 18:15:39 crc kubenswrapper[4700]: I0227 18:15:39.749311 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="f2325e29-9f4c-438f-9400-e643e2494808" containerName="galera" probeResult="failure" output="command timed out" Feb 27 18:15:39 crc kubenswrapper[4700]: E0227 18:15:39.982712 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:15:39 crc kubenswrapper[4700]: E0227 18:15:39.984571 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:15:43 crc kubenswrapper[4700]: I0227 18:15:43.982134 4700 scope.go:117] "RemoveContainer" containerID="454da4538d721f5f5da4ce85346adc3fbf77394a4e27f80b9bc979aa3bc37aa2" Feb 27 18:15:43 crc kubenswrapper[4700]: E0227 18:15:43.983499 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:15:44 crc kubenswrapper[4700]: E0227 18:15:44.356326 4700 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.222:49664->38.102.83.222:41369: write tcp 38.102.83.222:49664->38.102.83.222:41369: write: broken pipe Feb 27 18:15:50 crc kubenswrapper[4700]: E0227 18:15:50.998604 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:15:51 crc kubenswrapper[4700]: E0227 18:15:51.983110 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:15:53 crc kubenswrapper[4700]: E0227 18:15:53.984510 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:15:55 crc kubenswrapper[4700]: I0227 18:15:55.981039 4700 scope.go:117] "RemoveContainer" containerID="454da4538d721f5f5da4ce85346adc3fbf77394a4e27f80b9bc979aa3bc37aa2" Feb 27 18:15:55 crc kubenswrapper[4700]: E0227 18:15:55.981716 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:16:00 crc kubenswrapper[4700]: I0227 18:16:00.162440 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536936-qwrfm"] Feb 27 18:16:00 crc kubenswrapper[4700]: E0227 18:16:00.163915 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb" containerName="collect-profiles" Feb 27 18:16:00 crc kubenswrapper[4700]: I0227 18:16:00.163938 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb" containerName="collect-profiles" Feb 27 18:16:00 crc kubenswrapper[4700]: I0227 18:16:00.164286 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2b1edd3-d2e0-4f31-876f-0c7e11f4cfcb" containerName="collect-profiles" Feb 27 18:16:00 crc kubenswrapper[4700]: I0227 18:16:00.165487 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536936-qwrfm" Feb 27 18:16:00 crc kubenswrapper[4700]: I0227 18:16:00.203848 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536936-qwrfm"] Feb 27 18:16:00 crc kubenswrapper[4700]: I0227 18:16:00.269164 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vnsb\" (UniqueName: \"kubernetes.io/projected/951ae05f-f758-448b-9f39-07ddedc9e642-kube-api-access-9vnsb\") pod \"auto-csr-approver-29536936-qwrfm\" (UID: \"951ae05f-f758-448b-9f39-07ddedc9e642\") " pod="openshift-infra/auto-csr-approver-29536936-qwrfm" Feb 27 18:16:00 crc kubenswrapper[4700]: I0227 18:16:00.371012 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vnsb\" (UniqueName: \"kubernetes.io/projected/951ae05f-f758-448b-9f39-07ddedc9e642-kube-api-access-9vnsb\") pod \"auto-csr-approver-29536936-qwrfm\" (UID: \"951ae05f-f758-448b-9f39-07ddedc9e642\") " pod="openshift-infra/auto-csr-approver-29536936-qwrfm" Feb 27 18:16:00 crc kubenswrapper[4700]: I0227 18:16:00.404629 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vnsb\" (UniqueName: \"kubernetes.io/projected/951ae05f-f758-448b-9f39-07ddedc9e642-kube-api-access-9vnsb\") pod \"auto-csr-approver-29536936-qwrfm\" (UID: \"951ae05f-f758-448b-9f39-07ddedc9e642\") " pod="openshift-infra/auto-csr-approver-29536936-qwrfm" Feb 27 18:16:00 crc kubenswrapper[4700]: I0227 18:16:00.499194 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536936-qwrfm" Feb 27 18:16:01 crc kubenswrapper[4700]: I0227 18:16:01.049029 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536936-qwrfm"] Feb 27 18:16:01 crc kubenswrapper[4700]: I0227 18:16:01.689105 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536936-qwrfm" event={"ID":"951ae05f-f758-448b-9f39-07ddedc9e642","Type":"ContainerStarted","Data":"3adb06eb29c13049a72692e7970c2517fee5ab8f51ae904153f40b608aa73f21"} Feb 27 18:16:03 crc kubenswrapper[4700]: I0227 18:16:03.718986 4700 generic.go:334] "Generic (PLEG): container finished" podID="951ae05f-f758-448b-9f39-07ddedc9e642" containerID="464dbbb1633fded359d0f8206112c4ec17cd067a5cf5684b646af06470e6d45a" exitCode=0 Feb 27 18:16:03 crc kubenswrapper[4700]: I0227 18:16:03.719167 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536936-qwrfm" event={"ID":"951ae05f-f758-448b-9f39-07ddedc9e642","Type":"ContainerDied","Data":"464dbbb1633fded359d0f8206112c4ec17cd067a5cf5684b646af06470e6d45a"} Feb 27 18:16:05 crc kubenswrapper[4700]: E0227 18:16:05.000189 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:16:05 crc kubenswrapper[4700]: E0227 18:16:05.001165 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:16:05 crc kubenswrapper[4700]: I0227 18:16:05.268215 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536936-qwrfm" Feb 27 18:16:05 crc kubenswrapper[4700]: I0227 18:16:05.410299 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vnsb\" (UniqueName: \"kubernetes.io/projected/951ae05f-f758-448b-9f39-07ddedc9e642-kube-api-access-9vnsb\") pod \"951ae05f-f758-448b-9f39-07ddedc9e642\" (UID: \"951ae05f-f758-448b-9f39-07ddedc9e642\") " Feb 27 18:16:05 crc kubenswrapper[4700]: I0227 18:16:05.423280 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/951ae05f-f758-448b-9f39-07ddedc9e642-kube-api-access-9vnsb" (OuterVolumeSpecName: "kube-api-access-9vnsb") pod "951ae05f-f758-448b-9f39-07ddedc9e642" (UID: "951ae05f-f758-448b-9f39-07ddedc9e642"). InnerVolumeSpecName "kube-api-access-9vnsb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:16:05 crc kubenswrapper[4700]: I0227 18:16:05.513393 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vnsb\" (UniqueName: \"kubernetes.io/projected/951ae05f-f758-448b-9f39-07ddedc9e642-kube-api-access-9vnsb\") on node \"crc\" DevicePath \"\"" Feb 27 18:16:05 crc kubenswrapper[4700]: I0227 18:16:05.747792 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536936-qwrfm" event={"ID":"951ae05f-f758-448b-9f39-07ddedc9e642","Type":"ContainerDied","Data":"3adb06eb29c13049a72692e7970c2517fee5ab8f51ae904153f40b608aa73f21"} Feb 27 18:16:05 crc kubenswrapper[4700]: I0227 18:16:05.748117 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3adb06eb29c13049a72692e7970c2517fee5ab8f51ae904153f40b608aa73f21" Feb 27 18:16:05 crc kubenswrapper[4700]: I0227 18:16:05.747926 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536936-qwrfm" Feb 27 18:16:06 crc kubenswrapper[4700]: I0227 18:16:06.384201 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536924-7tpz9"] Feb 27 18:16:06 crc kubenswrapper[4700]: I0227 18:16:06.398057 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536924-7tpz9"] Feb 27 18:16:06 crc kubenswrapper[4700]: I0227 18:16:06.996777 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="860b240c-50d6-4d7d-88d6-d1ad82c7eece" path="/var/lib/kubelet/pods/860b240c-50d6-4d7d-88d6-d1ad82c7eece/volumes" Feb 27 18:16:07 crc kubenswrapper[4700]: I0227 18:16:07.982111 4700 scope.go:117] "RemoveContainer" containerID="454da4538d721f5f5da4ce85346adc3fbf77394a4e27f80b9bc979aa3bc37aa2" Feb 27 18:16:07 crc kubenswrapper[4700]: E0227 18:16:07.982812 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:16:07 crc kubenswrapper[4700]: E0227 18:16:07.985574 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:16:12 crc kubenswrapper[4700]: I0227 18:16:12.335503 4700 scope.go:117] "RemoveContainer" containerID="96f0a92db10c87c54ea74c7765c35e5bff7413044cf9c77d45a486e2b89f4ccc" Feb 27 18:16:16 crc kubenswrapper[4700]: E0227 18:16:16.006943 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 18:16:16 crc kubenswrapper[4700]: E0227 18:16:16.007587 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 18:16:16 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 18:16:16 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gtm8j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536930-xbg7f_openshift-infra(3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 18:16:16 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 18:16:16 crc kubenswrapper[4700]: E0227 18:16:16.009544 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:16:18 crc kubenswrapper[4700]: E0227 18:16:18.986012 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:16:18 crc kubenswrapper[4700]: E0227 18:16:18.986078 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:16:21 crc kubenswrapper[4700]: I0227 18:16:21.982463 4700 scope.go:117] "RemoveContainer" containerID="454da4538d721f5f5da4ce85346adc3fbf77394a4e27f80b9bc979aa3bc37aa2" Feb 27 18:16:21 crc kubenswrapper[4700]: E0227 18:16:21.983526 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:16:21 crc kubenswrapper[4700]: E0227 18:16:21.987036 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:16:26 crc kubenswrapper[4700]: E0227 18:16:26.985709 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:16:29 crc kubenswrapper[4700]: E0227 18:16:29.985259 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:16:30 crc kubenswrapper[4700]: E0227 18:16:30.992784 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:16:33 crc kubenswrapper[4700]: I0227 18:16:33.983270 4700 scope.go:117] "RemoveContainer" containerID="454da4538d721f5f5da4ce85346adc3fbf77394a4e27f80b9bc979aa3bc37aa2" Feb 27 18:16:33 crc kubenswrapper[4700]: E0227 18:16:33.984196 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:16:34 crc kubenswrapper[4700]: E0227 18:16:34.983800 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:16:39 crc kubenswrapper[4700]: I0227 18:16:39.750640 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="f2325e29-9f4c-438f-9400-e643e2494808" containerName="galera" probeResult="failure" output="command timed out" Feb 27 18:16:39 crc kubenswrapper[4700]: I0227 18:16:39.750665 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="f2325e29-9f4c-438f-9400-e643e2494808" containerName="galera" probeResult="failure" output="command timed out" Feb 27 18:16:41 crc kubenswrapper[4700]: E0227 18:16:41.986452 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:16:42 crc kubenswrapper[4700]: E0227 18:16:42.013686 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:16:44 crc kubenswrapper[4700]: E0227 18:16:44.984504 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:16:46 crc kubenswrapper[4700]: I0227 18:16:46.982080 4700 scope.go:117] "RemoveContainer" containerID="454da4538d721f5f5da4ce85346adc3fbf77394a4e27f80b9bc979aa3bc37aa2" Feb 27 18:16:46 crc kubenswrapper[4700]: E0227 18:16:46.982911 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:16:46 crc kubenswrapper[4700]: E0227 18:16:46.985055 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:16:53 crc kubenswrapper[4700]: E0227 18:16:53.984701 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:16:55 crc kubenswrapper[4700]: E0227 18:16:55.986484 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:16:59 crc kubenswrapper[4700]: I0227 18:16:59.982778 4700 scope.go:117] "RemoveContainer" containerID="454da4538d721f5f5da4ce85346adc3fbf77394a4e27f80b9bc979aa3bc37aa2" Feb 27 18:16:59 crc kubenswrapper[4700]: E0227 18:16:59.983935 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:16:59 crc kubenswrapper[4700]: E0227 18:16:59.985580 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:17:01 crc kubenswrapper[4700]: I0227 18:17:01.004674 4700 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 18:17:03 crc kubenswrapper[4700]: E0227 18:17:03.185396 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 18:17:03 crc kubenswrapper[4700]: E0227 18:17:03.186061 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 18:17:03 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 18:17:03 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qkjfz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536934-rw77f_openshift-infra(caa4e57d-b02d-4d65-944d-451b8294f884): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 18:17:03 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 18:17:03 crc kubenswrapper[4700]: E0227 18:17:03.187732 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:17:06 crc kubenswrapper[4700]: E0227 18:17:06.984767 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:17:07 crc kubenswrapper[4700]: E0227 18:17:07.985284 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:17:12 crc kubenswrapper[4700]: I0227 18:17:12.981915 4700 scope.go:117] "RemoveContainer" containerID="454da4538d721f5f5da4ce85346adc3fbf77394a4e27f80b9bc979aa3bc37aa2" Feb 27 18:17:12 crc kubenswrapper[4700]: E0227 18:17:12.982597 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:17:13 crc kubenswrapper[4700]: E0227 18:17:13.986503 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:17:14 crc kubenswrapper[4700]: E0227 18:17:14.864381 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 18:17:14 crc kubenswrapper[4700]: E0227 18:17:14.864936 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 18:17:14 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 18:17:14 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hjvdh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536926-s6ft8_openshift-infra(d37cbaaf-1dee-4efd-9327-7bee80986a35): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 18:17:14 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 18:17:14 crc kubenswrapper[4700]: E0227 18:17:14.866279 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:17:21 crc kubenswrapper[4700]: E0227 18:17:21.005690 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:17:22 crc kubenswrapper[4700]: E0227 18:17:22.699628 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 27 18:17:22 crc kubenswrapper[4700]: E0227 18:17:22.700056 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vkwzw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-d86g8_openshift-marketplace(243afaab-17c1-4631-974c-f7b693236b08): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 18:17:22 crc kubenswrapper[4700]: E0227 18:17:22.701269 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:17:26 crc kubenswrapper[4700]: I0227 18:17:26.981963 4700 scope.go:117] "RemoveContainer" containerID="454da4538d721f5f5da4ce85346adc3fbf77394a4e27f80b9bc979aa3bc37aa2" Feb 27 18:17:26 crc kubenswrapper[4700]: E0227 18:17:26.983192 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:17:28 crc kubenswrapper[4700]: E0227 18:17:28.983822 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:17:29 crc kubenswrapper[4700]: E0227 18:17:29.984597 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:17:34 crc kubenswrapper[4700]: E0227 18:17:34.983948 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:17:34 crc kubenswrapper[4700]: E0227 18:17:34.984018 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:17:41 crc kubenswrapper[4700]: I0227 18:17:41.982611 4700 scope.go:117] "RemoveContainer" containerID="454da4538d721f5f5da4ce85346adc3fbf77394a4e27f80b9bc979aa3bc37aa2" Feb 27 18:17:41 crc kubenswrapper[4700]: E0227 18:17:41.984021 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:17:41 crc kubenswrapper[4700]: E0227 18:17:41.984642 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:17:42 crc kubenswrapper[4700]: E0227 18:17:42.983766 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:17:46 crc kubenswrapper[4700]: E0227 18:17:46.985816 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:17:49 crc kubenswrapper[4700]: E0227 18:17:49.984736 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:17:54 crc kubenswrapper[4700]: E0227 18:17:54.990385 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:17:55 crc kubenswrapper[4700]: I0227 18:17:55.981846 4700 scope.go:117] "RemoveContainer" containerID="454da4538d721f5f5da4ce85346adc3fbf77394a4e27f80b9bc979aa3bc37aa2" Feb 27 18:17:55 crc kubenswrapper[4700]: E0227 18:17:55.982848 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:17:56 crc kubenswrapper[4700]: E0227 18:17:56.984738 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:17:59 crc kubenswrapper[4700]: E0227 18:17:59.986042 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:18:00 crc kubenswrapper[4700]: I0227 18:18:00.155577 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536938-bppsz"] Feb 27 18:18:00 crc kubenswrapper[4700]: E0227 18:18:00.156067 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="951ae05f-f758-448b-9f39-07ddedc9e642" containerName="oc" Feb 27 18:18:00 crc kubenswrapper[4700]: I0227 18:18:00.156089 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="951ae05f-f758-448b-9f39-07ddedc9e642" containerName="oc" Feb 27 18:18:00 crc kubenswrapper[4700]: I0227 18:18:00.156335 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="951ae05f-f758-448b-9f39-07ddedc9e642" containerName="oc" Feb 27 18:18:00 crc kubenswrapper[4700]: I0227 18:18:00.157121 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536938-bppsz" Feb 27 18:18:00 crc kubenswrapper[4700]: I0227 18:18:00.189772 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536938-bppsz"] Feb 27 18:18:00 crc kubenswrapper[4700]: I0227 18:18:00.221615 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4jqf\" (UniqueName: \"kubernetes.io/projected/90f1566e-7eeb-4455-9bcd-a9eed39c5eb1-kube-api-access-n4jqf\") pod \"auto-csr-approver-29536938-bppsz\" (UID: \"90f1566e-7eeb-4455-9bcd-a9eed39c5eb1\") " pod="openshift-infra/auto-csr-approver-29536938-bppsz" Feb 27 18:18:00 crc kubenswrapper[4700]: I0227 18:18:00.324972 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4jqf\" (UniqueName: \"kubernetes.io/projected/90f1566e-7eeb-4455-9bcd-a9eed39c5eb1-kube-api-access-n4jqf\") pod \"auto-csr-approver-29536938-bppsz\" (UID: \"90f1566e-7eeb-4455-9bcd-a9eed39c5eb1\") " pod="openshift-infra/auto-csr-approver-29536938-bppsz" Feb 27 18:18:00 crc kubenswrapper[4700]: I0227 18:18:00.349081 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4jqf\" (UniqueName: \"kubernetes.io/projected/90f1566e-7eeb-4455-9bcd-a9eed39c5eb1-kube-api-access-n4jqf\") pod \"auto-csr-approver-29536938-bppsz\" (UID: \"90f1566e-7eeb-4455-9bcd-a9eed39c5eb1\") " pod="openshift-infra/auto-csr-approver-29536938-bppsz" Feb 27 18:18:00 crc kubenswrapper[4700]: I0227 18:18:00.496848 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536938-bppsz" Feb 27 18:18:01 crc kubenswrapper[4700]: I0227 18:18:01.010180 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536938-bppsz"] Feb 27 18:18:01 crc kubenswrapper[4700]: I0227 18:18:01.169595 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536938-bppsz" event={"ID":"90f1566e-7eeb-4455-9bcd-a9eed39c5eb1","Type":"ContainerStarted","Data":"aa6c28ea7fa5f5c6f3cd3806890a91b1521dc45844bf8c8510fef51f81200098"} Feb 27 18:18:01 crc kubenswrapper[4700]: E0227 18:18:01.984274 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:18:02 crc kubenswrapper[4700]: E0227 18:18:02.345656 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 18:18:02 crc kubenswrapper[4700]: E0227 18:18:02.347959 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 18:18:02 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 18:18:02 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n4jqf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536938-bppsz_openshift-infra(90f1566e-7eeb-4455-9bcd-a9eed39c5eb1): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 18:18:02 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 18:18:02 crc kubenswrapper[4700]: E0227 18:18:02.350081 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536938-bppsz" podUID="90f1566e-7eeb-4455-9bcd-a9eed39c5eb1" Feb 27 18:18:03 crc kubenswrapper[4700]: E0227 18:18:03.194695 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536938-bppsz" podUID="90f1566e-7eeb-4455-9bcd-a9eed39c5eb1" Feb 27 18:18:07 crc kubenswrapper[4700]: I0227 18:18:07.981601 4700 scope.go:117] "RemoveContainer" containerID="454da4538d721f5f5da4ce85346adc3fbf77394a4e27f80b9bc979aa3bc37aa2" Feb 27 18:18:07 crc kubenswrapper[4700]: E0227 18:18:07.982445 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:18:08 crc kubenswrapper[4700]: E0227 18:18:08.984530 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:18:08 crc kubenswrapper[4700]: E0227 18:18:08.984649 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:18:11 crc kubenswrapper[4700]: E0227 18:18:11.985141 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:18:14 crc kubenswrapper[4700]: E0227 18:18:14.984895 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:18:19 crc kubenswrapper[4700]: E0227 18:18:19.059441 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 18:18:19 crc kubenswrapper[4700]: E0227 18:18:19.060617 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 18:18:19 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 18:18:19 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n4jqf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536938-bppsz_openshift-infra(90f1566e-7eeb-4455-9bcd-a9eed39c5eb1): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 18:18:19 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 18:18:19 crc kubenswrapper[4700]: E0227 18:18:19.062278 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536938-bppsz" podUID="90f1566e-7eeb-4455-9bcd-a9eed39c5eb1" Feb 27 18:18:21 crc kubenswrapper[4700]: E0227 18:18:20.998609 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:18:21 crc kubenswrapper[4700]: I0227 18:18:21.982190 4700 scope.go:117] "RemoveContainer" containerID="454da4538d721f5f5da4ce85346adc3fbf77394a4e27f80b9bc979aa3bc37aa2" Feb 27 18:18:21 crc kubenswrapper[4700]: E0227 18:18:21.982886 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:18:22 crc kubenswrapper[4700]: E0227 18:18:22.984982 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:18:23 crc kubenswrapper[4700]: E0227 18:18:23.983952 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:18:27 crc kubenswrapper[4700]: E0227 18:18:27.984541 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:18:31 crc kubenswrapper[4700]: E0227 18:18:30.999917 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536938-bppsz" podUID="90f1566e-7eeb-4455-9bcd-a9eed39c5eb1" Feb 27 18:18:32 crc kubenswrapper[4700]: I0227 18:18:32.982169 4700 scope.go:117] "RemoveContainer" containerID="454da4538d721f5f5da4ce85346adc3fbf77394a4e27f80b9bc979aa3bc37aa2" Feb 27 18:18:32 crc kubenswrapper[4700]: E0227 18:18:32.983163 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:18:34 crc kubenswrapper[4700]: E0227 18:18:34.983129 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:18:35 crc kubenswrapper[4700]: E0227 18:18:35.984652 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:18:36 crc kubenswrapper[4700]: E0227 18:18:36.985490 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:18:39 crc kubenswrapper[4700]: E0227 18:18:39.983762 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:18:43 crc kubenswrapper[4700]: E0227 18:18:43.245989 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 18:18:43 crc kubenswrapper[4700]: E0227 18:18:43.247548 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 18:18:43 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 18:18:43 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n4jqf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536938-bppsz_openshift-infra(90f1566e-7eeb-4455-9bcd-a9eed39c5eb1): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 18:18:43 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 18:18:43 crc kubenswrapper[4700]: E0227 18:18:43.248945 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536938-bppsz" podUID="90f1566e-7eeb-4455-9bcd-a9eed39c5eb1" Feb 27 18:18:45 crc kubenswrapper[4700]: E0227 18:18:45.986574 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:18:46 crc kubenswrapper[4700]: I0227 18:18:46.982034 4700 scope.go:117] "RemoveContainer" containerID="454da4538d721f5f5da4ce85346adc3fbf77394a4e27f80b9bc979aa3bc37aa2" Feb 27 18:18:46 crc kubenswrapper[4700]: E0227 18:18:46.982734 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:18:48 crc kubenswrapper[4700]: I0227 18:18:48.070068 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lzfcp"] Feb 27 18:18:48 crc kubenswrapper[4700]: I0227 18:18:48.072499 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lzfcp" Feb 27 18:18:48 crc kubenswrapper[4700]: I0227 18:18:48.085199 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lzfcp"] Feb 27 18:18:48 crc kubenswrapper[4700]: I0227 18:18:48.185115 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc031a5d-e87f-4625-9b92-a7d0e539506e-catalog-content\") pod \"community-operators-lzfcp\" (UID: \"bc031a5d-e87f-4625-9b92-a7d0e539506e\") " pod="openshift-marketplace/community-operators-lzfcp" Feb 27 18:18:48 crc kubenswrapper[4700]: I0227 18:18:48.185591 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc031a5d-e87f-4625-9b92-a7d0e539506e-utilities\") pod \"community-operators-lzfcp\" (UID: \"bc031a5d-e87f-4625-9b92-a7d0e539506e\") " pod="openshift-marketplace/community-operators-lzfcp" Feb 27 18:18:48 crc kubenswrapper[4700]: I0227 18:18:48.185836 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pz5gr\" (UniqueName: \"kubernetes.io/projected/bc031a5d-e87f-4625-9b92-a7d0e539506e-kube-api-access-pz5gr\") pod \"community-operators-lzfcp\" (UID: \"bc031a5d-e87f-4625-9b92-a7d0e539506e\") " pod="openshift-marketplace/community-operators-lzfcp" Feb 27 18:18:48 crc kubenswrapper[4700]: I0227 18:18:48.287993 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc031a5d-e87f-4625-9b92-a7d0e539506e-utilities\") pod \"community-operators-lzfcp\" (UID: \"bc031a5d-e87f-4625-9b92-a7d0e539506e\") " pod="openshift-marketplace/community-operators-lzfcp" Feb 27 18:18:48 crc kubenswrapper[4700]: I0227 18:18:48.288138 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pz5gr\" (UniqueName: \"kubernetes.io/projected/bc031a5d-e87f-4625-9b92-a7d0e539506e-kube-api-access-pz5gr\") pod \"community-operators-lzfcp\" (UID: \"bc031a5d-e87f-4625-9b92-a7d0e539506e\") " pod="openshift-marketplace/community-operators-lzfcp" Feb 27 18:18:48 crc kubenswrapper[4700]: I0227 18:18:48.288280 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc031a5d-e87f-4625-9b92-a7d0e539506e-catalog-content\") pod \"community-operators-lzfcp\" (UID: \"bc031a5d-e87f-4625-9b92-a7d0e539506e\") " pod="openshift-marketplace/community-operators-lzfcp" Feb 27 18:18:48 crc kubenswrapper[4700]: I0227 18:18:48.288542 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc031a5d-e87f-4625-9b92-a7d0e539506e-utilities\") pod \"community-operators-lzfcp\" (UID: \"bc031a5d-e87f-4625-9b92-a7d0e539506e\") " pod="openshift-marketplace/community-operators-lzfcp" Feb 27 18:18:48 crc kubenswrapper[4700]: I0227 18:18:48.288944 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc031a5d-e87f-4625-9b92-a7d0e539506e-catalog-content\") pod \"community-operators-lzfcp\" (UID: \"bc031a5d-e87f-4625-9b92-a7d0e539506e\") " pod="openshift-marketplace/community-operators-lzfcp" Feb 27 18:18:48 crc kubenswrapper[4700]: I0227 18:18:48.324886 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pz5gr\" (UniqueName: \"kubernetes.io/projected/bc031a5d-e87f-4625-9b92-a7d0e539506e-kube-api-access-pz5gr\") pod \"community-operators-lzfcp\" (UID: \"bc031a5d-e87f-4625-9b92-a7d0e539506e\") " pod="openshift-marketplace/community-operators-lzfcp" Feb 27 18:18:48 crc kubenswrapper[4700]: I0227 18:18:48.407355 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lzfcp" Feb 27 18:18:48 crc kubenswrapper[4700]: E0227 18:18:48.983943 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:18:49 crc kubenswrapper[4700]: I0227 18:18:49.794608 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lzfcp"] Feb 27 18:18:49 crc kubenswrapper[4700]: W0227 18:18:49.806078 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbc031a5d_e87f_4625_9b92_a7d0e539506e.slice/crio-bdbed22a09c7b6460661f6e86bf335d05d92e494cd5c110b19581d34a98cc190 WatchSource:0}: Error finding container bdbed22a09c7b6460661f6e86bf335d05d92e494cd5c110b19581d34a98cc190: Status 404 returned error can't find the container with id bdbed22a09c7b6460661f6e86bf335d05d92e494cd5c110b19581d34a98cc190 Feb 27 18:18:50 crc kubenswrapper[4700]: I0227 18:18:50.759965 4700 generic.go:334] "Generic (PLEG): container finished" podID="bc031a5d-e87f-4625-9b92-a7d0e539506e" containerID="85b204a3c1fd70456f22f3440f8d1014e104c1a8cbc3f47f3b7dc3703c125d13" exitCode=0 Feb 27 18:18:50 crc kubenswrapper[4700]: I0227 18:18:50.760380 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lzfcp" event={"ID":"bc031a5d-e87f-4625-9b92-a7d0e539506e","Type":"ContainerDied","Data":"85b204a3c1fd70456f22f3440f8d1014e104c1a8cbc3f47f3b7dc3703c125d13"} Feb 27 18:18:50 crc kubenswrapper[4700]: I0227 18:18:50.760414 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lzfcp" event={"ID":"bc031a5d-e87f-4625-9b92-a7d0e539506e","Type":"ContainerStarted","Data":"bdbed22a09c7b6460661f6e86bf335d05d92e494cd5c110b19581d34a98cc190"} Feb 27 18:18:50 crc kubenswrapper[4700]: E0227 18:18:50.997602 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:18:51 crc kubenswrapper[4700]: E0227 18:18:51.634023 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 27 18:18:51 crc kubenswrapper[4700]: E0227 18:18:51.634189 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pz5gr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-lzfcp_openshift-marketplace(bc031a5d-e87f-4625-9b92-a7d0e539506e): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 18:18:51 crc kubenswrapper[4700]: E0227 18:18:51.635351 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/community-operators-lzfcp" podUID="bc031a5d-e87f-4625-9b92-a7d0e539506e" Feb 27 18:18:51 crc kubenswrapper[4700]: E0227 18:18:51.787519 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-lzfcp" podUID="bc031a5d-e87f-4625-9b92-a7d0e539506e" Feb 27 18:18:51 crc kubenswrapper[4700]: E0227 18:18:51.984633 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:18:53 crc kubenswrapper[4700]: I0227 18:18:53.092894 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-n8hf9"] Feb 27 18:18:53 crc kubenswrapper[4700]: I0227 18:18:53.099965 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n8hf9" Feb 27 18:18:53 crc kubenswrapper[4700]: I0227 18:18:53.119867 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n8hf9"] Feb 27 18:18:53 crc kubenswrapper[4700]: I0227 18:18:53.139262 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88d71cab-ee75-4276-8c11-dd28e7ff49ef-catalog-content\") pod \"redhat-operators-n8hf9\" (UID: \"88d71cab-ee75-4276-8c11-dd28e7ff49ef\") " pod="openshift-marketplace/redhat-operators-n8hf9" Feb 27 18:18:53 crc kubenswrapper[4700]: I0227 18:18:53.141559 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsldw\" (UniqueName: \"kubernetes.io/projected/88d71cab-ee75-4276-8c11-dd28e7ff49ef-kube-api-access-zsldw\") pod \"redhat-operators-n8hf9\" (UID: \"88d71cab-ee75-4276-8c11-dd28e7ff49ef\") " pod="openshift-marketplace/redhat-operators-n8hf9" Feb 27 18:18:53 crc kubenswrapper[4700]: I0227 18:18:53.141756 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88d71cab-ee75-4276-8c11-dd28e7ff49ef-utilities\") pod \"redhat-operators-n8hf9\" (UID: \"88d71cab-ee75-4276-8c11-dd28e7ff49ef\") " pod="openshift-marketplace/redhat-operators-n8hf9" Feb 27 18:18:53 crc kubenswrapper[4700]: I0227 18:18:53.244328 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88d71cab-ee75-4276-8c11-dd28e7ff49ef-catalog-content\") pod \"redhat-operators-n8hf9\" (UID: \"88d71cab-ee75-4276-8c11-dd28e7ff49ef\") " pod="openshift-marketplace/redhat-operators-n8hf9" Feb 27 18:18:53 crc kubenswrapper[4700]: I0227 18:18:53.244453 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsldw\" (UniqueName: \"kubernetes.io/projected/88d71cab-ee75-4276-8c11-dd28e7ff49ef-kube-api-access-zsldw\") pod \"redhat-operators-n8hf9\" (UID: \"88d71cab-ee75-4276-8c11-dd28e7ff49ef\") " pod="openshift-marketplace/redhat-operators-n8hf9" Feb 27 18:18:53 crc kubenswrapper[4700]: I0227 18:18:53.244513 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88d71cab-ee75-4276-8c11-dd28e7ff49ef-utilities\") pod \"redhat-operators-n8hf9\" (UID: \"88d71cab-ee75-4276-8c11-dd28e7ff49ef\") " pod="openshift-marketplace/redhat-operators-n8hf9" Feb 27 18:18:53 crc kubenswrapper[4700]: I0227 18:18:53.245172 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88d71cab-ee75-4276-8c11-dd28e7ff49ef-catalog-content\") pod \"redhat-operators-n8hf9\" (UID: \"88d71cab-ee75-4276-8c11-dd28e7ff49ef\") " pod="openshift-marketplace/redhat-operators-n8hf9" Feb 27 18:18:53 crc kubenswrapper[4700]: I0227 18:18:53.245213 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88d71cab-ee75-4276-8c11-dd28e7ff49ef-utilities\") pod \"redhat-operators-n8hf9\" (UID: \"88d71cab-ee75-4276-8c11-dd28e7ff49ef\") " pod="openshift-marketplace/redhat-operators-n8hf9" Feb 27 18:18:53 crc kubenswrapper[4700]: I0227 18:18:53.273647 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsldw\" (UniqueName: \"kubernetes.io/projected/88d71cab-ee75-4276-8c11-dd28e7ff49ef-kube-api-access-zsldw\") pod \"redhat-operators-n8hf9\" (UID: \"88d71cab-ee75-4276-8c11-dd28e7ff49ef\") " pod="openshift-marketplace/redhat-operators-n8hf9" Feb 27 18:18:53 crc kubenswrapper[4700]: I0227 18:18:53.451358 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n8hf9" Feb 27 18:18:53 crc kubenswrapper[4700]: I0227 18:18:53.975125 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n8hf9"] Feb 27 18:18:53 crc kubenswrapper[4700]: W0227 18:18:53.978429 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod88d71cab_ee75_4276_8c11_dd28e7ff49ef.slice/crio-4728795ed9c9cf014acc4f7839010ed9387bf0efb901b385688221739e94fb21 WatchSource:0}: Error finding container 4728795ed9c9cf014acc4f7839010ed9387bf0efb901b385688221739e94fb21: Status 404 returned error can't find the container with id 4728795ed9c9cf014acc4f7839010ed9387bf0efb901b385688221739e94fb21 Feb 27 18:18:53 crc kubenswrapper[4700]: E0227 18:18:53.982064 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536938-bppsz" podUID="90f1566e-7eeb-4455-9bcd-a9eed39c5eb1" Feb 27 18:18:54 crc kubenswrapper[4700]: I0227 18:18:54.813142 4700 generic.go:334] "Generic (PLEG): container finished" podID="88d71cab-ee75-4276-8c11-dd28e7ff49ef" containerID="3ceb6b1b013043d1716da8a19d273b0488e5d13ce5034afe612c4f761cd5f083" exitCode=0 Feb 27 18:18:54 crc kubenswrapper[4700]: I0227 18:18:54.813399 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n8hf9" event={"ID":"88d71cab-ee75-4276-8c11-dd28e7ff49ef","Type":"ContainerDied","Data":"3ceb6b1b013043d1716da8a19d273b0488e5d13ce5034afe612c4f761cd5f083"} Feb 27 18:18:54 crc kubenswrapper[4700]: I0227 18:18:54.813514 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n8hf9" event={"ID":"88d71cab-ee75-4276-8c11-dd28e7ff49ef","Type":"ContainerStarted","Data":"4728795ed9c9cf014acc4f7839010ed9387bf0efb901b385688221739e94fb21"} Feb 27 18:18:55 crc kubenswrapper[4700]: E0227 18:18:55.547152 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 27 18:18:55 crc kubenswrapper[4700]: E0227 18:18:55.547859 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zsldw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-n8hf9_openshift-marketplace(88d71cab-ee75-4276-8c11-dd28e7ff49ef): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 18:18:55 crc kubenswrapper[4700]: E0227 18:18:55.549182 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=340dbaa786c584e5ffe05a0f79571b9c2fe7d16a1a1fb390e5d83b437d7a1ff3/signature-3: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-operators-n8hf9" podUID="88d71cab-ee75-4276-8c11-dd28e7ff49ef" Feb 27 18:18:55 crc kubenswrapper[4700]: E0227 18:18:55.853931 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-n8hf9" podUID="88d71cab-ee75-4276-8c11-dd28e7ff49ef" Feb 27 18:18:57 crc kubenswrapper[4700]: I0227 18:18:57.982034 4700 scope.go:117] "RemoveContainer" containerID="454da4538d721f5f5da4ce85346adc3fbf77394a4e27f80b9bc979aa3bc37aa2" Feb 27 18:18:57 crc kubenswrapper[4700]: E0227 18:18:57.982821 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:18:57 crc kubenswrapper[4700]: E0227 18:18:57.986803 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:19:03 crc kubenswrapper[4700]: E0227 18:19:03.984994 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:19:04 crc kubenswrapper[4700]: E0227 18:19:04.621629 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 27 18:19:04 crc kubenswrapper[4700]: E0227 18:19:04.621829 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pz5gr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-lzfcp_openshift-marketplace(bc031a5d-e87f-4625-9b92-a7d0e539506e): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 18:19:04 crc kubenswrapper[4700]: E0227 18:19:04.623117 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/community-operators-lzfcp" podUID="bc031a5d-e87f-4625-9b92-a7d0e539506e" Feb 27 18:19:06 crc kubenswrapper[4700]: E0227 18:19:06.987270 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:19:06 crc kubenswrapper[4700]: E0227 18:19:06.987511 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536938-bppsz" podUID="90f1566e-7eeb-4455-9bcd-a9eed39c5eb1" Feb 27 18:19:06 crc kubenswrapper[4700]: E0227 18:19:06.987593 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:19:09 crc kubenswrapper[4700]: I0227 18:19:09.023553 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n8hf9" event={"ID":"88d71cab-ee75-4276-8c11-dd28e7ff49ef","Type":"ContainerStarted","Data":"512d4dbd5bfdadd93ebec56d852bedbd780a6928f31978424ff9956eab4a74ae"} Feb 27 18:19:11 crc kubenswrapper[4700]: I0227 18:19:10.999027 4700 scope.go:117] "RemoveContainer" containerID="454da4538d721f5f5da4ce85346adc3fbf77394a4e27f80b9bc979aa3bc37aa2" Feb 27 18:19:12 crc kubenswrapper[4700]: I0227 18:19:12.069054 4700 generic.go:334] "Generic (PLEG): container finished" podID="88d71cab-ee75-4276-8c11-dd28e7ff49ef" containerID="512d4dbd5bfdadd93ebec56d852bedbd780a6928f31978424ff9956eab4a74ae" exitCode=0 Feb 27 18:19:12 crc kubenswrapper[4700]: I0227 18:19:12.069126 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n8hf9" event={"ID":"88d71cab-ee75-4276-8c11-dd28e7ff49ef","Type":"ContainerDied","Data":"512d4dbd5bfdadd93ebec56d852bedbd780a6928f31978424ff9956eab4a74ae"} Feb 27 18:19:12 crc kubenswrapper[4700]: I0227 18:19:12.088876 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerStarted","Data":"d616594fff3f3a84ae3299a2a56f70ba23d3b8c11784f6e69c706e5bb75192a5"} Feb 27 18:19:12 crc kubenswrapper[4700]: E0227 18:19:12.984823 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:19:13 crc kubenswrapper[4700]: I0227 18:19:13.109264 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n8hf9" event={"ID":"88d71cab-ee75-4276-8c11-dd28e7ff49ef","Type":"ContainerStarted","Data":"7df0d8cf9eae44f0125d89bfc3f64081595d308b4a34b2b509e7f8fe2ac60f9d"} Feb 27 18:19:13 crc kubenswrapper[4700]: I0227 18:19:13.152267 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-n8hf9" podStartSLOduration=2.440792073 podStartE2EDuration="20.152244654s" podCreationTimestamp="2026-02-27 18:18:53 +0000 UTC" firstStartedPulling="2026-02-27 18:18:54.815529006 +0000 UTC m=+4694.800841763" lastFinishedPulling="2026-02-27 18:19:12.526981597 +0000 UTC m=+4712.512294344" observedRunningTime="2026-02-27 18:19:13.133865715 +0000 UTC m=+4713.119178472" watchObservedRunningTime="2026-02-27 18:19:13.152244654 +0000 UTC m=+4713.137557411" Feb 27 18:19:13 crc kubenswrapper[4700]: I0227 18:19:13.452291 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-n8hf9" Feb 27 18:19:13 crc kubenswrapper[4700]: I0227 18:19:13.452361 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-n8hf9" Feb 27 18:19:14 crc kubenswrapper[4700]: I0227 18:19:14.529164 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-n8hf9" podUID="88d71cab-ee75-4276-8c11-dd28e7ff49ef" containerName="registry-server" probeResult="failure" output=< Feb 27 18:19:14 crc kubenswrapper[4700]: timeout: failed to connect service ":50051" within 1s Feb 27 18:19:14 crc kubenswrapper[4700]: > Feb 27 18:19:15 crc kubenswrapper[4700]: E0227 18:19:15.986747 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-lzfcp" podUID="bc031a5d-e87f-4625-9b92-a7d0e539506e" Feb 27 18:19:16 crc kubenswrapper[4700]: E0227 18:19:16.983863 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:19:18 crc kubenswrapper[4700]: E0227 18:19:18.988836 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:19:21 crc kubenswrapper[4700]: E0227 18:19:21.003170 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536938-bppsz" podUID="90f1566e-7eeb-4455-9bcd-a9eed39c5eb1" Feb 27 18:19:21 crc kubenswrapper[4700]: E0227 18:19:21.984007 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:19:24 crc kubenswrapper[4700]: I0227 18:19:24.528170 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-n8hf9" podUID="88d71cab-ee75-4276-8c11-dd28e7ff49ef" containerName="registry-server" probeResult="failure" output=< Feb 27 18:19:24 crc kubenswrapper[4700]: timeout: failed to connect service ":50051" within 1s Feb 27 18:19:24 crc kubenswrapper[4700]: > Feb 27 18:19:26 crc kubenswrapper[4700]: E0227 18:19:26.982951 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:19:27 crc kubenswrapper[4700]: E0227 18:19:27.984235 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:19:32 crc kubenswrapper[4700]: I0227 18:19:32.382663 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lzfcp" event={"ID":"bc031a5d-e87f-4625-9b92-a7d0e539506e","Type":"ContainerStarted","Data":"5e02003e4306bbed4552ae2330440c9bd350e7b19eb9fb02626d2ff1464a649b"} Feb 27 18:19:33 crc kubenswrapper[4700]: I0227 18:19:33.400002 4700 generic.go:334] "Generic (PLEG): container finished" podID="bc031a5d-e87f-4625-9b92-a7d0e539506e" containerID="5e02003e4306bbed4552ae2330440c9bd350e7b19eb9fb02626d2ff1464a649b" exitCode=0 Feb 27 18:19:33 crc kubenswrapper[4700]: I0227 18:19:33.400093 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lzfcp" event={"ID":"bc031a5d-e87f-4625-9b92-a7d0e539506e","Type":"ContainerDied","Data":"5e02003e4306bbed4552ae2330440c9bd350e7b19eb9fb02626d2ff1464a649b"} Feb 27 18:19:33 crc kubenswrapper[4700]: I0227 18:19:33.553856 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-n8hf9" Feb 27 18:19:33 crc kubenswrapper[4700]: I0227 18:19:33.643937 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-n8hf9" Feb 27 18:19:33 crc kubenswrapper[4700]: E0227 18:19:33.984497 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:19:34 crc kubenswrapper[4700]: I0227 18:19:34.418254 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lzfcp" event={"ID":"bc031a5d-e87f-4625-9b92-a7d0e539506e","Type":"ContainerStarted","Data":"539c9833b4fd20b65e274a0ea65f56274d65da1aeccb49929bff2cab13221aa9"} Feb 27 18:19:34 crc kubenswrapper[4700]: I0227 18:19:34.457557 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lzfcp" podStartSLOduration=3.380543583 podStartE2EDuration="46.457524825s" podCreationTimestamp="2026-02-27 18:18:48 +0000 UTC" firstStartedPulling="2026-02-27 18:18:50.763597963 +0000 UTC m=+4690.748910720" lastFinishedPulling="2026-02-27 18:19:33.840579205 +0000 UTC m=+4733.825891962" observedRunningTime="2026-02-27 18:19:34.442972116 +0000 UTC m=+4734.428284893" watchObservedRunningTime="2026-02-27 18:19:34.457524825 +0000 UTC m=+4734.442837602" Feb 27 18:19:35 crc kubenswrapper[4700]: E0227 18:19:35.003966 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:19:35 crc kubenswrapper[4700]: I0227 18:19:35.441367 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536938-bppsz" event={"ID":"90f1566e-7eeb-4455-9bcd-a9eed39c5eb1","Type":"ContainerStarted","Data":"74145e784d29840559779fda9c00434464fd79d87683569916a30512b3a7987b"} Feb 27 18:19:35 crc kubenswrapper[4700]: I0227 18:19:35.461776 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536938-bppsz" podStartSLOduration=1.522753367 podStartE2EDuration="1m35.461753093s" podCreationTimestamp="2026-02-27 18:18:00 +0000 UTC" firstStartedPulling="2026-02-27 18:18:01.019968121 +0000 UTC m=+4641.005280868" lastFinishedPulling="2026-02-27 18:19:34.958967817 +0000 UTC m=+4734.944280594" observedRunningTime="2026-02-27 18:19:35.458630932 +0000 UTC m=+4735.443943689" watchObservedRunningTime="2026-02-27 18:19:35.461753093 +0000 UTC m=+4735.447065850" Feb 27 18:19:35 crc kubenswrapper[4700]: I0227 18:19:35.858533 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n8hf9"] Feb 27 18:19:35 crc kubenswrapper[4700]: I0227 18:19:35.859262 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-n8hf9" podUID="88d71cab-ee75-4276-8c11-dd28e7ff49ef" containerName="registry-server" containerID="cri-o://7df0d8cf9eae44f0125d89bfc3f64081595d308b4a34b2b509e7f8fe2ac60f9d" gracePeriod=2 Feb 27 18:19:36 crc kubenswrapper[4700]: I0227 18:19:36.463066 4700 generic.go:334] "Generic (PLEG): container finished" podID="88d71cab-ee75-4276-8c11-dd28e7ff49ef" containerID="7df0d8cf9eae44f0125d89bfc3f64081595d308b4a34b2b509e7f8fe2ac60f9d" exitCode=0 Feb 27 18:19:36 crc kubenswrapper[4700]: I0227 18:19:36.463175 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n8hf9" event={"ID":"88d71cab-ee75-4276-8c11-dd28e7ff49ef","Type":"ContainerDied","Data":"7df0d8cf9eae44f0125d89bfc3f64081595d308b4a34b2b509e7f8fe2ac60f9d"} Feb 27 18:19:36 crc kubenswrapper[4700]: I0227 18:19:36.466530 4700 generic.go:334] "Generic (PLEG): container finished" podID="90f1566e-7eeb-4455-9bcd-a9eed39c5eb1" containerID="74145e784d29840559779fda9c00434464fd79d87683569916a30512b3a7987b" exitCode=0 Feb 27 18:19:36 crc kubenswrapper[4700]: I0227 18:19:36.466599 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536938-bppsz" event={"ID":"90f1566e-7eeb-4455-9bcd-a9eed39c5eb1","Type":"ContainerDied","Data":"74145e784d29840559779fda9c00434464fd79d87683569916a30512b3a7987b"} Feb 27 18:19:36 crc kubenswrapper[4700]: I0227 18:19:36.795867 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n8hf9" Feb 27 18:19:36 crc kubenswrapper[4700]: I0227 18:19:36.907815 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88d71cab-ee75-4276-8c11-dd28e7ff49ef-utilities\") pod \"88d71cab-ee75-4276-8c11-dd28e7ff49ef\" (UID: \"88d71cab-ee75-4276-8c11-dd28e7ff49ef\") " Feb 27 18:19:36 crc kubenswrapper[4700]: I0227 18:19:36.907987 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88d71cab-ee75-4276-8c11-dd28e7ff49ef-catalog-content\") pod \"88d71cab-ee75-4276-8c11-dd28e7ff49ef\" (UID: \"88d71cab-ee75-4276-8c11-dd28e7ff49ef\") " Feb 27 18:19:36 crc kubenswrapper[4700]: I0227 18:19:36.908066 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zsldw\" (UniqueName: \"kubernetes.io/projected/88d71cab-ee75-4276-8c11-dd28e7ff49ef-kube-api-access-zsldw\") pod \"88d71cab-ee75-4276-8c11-dd28e7ff49ef\" (UID: \"88d71cab-ee75-4276-8c11-dd28e7ff49ef\") " Feb 27 18:19:36 crc kubenswrapper[4700]: I0227 18:19:36.911779 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88d71cab-ee75-4276-8c11-dd28e7ff49ef-utilities" (OuterVolumeSpecName: "utilities") pod "88d71cab-ee75-4276-8c11-dd28e7ff49ef" (UID: "88d71cab-ee75-4276-8c11-dd28e7ff49ef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 18:19:36 crc kubenswrapper[4700]: I0227 18:19:36.920842 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88d71cab-ee75-4276-8c11-dd28e7ff49ef-kube-api-access-zsldw" (OuterVolumeSpecName: "kube-api-access-zsldw") pod "88d71cab-ee75-4276-8c11-dd28e7ff49ef" (UID: "88d71cab-ee75-4276-8c11-dd28e7ff49ef"). InnerVolumeSpecName "kube-api-access-zsldw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:19:37 crc kubenswrapper[4700]: I0227 18:19:37.011816 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88d71cab-ee75-4276-8c11-dd28e7ff49ef-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 18:19:37 crc kubenswrapper[4700]: I0227 18:19:37.011864 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zsldw\" (UniqueName: \"kubernetes.io/projected/88d71cab-ee75-4276-8c11-dd28e7ff49ef-kube-api-access-zsldw\") on node \"crc\" DevicePath \"\"" Feb 27 18:19:37 crc kubenswrapper[4700]: I0227 18:19:37.087994 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88d71cab-ee75-4276-8c11-dd28e7ff49ef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "88d71cab-ee75-4276-8c11-dd28e7ff49ef" (UID: "88d71cab-ee75-4276-8c11-dd28e7ff49ef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 18:19:37 crc kubenswrapper[4700]: I0227 18:19:37.114838 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88d71cab-ee75-4276-8c11-dd28e7ff49ef-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 18:19:37 crc kubenswrapper[4700]: I0227 18:19:37.492671 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n8hf9" event={"ID":"88d71cab-ee75-4276-8c11-dd28e7ff49ef","Type":"ContainerDied","Data":"4728795ed9c9cf014acc4f7839010ed9387bf0efb901b385688221739e94fb21"} Feb 27 18:19:37 crc kubenswrapper[4700]: I0227 18:19:37.492771 4700 scope.go:117] "RemoveContainer" containerID="7df0d8cf9eae44f0125d89bfc3f64081595d308b4a34b2b509e7f8fe2ac60f9d" Feb 27 18:19:37 crc kubenswrapper[4700]: I0227 18:19:37.492816 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n8hf9" Feb 27 18:19:37 crc kubenswrapper[4700]: I0227 18:19:37.523886 4700 scope.go:117] "RemoveContainer" containerID="512d4dbd5bfdadd93ebec56d852bedbd780a6928f31978424ff9956eab4a74ae" Feb 27 18:19:37 crc kubenswrapper[4700]: I0227 18:19:37.564775 4700 scope.go:117] "RemoveContainer" containerID="3ceb6b1b013043d1716da8a19d273b0488e5d13ce5034afe612c4f761cd5f083" Feb 27 18:19:37 crc kubenswrapper[4700]: I0227 18:19:37.574245 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n8hf9"] Feb 27 18:19:37 crc kubenswrapper[4700]: I0227 18:19:37.595231 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-n8hf9"] Feb 27 18:19:37 crc kubenswrapper[4700]: I0227 18:19:37.929000 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536938-bppsz" Feb 27 18:19:38 crc kubenswrapper[4700]: I0227 18:19:38.046857 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4jqf\" (UniqueName: \"kubernetes.io/projected/90f1566e-7eeb-4455-9bcd-a9eed39c5eb1-kube-api-access-n4jqf\") pod \"90f1566e-7eeb-4455-9bcd-a9eed39c5eb1\" (UID: \"90f1566e-7eeb-4455-9bcd-a9eed39c5eb1\") " Feb 27 18:19:38 crc kubenswrapper[4700]: I0227 18:19:38.059211 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90f1566e-7eeb-4455-9bcd-a9eed39c5eb1-kube-api-access-n4jqf" (OuterVolumeSpecName: "kube-api-access-n4jqf") pod "90f1566e-7eeb-4455-9bcd-a9eed39c5eb1" (UID: "90f1566e-7eeb-4455-9bcd-a9eed39c5eb1"). InnerVolumeSpecName "kube-api-access-n4jqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:19:38 crc kubenswrapper[4700]: I0227 18:19:38.150087 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4jqf\" (UniqueName: \"kubernetes.io/projected/90f1566e-7eeb-4455-9bcd-a9eed39c5eb1-kube-api-access-n4jqf\") on node \"crc\" DevicePath \"\"" Feb 27 18:19:38 crc kubenswrapper[4700]: I0227 18:19:38.408214 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lzfcp" Feb 27 18:19:38 crc kubenswrapper[4700]: I0227 18:19:38.409065 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lzfcp" Feb 27 18:19:38 crc kubenswrapper[4700]: I0227 18:19:38.512222 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536938-bppsz" event={"ID":"90f1566e-7eeb-4455-9bcd-a9eed39c5eb1","Type":"ContainerDied","Data":"aa6c28ea7fa5f5c6f3cd3806890a91b1521dc45844bf8c8510fef51f81200098"} Feb 27 18:19:38 crc kubenswrapper[4700]: I0227 18:19:38.512286 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa6c28ea7fa5f5c6f3cd3806890a91b1521dc45844bf8c8510fef51f81200098" Feb 27 18:19:38 crc kubenswrapper[4700]: I0227 18:19:38.512370 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536938-bppsz" Feb 27 18:19:38 crc kubenswrapper[4700]: I0227 18:19:38.560244 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536928-hphz9"] Feb 27 18:19:38 crc kubenswrapper[4700]: I0227 18:19:38.581819 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536928-hphz9"] Feb 27 18:19:39 crc kubenswrapper[4700]: I0227 18:19:39.007834 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88d71cab-ee75-4276-8c11-dd28e7ff49ef" path="/var/lib/kubelet/pods/88d71cab-ee75-4276-8c11-dd28e7ff49ef/volumes" Feb 27 18:19:39 crc kubenswrapper[4700]: I0227 18:19:39.009575 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e89023d8-dd8a-49ad-8749-5e54c469007d" path="/var/lib/kubelet/pods/e89023d8-dd8a-49ad-8749-5e54c469007d/volumes" Feb 27 18:19:40 crc kubenswrapper[4700]: I0227 18:19:40.321541 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-lzfcp" podUID="bc031a5d-e87f-4625-9b92-a7d0e539506e" containerName="registry-server" probeResult="failure" output=< Feb 27 18:19:40 crc kubenswrapper[4700]: timeout: failed to connect service ":50051" within 1s Feb 27 18:19:40 crc kubenswrapper[4700]: > Feb 27 18:19:41 crc kubenswrapper[4700]: E0227 18:19:41.984115 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:19:42 crc kubenswrapper[4700]: E0227 18:19:42.984636 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:19:44 crc kubenswrapper[4700]: E0227 18:19:44.985153 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:19:45 crc kubenswrapper[4700]: E0227 18:19:45.989781 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:19:48 crc kubenswrapper[4700]: I0227 18:19:48.505149 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lzfcp" Feb 27 18:19:48 crc kubenswrapper[4700]: I0227 18:19:48.579975 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lzfcp" Feb 27 18:19:54 crc kubenswrapper[4700]: E0227 18:19:54.985074 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:19:55 crc kubenswrapper[4700]: I0227 18:19:55.665377 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lzfcp"] Feb 27 18:19:55 crc kubenswrapper[4700]: I0227 18:19:55.665741 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lzfcp" podUID="bc031a5d-e87f-4625-9b92-a7d0e539506e" containerName="registry-server" containerID="cri-o://539c9833b4fd20b65e274a0ea65f56274d65da1aeccb49929bff2cab13221aa9" gracePeriod=2 Feb 27 18:19:55 crc kubenswrapper[4700]: E0227 18:19:55.983206 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:19:56 crc kubenswrapper[4700]: I0227 18:19:56.236161 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lzfcp" Feb 27 18:19:56 crc kubenswrapper[4700]: I0227 18:19:56.279507 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pz5gr\" (UniqueName: \"kubernetes.io/projected/bc031a5d-e87f-4625-9b92-a7d0e539506e-kube-api-access-pz5gr\") pod \"bc031a5d-e87f-4625-9b92-a7d0e539506e\" (UID: \"bc031a5d-e87f-4625-9b92-a7d0e539506e\") " Feb 27 18:19:56 crc kubenswrapper[4700]: I0227 18:19:56.279718 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc031a5d-e87f-4625-9b92-a7d0e539506e-utilities\") pod \"bc031a5d-e87f-4625-9b92-a7d0e539506e\" (UID: \"bc031a5d-e87f-4625-9b92-a7d0e539506e\") " Feb 27 18:19:56 crc kubenswrapper[4700]: I0227 18:19:56.279883 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc031a5d-e87f-4625-9b92-a7d0e539506e-catalog-content\") pod \"bc031a5d-e87f-4625-9b92-a7d0e539506e\" (UID: \"bc031a5d-e87f-4625-9b92-a7d0e539506e\") " Feb 27 18:19:56 crc kubenswrapper[4700]: I0227 18:19:56.281685 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc031a5d-e87f-4625-9b92-a7d0e539506e-utilities" (OuterVolumeSpecName: "utilities") pod "bc031a5d-e87f-4625-9b92-a7d0e539506e" (UID: "bc031a5d-e87f-4625-9b92-a7d0e539506e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 18:19:56 crc kubenswrapper[4700]: I0227 18:19:56.303599 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc031a5d-e87f-4625-9b92-a7d0e539506e-kube-api-access-pz5gr" (OuterVolumeSpecName: "kube-api-access-pz5gr") pod "bc031a5d-e87f-4625-9b92-a7d0e539506e" (UID: "bc031a5d-e87f-4625-9b92-a7d0e539506e"). InnerVolumeSpecName "kube-api-access-pz5gr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:19:56 crc kubenswrapper[4700]: I0227 18:19:56.382104 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pz5gr\" (UniqueName: \"kubernetes.io/projected/bc031a5d-e87f-4625-9b92-a7d0e539506e-kube-api-access-pz5gr\") on node \"crc\" DevicePath \"\"" Feb 27 18:19:56 crc kubenswrapper[4700]: I0227 18:19:56.382496 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc031a5d-e87f-4625-9b92-a7d0e539506e-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 18:19:56 crc kubenswrapper[4700]: I0227 18:19:56.386937 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc031a5d-e87f-4625-9b92-a7d0e539506e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bc031a5d-e87f-4625-9b92-a7d0e539506e" (UID: "bc031a5d-e87f-4625-9b92-a7d0e539506e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 18:19:56 crc kubenswrapper[4700]: I0227 18:19:56.484254 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc031a5d-e87f-4625-9b92-a7d0e539506e-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 18:19:56 crc kubenswrapper[4700]: I0227 18:19:56.783073 4700 generic.go:334] "Generic (PLEG): container finished" podID="bc031a5d-e87f-4625-9b92-a7d0e539506e" containerID="539c9833b4fd20b65e274a0ea65f56274d65da1aeccb49929bff2cab13221aa9" exitCode=0 Feb 27 18:19:56 crc kubenswrapper[4700]: I0227 18:19:56.783150 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lzfcp" event={"ID":"bc031a5d-e87f-4625-9b92-a7d0e539506e","Type":"ContainerDied","Data":"539c9833b4fd20b65e274a0ea65f56274d65da1aeccb49929bff2cab13221aa9"} Feb 27 18:19:56 crc kubenswrapper[4700]: I0227 18:19:56.783198 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lzfcp" event={"ID":"bc031a5d-e87f-4625-9b92-a7d0e539506e","Type":"ContainerDied","Data":"bdbed22a09c7b6460661f6e86bf335d05d92e494cd5c110b19581d34a98cc190"} Feb 27 18:19:56 crc kubenswrapper[4700]: I0227 18:19:56.783262 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lzfcp" Feb 27 18:19:56 crc kubenswrapper[4700]: I0227 18:19:56.783263 4700 scope.go:117] "RemoveContainer" containerID="539c9833b4fd20b65e274a0ea65f56274d65da1aeccb49929bff2cab13221aa9" Feb 27 18:19:56 crc kubenswrapper[4700]: I0227 18:19:56.824316 4700 scope.go:117] "RemoveContainer" containerID="5e02003e4306bbed4552ae2330440c9bd350e7b19eb9fb02626d2ff1464a649b" Feb 27 18:19:56 crc kubenswrapper[4700]: I0227 18:19:56.841648 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lzfcp"] Feb 27 18:19:56 crc kubenswrapper[4700]: I0227 18:19:56.861811 4700 scope.go:117] "RemoveContainer" containerID="85b204a3c1fd70456f22f3440f8d1014e104c1a8cbc3f47f3b7dc3703c125d13" Feb 27 18:19:56 crc kubenswrapper[4700]: I0227 18:19:56.863889 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lzfcp"] Feb 27 18:19:56 crc kubenswrapper[4700]: I0227 18:19:56.926978 4700 scope.go:117] "RemoveContainer" containerID="539c9833b4fd20b65e274a0ea65f56274d65da1aeccb49929bff2cab13221aa9" Feb 27 18:19:56 crc kubenswrapper[4700]: E0227 18:19:56.928011 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"539c9833b4fd20b65e274a0ea65f56274d65da1aeccb49929bff2cab13221aa9\": container with ID starting with 539c9833b4fd20b65e274a0ea65f56274d65da1aeccb49929bff2cab13221aa9 not found: ID does not exist" containerID="539c9833b4fd20b65e274a0ea65f56274d65da1aeccb49929bff2cab13221aa9" Feb 27 18:19:56 crc kubenswrapper[4700]: I0227 18:19:56.928074 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"539c9833b4fd20b65e274a0ea65f56274d65da1aeccb49929bff2cab13221aa9"} err="failed to get container status \"539c9833b4fd20b65e274a0ea65f56274d65da1aeccb49929bff2cab13221aa9\": rpc error: code = NotFound desc = could not find container \"539c9833b4fd20b65e274a0ea65f56274d65da1aeccb49929bff2cab13221aa9\": container with ID starting with 539c9833b4fd20b65e274a0ea65f56274d65da1aeccb49929bff2cab13221aa9 not found: ID does not exist" Feb 27 18:19:56 crc kubenswrapper[4700]: I0227 18:19:56.928116 4700 scope.go:117] "RemoveContainer" containerID="5e02003e4306bbed4552ae2330440c9bd350e7b19eb9fb02626d2ff1464a649b" Feb 27 18:19:56 crc kubenswrapper[4700]: E0227 18:19:56.928787 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e02003e4306bbed4552ae2330440c9bd350e7b19eb9fb02626d2ff1464a649b\": container with ID starting with 5e02003e4306bbed4552ae2330440c9bd350e7b19eb9fb02626d2ff1464a649b not found: ID does not exist" containerID="5e02003e4306bbed4552ae2330440c9bd350e7b19eb9fb02626d2ff1464a649b" Feb 27 18:19:56 crc kubenswrapper[4700]: I0227 18:19:56.928949 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e02003e4306bbed4552ae2330440c9bd350e7b19eb9fb02626d2ff1464a649b"} err="failed to get container status \"5e02003e4306bbed4552ae2330440c9bd350e7b19eb9fb02626d2ff1464a649b\": rpc error: code = NotFound desc = could not find container \"5e02003e4306bbed4552ae2330440c9bd350e7b19eb9fb02626d2ff1464a649b\": container with ID starting with 5e02003e4306bbed4552ae2330440c9bd350e7b19eb9fb02626d2ff1464a649b not found: ID does not exist" Feb 27 18:19:56 crc kubenswrapper[4700]: I0227 18:19:56.929005 4700 scope.go:117] "RemoveContainer" containerID="85b204a3c1fd70456f22f3440f8d1014e104c1a8cbc3f47f3b7dc3703c125d13" Feb 27 18:19:56 crc kubenswrapper[4700]: E0227 18:19:56.929775 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85b204a3c1fd70456f22f3440f8d1014e104c1a8cbc3f47f3b7dc3703c125d13\": container with ID starting with 85b204a3c1fd70456f22f3440f8d1014e104c1a8cbc3f47f3b7dc3703c125d13 not found: ID does not exist" containerID="85b204a3c1fd70456f22f3440f8d1014e104c1a8cbc3f47f3b7dc3703c125d13" Feb 27 18:19:56 crc kubenswrapper[4700]: I0227 18:19:56.929841 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85b204a3c1fd70456f22f3440f8d1014e104c1a8cbc3f47f3b7dc3703c125d13"} err="failed to get container status \"85b204a3c1fd70456f22f3440f8d1014e104c1a8cbc3f47f3b7dc3703c125d13\": rpc error: code = NotFound desc = could not find container \"85b204a3c1fd70456f22f3440f8d1014e104c1a8cbc3f47f3b7dc3703c125d13\": container with ID starting with 85b204a3c1fd70456f22f3440f8d1014e104c1a8cbc3f47f3b7dc3703c125d13 not found: ID does not exist" Feb 27 18:19:56 crc kubenswrapper[4700]: E0227 18:19:56.984078 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:19:56 crc kubenswrapper[4700]: I0227 18:19:56.997860 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc031a5d-e87f-4625-9b92-a7d0e539506e" path="/var/lib/kubelet/pods/bc031a5d-e87f-4625-9b92-a7d0e539506e/volumes" Feb 27 18:19:58 crc kubenswrapper[4700]: E0227 18:19:58.291308 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 18:19:58 crc kubenswrapper[4700]: E0227 18:19:58.292659 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 18:19:58 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 18:19:58 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qkjfz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536934-rw77f_openshift-infra(caa4e57d-b02d-4d65-944d-451b8294f884): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 18:19:58 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 18:19:58 crc kubenswrapper[4700]: E0227 18:19:58.294118 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:20:00 crc kubenswrapper[4700]: I0227 18:20:00.159160 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536940-2ffgp"] Feb 27 18:20:00 crc kubenswrapper[4700]: E0227 18:20:00.160329 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc031a5d-e87f-4625-9b92-a7d0e539506e" containerName="extract-utilities" Feb 27 18:20:00 crc kubenswrapper[4700]: I0227 18:20:00.160346 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc031a5d-e87f-4625-9b92-a7d0e539506e" containerName="extract-utilities" Feb 27 18:20:00 crc kubenswrapper[4700]: E0227 18:20:00.160387 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90f1566e-7eeb-4455-9bcd-a9eed39c5eb1" containerName="oc" Feb 27 18:20:00 crc kubenswrapper[4700]: I0227 18:20:00.160396 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="90f1566e-7eeb-4455-9bcd-a9eed39c5eb1" containerName="oc" Feb 27 18:20:00 crc kubenswrapper[4700]: E0227 18:20:00.160409 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc031a5d-e87f-4625-9b92-a7d0e539506e" containerName="extract-content" Feb 27 18:20:00 crc kubenswrapper[4700]: I0227 18:20:00.160419 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc031a5d-e87f-4625-9b92-a7d0e539506e" containerName="extract-content" Feb 27 18:20:00 crc kubenswrapper[4700]: E0227 18:20:00.160431 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc031a5d-e87f-4625-9b92-a7d0e539506e" containerName="registry-server" Feb 27 18:20:00 crc kubenswrapper[4700]: I0227 18:20:00.160439 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc031a5d-e87f-4625-9b92-a7d0e539506e" containerName="registry-server" Feb 27 18:20:00 crc kubenswrapper[4700]: E0227 18:20:00.160449 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88d71cab-ee75-4276-8c11-dd28e7ff49ef" containerName="registry-server" Feb 27 18:20:00 crc kubenswrapper[4700]: I0227 18:20:00.160477 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="88d71cab-ee75-4276-8c11-dd28e7ff49ef" containerName="registry-server" Feb 27 18:20:00 crc kubenswrapper[4700]: E0227 18:20:00.160495 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88d71cab-ee75-4276-8c11-dd28e7ff49ef" containerName="extract-utilities" Feb 27 18:20:00 crc kubenswrapper[4700]: I0227 18:20:00.160506 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="88d71cab-ee75-4276-8c11-dd28e7ff49ef" containerName="extract-utilities" Feb 27 18:20:00 crc kubenswrapper[4700]: E0227 18:20:00.160539 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88d71cab-ee75-4276-8c11-dd28e7ff49ef" containerName="extract-content" Feb 27 18:20:00 crc kubenswrapper[4700]: I0227 18:20:00.160546 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="88d71cab-ee75-4276-8c11-dd28e7ff49ef" containerName="extract-content" Feb 27 18:20:00 crc kubenswrapper[4700]: I0227 18:20:00.160798 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="90f1566e-7eeb-4455-9bcd-a9eed39c5eb1" containerName="oc" Feb 27 18:20:00 crc kubenswrapper[4700]: I0227 18:20:00.160825 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc031a5d-e87f-4625-9b92-a7d0e539506e" containerName="registry-server" Feb 27 18:20:00 crc kubenswrapper[4700]: I0227 18:20:00.160848 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="88d71cab-ee75-4276-8c11-dd28e7ff49ef" containerName="registry-server" Feb 27 18:20:00 crc kubenswrapper[4700]: I0227 18:20:00.161664 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536940-2ffgp" Feb 27 18:20:00 crc kubenswrapper[4700]: I0227 18:20:00.191293 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536940-2ffgp"] Feb 27 18:20:00 crc kubenswrapper[4700]: I0227 18:20:00.325509 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2ws2\" (UniqueName: \"kubernetes.io/projected/fdeb63e9-124e-40bb-a865-2463bba1aef2-kube-api-access-b2ws2\") pod \"auto-csr-approver-29536940-2ffgp\" (UID: \"fdeb63e9-124e-40bb-a865-2463bba1aef2\") " pod="openshift-infra/auto-csr-approver-29536940-2ffgp" Feb 27 18:20:00 crc kubenswrapper[4700]: I0227 18:20:00.429202 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2ws2\" (UniqueName: \"kubernetes.io/projected/fdeb63e9-124e-40bb-a865-2463bba1aef2-kube-api-access-b2ws2\") pod \"auto-csr-approver-29536940-2ffgp\" (UID: \"fdeb63e9-124e-40bb-a865-2463bba1aef2\") " pod="openshift-infra/auto-csr-approver-29536940-2ffgp" Feb 27 18:20:00 crc kubenswrapper[4700]: I0227 18:20:00.456365 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2ws2\" (UniqueName: \"kubernetes.io/projected/fdeb63e9-124e-40bb-a865-2463bba1aef2-kube-api-access-b2ws2\") pod \"auto-csr-approver-29536940-2ffgp\" (UID: \"fdeb63e9-124e-40bb-a865-2463bba1aef2\") " pod="openshift-infra/auto-csr-approver-29536940-2ffgp" Feb 27 18:20:00 crc kubenswrapper[4700]: I0227 18:20:00.498700 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536940-2ffgp" Feb 27 18:20:01 crc kubenswrapper[4700]: I0227 18:20:01.061817 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536940-2ffgp"] Feb 27 18:20:01 crc kubenswrapper[4700]: W0227 18:20:01.067484 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfdeb63e9_124e_40bb_a865_2463bba1aef2.slice/crio-1c63b80875eaf1674257e133948e90bca3de387ebb5c02cc93f0dc6514a2d0b7 WatchSource:0}: Error finding container 1c63b80875eaf1674257e133948e90bca3de387ebb5c02cc93f0dc6514a2d0b7: Status 404 returned error can't find the container with id 1c63b80875eaf1674257e133948e90bca3de387ebb5c02cc93f0dc6514a2d0b7 Feb 27 18:20:01 crc kubenswrapper[4700]: I0227 18:20:01.847223 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536940-2ffgp" event={"ID":"fdeb63e9-124e-40bb-a865-2463bba1aef2","Type":"ContainerStarted","Data":"1c63b80875eaf1674257e133948e90bca3de387ebb5c02cc93f0dc6514a2d0b7"} Feb 27 18:20:02 crc kubenswrapper[4700]: I0227 18:20:02.866612 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536940-2ffgp" event={"ID":"fdeb63e9-124e-40bb-a865-2463bba1aef2","Type":"ContainerStarted","Data":"626b83a35710ba8583357907106bf3ee7bf7f041e6cfd87b21a5822ae18d3c2e"} Feb 27 18:20:02 crc kubenswrapper[4700]: I0227 18:20:02.900841 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536940-2ffgp" podStartSLOduration=1.573291886 podStartE2EDuration="2.900818825s" podCreationTimestamp="2026-02-27 18:20:00 +0000 UTC" firstStartedPulling="2026-02-27 18:20:01.070441838 +0000 UTC m=+4761.055754585" lastFinishedPulling="2026-02-27 18:20:02.397968737 +0000 UTC m=+4762.383281524" observedRunningTime="2026-02-27 18:20:02.887606691 +0000 UTC m=+4762.872919448" watchObservedRunningTime="2026-02-27 18:20:02.900818825 +0000 UTC m=+4762.886131582" Feb 27 18:20:03 crc kubenswrapper[4700]: I0227 18:20:03.885533 4700 generic.go:334] "Generic (PLEG): container finished" podID="fdeb63e9-124e-40bb-a865-2463bba1aef2" containerID="626b83a35710ba8583357907106bf3ee7bf7f041e6cfd87b21a5822ae18d3c2e" exitCode=0 Feb 27 18:20:03 crc kubenswrapper[4700]: I0227 18:20:03.885655 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536940-2ffgp" event={"ID":"fdeb63e9-124e-40bb-a865-2463bba1aef2","Type":"ContainerDied","Data":"626b83a35710ba8583357907106bf3ee7bf7f041e6cfd87b21a5822ae18d3c2e"} Feb 27 18:20:05 crc kubenswrapper[4700]: I0227 18:20:05.375041 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536940-2ffgp" Feb 27 18:20:05 crc kubenswrapper[4700]: I0227 18:20:05.495069 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2ws2\" (UniqueName: \"kubernetes.io/projected/fdeb63e9-124e-40bb-a865-2463bba1aef2-kube-api-access-b2ws2\") pod \"fdeb63e9-124e-40bb-a865-2463bba1aef2\" (UID: \"fdeb63e9-124e-40bb-a865-2463bba1aef2\") " Feb 27 18:20:05 crc kubenswrapper[4700]: I0227 18:20:05.507629 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdeb63e9-124e-40bb-a865-2463bba1aef2-kube-api-access-b2ws2" (OuterVolumeSpecName: "kube-api-access-b2ws2") pod "fdeb63e9-124e-40bb-a865-2463bba1aef2" (UID: "fdeb63e9-124e-40bb-a865-2463bba1aef2"). InnerVolumeSpecName "kube-api-access-b2ws2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:20:05 crc kubenswrapper[4700]: I0227 18:20:05.599378 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2ws2\" (UniqueName: \"kubernetes.io/projected/fdeb63e9-124e-40bb-a865-2463bba1aef2-kube-api-access-b2ws2\") on node \"crc\" DevicePath \"\"" Feb 27 18:20:05 crc kubenswrapper[4700]: I0227 18:20:05.927840 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536940-2ffgp" event={"ID":"fdeb63e9-124e-40bb-a865-2463bba1aef2","Type":"ContainerDied","Data":"1c63b80875eaf1674257e133948e90bca3de387ebb5c02cc93f0dc6514a2d0b7"} Feb 27 18:20:05 crc kubenswrapper[4700]: I0227 18:20:05.927933 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c63b80875eaf1674257e133948e90bca3de387ebb5c02cc93f0dc6514a2d0b7" Feb 27 18:20:05 crc kubenswrapper[4700]: I0227 18:20:05.927962 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536940-2ffgp" Feb 27 18:20:06 crc kubenswrapper[4700]: I0227 18:20:06.021249 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536932-l4s4p"] Feb 27 18:20:06 crc kubenswrapper[4700]: I0227 18:20:06.035822 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536932-l4s4p"] Feb 27 18:20:06 crc kubenswrapper[4700]: E0227 18:20:06.988205 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:20:07 crc kubenswrapper[4700]: I0227 18:20:07.002687 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d8e8a24-2862-4552-82c1-3b445fb6628d" path="/var/lib/kubelet/pods/3d8e8a24-2862-4552-82c1-3b445fb6628d/volumes" Feb 27 18:20:09 crc kubenswrapper[4700]: E0227 18:20:09.984420 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:20:09 crc kubenswrapper[4700]: E0227 18:20:09.985183 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:20:11 crc kubenswrapper[4700]: E0227 18:20:11.983572 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:20:12 crc kubenswrapper[4700]: I0227 18:20:12.506449 4700 scope.go:117] "RemoveContainer" containerID="d81845e950d957edbe8fade1a8eb3a28a30ca213b35a6ec2f6568db90f5b452c" Feb 27 18:20:12 crc kubenswrapper[4700]: I0227 18:20:12.599269 4700 scope.go:117] "RemoveContainer" containerID="6011bbefc46aa294432da75f46bd16d87194a34721c8efb015a8d11d76420056" Feb 27 18:20:19 crc kubenswrapper[4700]: E0227 18:20:19.983591 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:20:20 crc kubenswrapper[4700]: E0227 18:20:20.994282 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:20:23 crc kubenswrapper[4700]: E0227 18:20:23.984734 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:20:25 crc kubenswrapper[4700]: E0227 18:20:25.990481 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:20:32 crc kubenswrapper[4700]: E0227 18:20:32.984409 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:20:35 crc kubenswrapper[4700]: E0227 18:20:35.985589 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:20:36 crc kubenswrapper[4700]: E0227 18:20:36.984652 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:20:37 crc kubenswrapper[4700]: E0227 18:20:37.982418 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:20:46 crc kubenswrapper[4700]: E0227 18:20:46.985262 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:20:47 crc kubenswrapper[4700]: E0227 18:20:47.985906 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:20:48 crc kubenswrapper[4700]: E0227 18:20:48.984132 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:20:49 crc kubenswrapper[4700]: E0227 18:20:49.983402 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:20:58 crc kubenswrapper[4700]: E0227 18:20:58.984448 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:20:59 crc kubenswrapper[4700]: E0227 18:20:59.983539 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:21:01 crc kubenswrapper[4700]: E0227 18:21:01.986014 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:21:02 crc kubenswrapper[4700]: E0227 18:21:02.983727 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:21:11 crc kubenswrapper[4700]: E0227 18:21:11.002309 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:21:11 crc kubenswrapper[4700]: E0227 18:21:11.984544 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:21:12 crc kubenswrapper[4700]: E0227 18:21:12.984585 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:21:15 crc kubenswrapper[4700]: E0227 18:21:15.986401 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:21:24 crc kubenswrapper[4700]: E0227 18:21:24.986059 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:21:26 crc kubenswrapper[4700]: E0227 18:21:26.985518 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:21:27 crc kubenswrapper[4700]: E0227 18:21:27.985443 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:21:36 crc kubenswrapper[4700]: I0227 18:21:36.411074 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 18:21:36 crc kubenswrapper[4700]: I0227 18:21:36.411858 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 18:21:38 crc kubenswrapper[4700]: E0227 18:21:38.984855 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:21:39 crc kubenswrapper[4700]: E0227 18:21:39.985800 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:21:43 crc kubenswrapper[4700]: E0227 18:21:43.292436 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:21:53 crc kubenswrapper[4700]: E0227 18:21:53.984041 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:21:53 crc kubenswrapper[4700]: E0227 18:21:53.984134 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:21:56 crc kubenswrapper[4700]: E0227 18:21:56.984478 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:22:00 crc kubenswrapper[4700]: I0227 18:22:00.166704 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536942-st4kg"] Feb 27 18:22:00 crc kubenswrapper[4700]: E0227 18:22:00.168035 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdeb63e9-124e-40bb-a865-2463bba1aef2" containerName="oc" Feb 27 18:22:00 crc kubenswrapper[4700]: I0227 18:22:00.168061 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdeb63e9-124e-40bb-a865-2463bba1aef2" containerName="oc" Feb 27 18:22:00 crc kubenswrapper[4700]: I0227 18:22:00.168410 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdeb63e9-124e-40bb-a865-2463bba1aef2" containerName="oc" Feb 27 18:22:00 crc kubenswrapper[4700]: I0227 18:22:00.169595 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536942-st4kg" Feb 27 18:22:00 crc kubenswrapper[4700]: I0227 18:22:00.182144 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536942-st4kg"] Feb 27 18:22:00 crc kubenswrapper[4700]: I0227 18:22:00.262542 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvr97\" (UniqueName: \"kubernetes.io/projected/237d2183-87cf-4196-944e-ddbda807b62b-kube-api-access-wvr97\") pod \"auto-csr-approver-29536942-st4kg\" (UID: \"237d2183-87cf-4196-944e-ddbda807b62b\") " pod="openshift-infra/auto-csr-approver-29536942-st4kg" Feb 27 18:22:00 crc kubenswrapper[4700]: I0227 18:22:00.365353 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvr97\" (UniqueName: \"kubernetes.io/projected/237d2183-87cf-4196-944e-ddbda807b62b-kube-api-access-wvr97\") pod \"auto-csr-approver-29536942-st4kg\" (UID: \"237d2183-87cf-4196-944e-ddbda807b62b\") " pod="openshift-infra/auto-csr-approver-29536942-st4kg" Feb 27 18:22:00 crc kubenswrapper[4700]: I0227 18:22:00.397556 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvr97\" (UniqueName: \"kubernetes.io/projected/237d2183-87cf-4196-944e-ddbda807b62b-kube-api-access-wvr97\") pod \"auto-csr-approver-29536942-st4kg\" (UID: \"237d2183-87cf-4196-944e-ddbda807b62b\") " pod="openshift-infra/auto-csr-approver-29536942-st4kg" Feb 27 18:22:00 crc kubenswrapper[4700]: I0227 18:22:00.492480 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536942-st4kg" Feb 27 18:22:01 crc kubenswrapper[4700]: I0227 18:22:01.009232 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536942-st4kg"] Feb 27 18:22:01 crc kubenswrapper[4700]: I0227 18:22:01.020828 4700 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 18:22:01 crc kubenswrapper[4700]: I0227 18:22:01.503292 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536942-st4kg" event={"ID":"237d2183-87cf-4196-944e-ddbda807b62b","Type":"ContainerStarted","Data":"d551231e7e4f4775a0eb15a70731671ccdd011200a1c99bd1cde49b67511b349"} Feb 27 18:22:03 crc kubenswrapper[4700]: I0227 18:22:03.547640 4700 generic.go:334] "Generic (PLEG): container finished" podID="237d2183-87cf-4196-944e-ddbda807b62b" containerID="cdff47536c8d552a075a8e0a9cf511f2b71c81f2d61d3cb94c490b7d1f5176ee" exitCode=0 Feb 27 18:22:03 crc kubenswrapper[4700]: I0227 18:22:03.547623 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536942-st4kg" event={"ID":"237d2183-87cf-4196-944e-ddbda807b62b","Type":"ContainerDied","Data":"cdff47536c8d552a075a8e0a9cf511f2b71c81f2d61d3cb94c490b7d1f5176ee"} Feb 27 18:22:05 crc kubenswrapper[4700]: I0227 18:22:05.703228 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536942-st4kg" Feb 27 18:22:05 crc kubenswrapper[4700]: I0227 18:22:05.797822 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvr97\" (UniqueName: \"kubernetes.io/projected/237d2183-87cf-4196-944e-ddbda807b62b-kube-api-access-wvr97\") pod \"237d2183-87cf-4196-944e-ddbda807b62b\" (UID: \"237d2183-87cf-4196-944e-ddbda807b62b\") " Feb 27 18:22:05 crc kubenswrapper[4700]: I0227 18:22:05.804778 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/237d2183-87cf-4196-944e-ddbda807b62b-kube-api-access-wvr97" (OuterVolumeSpecName: "kube-api-access-wvr97") pod "237d2183-87cf-4196-944e-ddbda807b62b" (UID: "237d2183-87cf-4196-944e-ddbda807b62b"). InnerVolumeSpecName "kube-api-access-wvr97". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:22:05 crc kubenswrapper[4700]: I0227 18:22:05.901015 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvr97\" (UniqueName: \"kubernetes.io/projected/237d2183-87cf-4196-944e-ddbda807b62b-kube-api-access-wvr97\") on node \"crc\" DevicePath \"\"" Feb 27 18:22:05 crc kubenswrapper[4700]: E0227 18:22:05.984494 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:22:06 crc kubenswrapper[4700]: I0227 18:22:06.411100 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 18:22:06 crc kubenswrapper[4700]: I0227 18:22:06.411174 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 18:22:06 crc kubenswrapper[4700]: I0227 18:22:06.594007 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536942-st4kg" event={"ID":"237d2183-87cf-4196-944e-ddbda807b62b","Type":"ContainerDied","Data":"d551231e7e4f4775a0eb15a70731671ccdd011200a1c99bd1cde49b67511b349"} Feb 27 18:22:06 crc kubenswrapper[4700]: I0227 18:22:06.594095 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d551231e7e4f4775a0eb15a70731671ccdd011200a1c99bd1cde49b67511b349" Feb 27 18:22:06 crc kubenswrapper[4700]: I0227 18:22:06.594152 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536942-st4kg" Feb 27 18:22:06 crc kubenswrapper[4700]: I0227 18:22:06.808654 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536936-qwrfm"] Feb 27 18:22:06 crc kubenswrapper[4700]: I0227 18:22:06.819404 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536936-qwrfm"] Feb 27 18:22:06 crc kubenswrapper[4700]: E0227 18:22:06.985693 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:22:06 crc kubenswrapper[4700]: I0227 18:22:06.999593 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="951ae05f-f758-448b-9f39-07ddedc9e642" path="/var/lib/kubelet/pods/951ae05f-f758-448b-9f39-07ddedc9e642/volumes" Feb 27 18:22:11 crc kubenswrapper[4700]: E0227 18:22:11.985180 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:22:12 crc kubenswrapper[4700]: I0227 18:22:12.836914 4700 scope.go:117] "RemoveContainer" containerID="464dbbb1633fded359d0f8206112c4ec17cd067a5cf5684b646af06470e6d45a" Feb 27 18:22:20 crc kubenswrapper[4700]: E0227 18:22:20.005088 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:22:20 crc kubenswrapper[4700]: E0227 18:22:20.997666 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" Feb 27 18:22:27 crc kubenswrapper[4700]: E0227 18:22:27.137705 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 18:22:27 crc kubenswrapper[4700]: E0227 18:22:27.138276 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 18:22:27 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 18:22:27 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gtm8j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536930-xbg7f_openshift-infra(3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 18:22:27 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 18:22:27 crc kubenswrapper[4700]: E0227 18:22:27.139623 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:22:27 crc kubenswrapper[4700]: E0227 18:22:27.821939 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 18:22:27 crc kubenswrapper[4700]: E0227 18:22:27.822242 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 18:22:27 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 18:22:27 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hjvdh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536926-s6ft8_openshift-infra(d37cbaaf-1dee-4efd-9327-7bee80986a35): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 18:22:27 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 18:22:27 crc kubenswrapper[4700]: E0227 18:22:27.824204 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:22:31 crc kubenswrapper[4700]: E0227 18:22:31.983319 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:22:34 crc kubenswrapper[4700]: I0227 18:22:34.970351 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d86g8" event={"ID":"243afaab-17c1-4631-974c-f7b693236b08","Type":"ContainerStarted","Data":"424d19170d51b0ec2a2e662c79027264e27a58bed662fc690382cf464322fc63"} Feb 27 18:22:36 crc kubenswrapper[4700]: I0227 18:22:36.001116 4700 generic.go:334] "Generic (PLEG): container finished" podID="243afaab-17c1-4631-974c-f7b693236b08" containerID="424d19170d51b0ec2a2e662c79027264e27a58bed662fc690382cf464322fc63" exitCode=0 Feb 27 18:22:36 crc kubenswrapper[4700]: I0227 18:22:36.001378 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d86g8" event={"ID":"243afaab-17c1-4631-974c-f7b693236b08","Type":"ContainerDied","Data":"424d19170d51b0ec2a2e662c79027264e27a58bed662fc690382cf464322fc63"} Feb 27 18:22:36 crc kubenswrapper[4700]: I0227 18:22:36.410765 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 18:22:36 crc kubenswrapper[4700]: I0227 18:22:36.410888 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 18:22:36 crc kubenswrapper[4700]: I0227 18:22:36.410978 4700 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" Feb 27 18:22:36 crc kubenswrapper[4700]: I0227 18:22:36.412829 4700 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d616594fff3f3a84ae3299a2a56f70ba23d3b8c11784f6e69c706e5bb75192a5"} pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 18:22:36 crc kubenswrapper[4700]: I0227 18:22:36.412929 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" containerID="cri-o://d616594fff3f3a84ae3299a2a56f70ba23d3b8c11784f6e69c706e5bb75192a5" gracePeriod=600 Feb 27 18:22:37 crc kubenswrapper[4700]: I0227 18:22:37.020764 4700 generic.go:334] "Generic (PLEG): container finished" podID="84b45511-b94c-479f-98db-fd2c4eceec46" containerID="d616594fff3f3a84ae3299a2a56f70ba23d3b8c11784f6e69c706e5bb75192a5" exitCode=0 Feb 27 18:22:37 crc kubenswrapper[4700]: I0227 18:22:37.020837 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerDied","Data":"d616594fff3f3a84ae3299a2a56f70ba23d3b8c11784f6e69c706e5bb75192a5"} Feb 27 18:22:37 crc kubenswrapper[4700]: I0227 18:22:37.020888 4700 scope.go:117] "RemoveContainer" containerID="454da4538d721f5f5da4ce85346adc3fbf77394a4e27f80b9bc979aa3bc37aa2" Feb 27 18:22:37 crc kubenswrapper[4700]: E0227 18:22:37.984814 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:22:38 crc kubenswrapper[4700]: I0227 18:22:38.033711 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerStarted","Data":"b3fdfb06db2c06428e3b3e96ef9804be97292a50a8f04f891caec5e729731adf"} Feb 27 18:22:38 crc kubenswrapper[4700]: I0227 18:22:38.037333 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d86g8" event={"ID":"243afaab-17c1-4631-974c-f7b693236b08","Type":"ContainerStarted","Data":"1997a0abf345d10034cfa408684ada4628e4101c046391ef96dc8d365163248c"} Feb 27 18:22:38 crc kubenswrapper[4700]: I0227 18:22:38.077125 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-d86g8" podStartSLOduration=3.78117146 podStartE2EDuration="11m6.077103886s" podCreationTimestamp="2026-02-27 18:11:32 +0000 UTC" firstStartedPulling="2026-02-27 18:11:34.170387338 +0000 UTC m=+4254.155700095" lastFinishedPulling="2026-02-27 18:22:36.466319734 +0000 UTC m=+4916.451632521" observedRunningTime="2026-02-27 18:22:38.068859281 +0000 UTC m=+4918.054172058" watchObservedRunningTime="2026-02-27 18:22:38.077103886 +0000 UTC m=+4918.062416633" Feb 27 18:22:41 crc kubenswrapper[4700]: E0227 18:22:41.013249 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:22:42 crc kubenswrapper[4700]: E0227 18:22:42.984668 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:22:43 crc kubenswrapper[4700]: I0227 18:22:43.361084 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-d86g8" Feb 27 18:22:43 crc kubenswrapper[4700]: I0227 18:22:43.361152 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-d86g8" Feb 27 18:22:43 crc kubenswrapper[4700]: I0227 18:22:43.460058 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-d86g8" Feb 27 18:22:44 crc kubenswrapper[4700]: I0227 18:22:44.206907 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-d86g8" Feb 27 18:22:44 crc kubenswrapper[4700]: I0227 18:22:44.285142 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d86g8"] Feb 27 18:22:46 crc kubenswrapper[4700]: I0227 18:22:46.137291 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-d86g8" podUID="243afaab-17c1-4631-974c-f7b693236b08" containerName="registry-server" containerID="cri-o://1997a0abf345d10034cfa408684ada4628e4101c046391ef96dc8d365163248c" gracePeriod=2 Feb 27 18:22:46 crc kubenswrapper[4700]: I0227 18:22:46.657606 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d86g8" Feb 27 18:22:46 crc kubenswrapper[4700]: I0227 18:22:46.834595 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vkwzw\" (UniqueName: \"kubernetes.io/projected/243afaab-17c1-4631-974c-f7b693236b08-kube-api-access-vkwzw\") pod \"243afaab-17c1-4631-974c-f7b693236b08\" (UID: \"243afaab-17c1-4631-974c-f7b693236b08\") " Feb 27 18:22:46 crc kubenswrapper[4700]: I0227 18:22:46.834663 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/243afaab-17c1-4631-974c-f7b693236b08-catalog-content\") pod \"243afaab-17c1-4631-974c-f7b693236b08\" (UID: \"243afaab-17c1-4631-974c-f7b693236b08\") " Feb 27 18:22:46 crc kubenswrapper[4700]: I0227 18:22:46.834703 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/243afaab-17c1-4631-974c-f7b693236b08-utilities\") pod \"243afaab-17c1-4631-974c-f7b693236b08\" (UID: \"243afaab-17c1-4631-974c-f7b693236b08\") " Feb 27 18:22:46 crc kubenswrapper[4700]: I0227 18:22:46.835830 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/243afaab-17c1-4631-974c-f7b693236b08-utilities" (OuterVolumeSpecName: "utilities") pod "243afaab-17c1-4631-974c-f7b693236b08" (UID: "243afaab-17c1-4631-974c-f7b693236b08"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 18:22:46 crc kubenswrapper[4700]: I0227 18:22:46.843029 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/243afaab-17c1-4631-974c-f7b693236b08-kube-api-access-vkwzw" (OuterVolumeSpecName: "kube-api-access-vkwzw") pod "243afaab-17c1-4631-974c-f7b693236b08" (UID: "243afaab-17c1-4631-974c-f7b693236b08"). InnerVolumeSpecName "kube-api-access-vkwzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:22:46 crc kubenswrapper[4700]: I0227 18:22:46.900095 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/243afaab-17c1-4631-974c-f7b693236b08-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "243afaab-17c1-4631-974c-f7b693236b08" (UID: "243afaab-17c1-4631-974c-f7b693236b08"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 18:22:46 crc kubenswrapper[4700]: I0227 18:22:46.937205 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vkwzw\" (UniqueName: \"kubernetes.io/projected/243afaab-17c1-4631-974c-f7b693236b08-kube-api-access-vkwzw\") on node \"crc\" DevicePath \"\"" Feb 27 18:22:46 crc kubenswrapper[4700]: I0227 18:22:46.937243 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/243afaab-17c1-4631-974c-f7b693236b08-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 18:22:46 crc kubenswrapper[4700]: I0227 18:22:46.937256 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/243afaab-17c1-4631-974c-f7b693236b08-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 18:22:47 crc kubenswrapper[4700]: I0227 18:22:47.161342 4700 generic.go:334] "Generic (PLEG): container finished" podID="243afaab-17c1-4631-974c-f7b693236b08" containerID="1997a0abf345d10034cfa408684ada4628e4101c046391ef96dc8d365163248c" exitCode=0 Feb 27 18:22:47 crc kubenswrapper[4700]: I0227 18:22:47.161496 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d86g8" Feb 27 18:22:47 crc kubenswrapper[4700]: I0227 18:22:47.161510 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d86g8" event={"ID":"243afaab-17c1-4631-974c-f7b693236b08","Type":"ContainerDied","Data":"1997a0abf345d10034cfa408684ada4628e4101c046391ef96dc8d365163248c"} Feb 27 18:22:47 crc kubenswrapper[4700]: I0227 18:22:47.162053 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d86g8" event={"ID":"243afaab-17c1-4631-974c-f7b693236b08","Type":"ContainerDied","Data":"16257cbd86f69bef2cc6b3c9a37f67f6da2159e9766be7be5d802dee309affd4"} Feb 27 18:22:47 crc kubenswrapper[4700]: I0227 18:22:47.162078 4700 scope.go:117] "RemoveContainer" containerID="1997a0abf345d10034cfa408684ada4628e4101c046391ef96dc8d365163248c" Feb 27 18:22:47 crc kubenswrapper[4700]: I0227 18:22:47.196561 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d86g8"] Feb 27 18:22:47 crc kubenswrapper[4700]: I0227 18:22:47.198961 4700 scope.go:117] "RemoveContainer" containerID="424d19170d51b0ec2a2e662c79027264e27a58bed662fc690382cf464322fc63" Feb 27 18:22:47 crc kubenswrapper[4700]: I0227 18:22:47.205900 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-d86g8"] Feb 27 18:22:47 crc kubenswrapper[4700]: I0227 18:22:47.231704 4700 scope.go:117] "RemoveContainer" containerID="3b8324dc261d79c510a5a6fe4498764a5ccfedc2a55e39d88d50dce728b771f4" Feb 27 18:22:47 crc kubenswrapper[4700]: I0227 18:22:47.259620 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tn4s8"] Feb 27 18:22:47 crc kubenswrapper[4700]: E0227 18:22:47.260318 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="243afaab-17c1-4631-974c-f7b693236b08" containerName="registry-server" Feb 27 18:22:47 crc kubenswrapper[4700]: I0227 18:22:47.260337 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="243afaab-17c1-4631-974c-f7b693236b08" containerName="registry-server" Feb 27 18:22:47 crc kubenswrapper[4700]: E0227 18:22:47.260348 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="237d2183-87cf-4196-944e-ddbda807b62b" containerName="oc" Feb 27 18:22:47 crc kubenswrapper[4700]: I0227 18:22:47.260358 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="237d2183-87cf-4196-944e-ddbda807b62b" containerName="oc" Feb 27 18:22:47 crc kubenswrapper[4700]: E0227 18:22:47.260372 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="243afaab-17c1-4631-974c-f7b693236b08" containerName="extract-utilities" Feb 27 18:22:47 crc kubenswrapper[4700]: I0227 18:22:47.260379 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="243afaab-17c1-4631-974c-f7b693236b08" containerName="extract-utilities" Feb 27 18:22:47 crc kubenswrapper[4700]: E0227 18:22:47.260393 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="243afaab-17c1-4631-974c-f7b693236b08" containerName="extract-content" Feb 27 18:22:47 crc kubenswrapper[4700]: I0227 18:22:47.260399 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="243afaab-17c1-4631-974c-f7b693236b08" containerName="extract-content" Feb 27 18:22:47 crc kubenswrapper[4700]: I0227 18:22:47.260610 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="243afaab-17c1-4631-974c-f7b693236b08" containerName="registry-server" Feb 27 18:22:47 crc kubenswrapper[4700]: I0227 18:22:47.260637 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="237d2183-87cf-4196-944e-ddbda807b62b" containerName="oc" Feb 27 18:22:47 crc kubenswrapper[4700]: I0227 18:22:47.262093 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tn4s8" Feb 27 18:22:47 crc kubenswrapper[4700]: I0227 18:22:47.289870 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tn4s8"] Feb 27 18:22:47 crc kubenswrapper[4700]: I0227 18:22:47.326585 4700 scope.go:117] "RemoveContainer" containerID="1997a0abf345d10034cfa408684ada4628e4101c046391ef96dc8d365163248c" Feb 27 18:22:47 crc kubenswrapper[4700]: E0227 18:22:47.327049 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1997a0abf345d10034cfa408684ada4628e4101c046391ef96dc8d365163248c\": container with ID starting with 1997a0abf345d10034cfa408684ada4628e4101c046391ef96dc8d365163248c not found: ID does not exist" containerID="1997a0abf345d10034cfa408684ada4628e4101c046391ef96dc8d365163248c" Feb 27 18:22:47 crc kubenswrapper[4700]: I0227 18:22:47.327090 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1997a0abf345d10034cfa408684ada4628e4101c046391ef96dc8d365163248c"} err="failed to get container status \"1997a0abf345d10034cfa408684ada4628e4101c046391ef96dc8d365163248c\": rpc error: code = NotFound desc = could not find container \"1997a0abf345d10034cfa408684ada4628e4101c046391ef96dc8d365163248c\": container with ID starting with 1997a0abf345d10034cfa408684ada4628e4101c046391ef96dc8d365163248c not found: ID does not exist" Feb 27 18:22:47 crc kubenswrapper[4700]: I0227 18:22:47.327114 4700 scope.go:117] "RemoveContainer" containerID="424d19170d51b0ec2a2e662c79027264e27a58bed662fc690382cf464322fc63" Feb 27 18:22:47 crc kubenswrapper[4700]: E0227 18:22:47.327343 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"424d19170d51b0ec2a2e662c79027264e27a58bed662fc690382cf464322fc63\": container with ID starting with 424d19170d51b0ec2a2e662c79027264e27a58bed662fc690382cf464322fc63 not found: ID does not exist" containerID="424d19170d51b0ec2a2e662c79027264e27a58bed662fc690382cf464322fc63" Feb 27 18:22:47 crc kubenswrapper[4700]: I0227 18:22:47.327371 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"424d19170d51b0ec2a2e662c79027264e27a58bed662fc690382cf464322fc63"} err="failed to get container status \"424d19170d51b0ec2a2e662c79027264e27a58bed662fc690382cf464322fc63\": rpc error: code = NotFound desc = could not find container \"424d19170d51b0ec2a2e662c79027264e27a58bed662fc690382cf464322fc63\": container with ID starting with 424d19170d51b0ec2a2e662c79027264e27a58bed662fc690382cf464322fc63 not found: ID does not exist" Feb 27 18:22:47 crc kubenswrapper[4700]: I0227 18:22:47.327391 4700 scope.go:117] "RemoveContainer" containerID="3b8324dc261d79c510a5a6fe4498764a5ccfedc2a55e39d88d50dce728b771f4" Feb 27 18:22:47 crc kubenswrapper[4700]: E0227 18:22:47.327742 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b8324dc261d79c510a5a6fe4498764a5ccfedc2a55e39d88d50dce728b771f4\": container with ID starting with 3b8324dc261d79c510a5a6fe4498764a5ccfedc2a55e39d88d50dce728b771f4 not found: ID does not exist" containerID="3b8324dc261d79c510a5a6fe4498764a5ccfedc2a55e39d88d50dce728b771f4" Feb 27 18:22:47 crc kubenswrapper[4700]: I0227 18:22:47.327790 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b8324dc261d79c510a5a6fe4498764a5ccfedc2a55e39d88d50dce728b771f4"} err="failed to get container status \"3b8324dc261d79c510a5a6fe4498764a5ccfedc2a55e39d88d50dce728b771f4\": rpc error: code = NotFound desc = could not find container \"3b8324dc261d79c510a5a6fe4498764a5ccfedc2a55e39d88d50dce728b771f4\": container with ID starting with 3b8324dc261d79c510a5a6fe4498764a5ccfedc2a55e39d88d50dce728b771f4 not found: ID does not exist" Feb 27 18:22:47 crc kubenswrapper[4700]: I0227 18:22:47.449165 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba9a02be-b313-47fe-8cc7-c68c8c310a37-utilities\") pod \"certified-operators-tn4s8\" (UID: \"ba9a02be-b313-47fe-8cc7-c68c8c310a37\") " pod="openshift-marketplace/certified-operators-tn4s8" Feb 27 18:22:47 crc kubenswrapper[4700]: I0227 18:22:47.449234 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba9a02be-b313-47fe-8cc7-c68c8c310a37-catalog-content\") pod \"certified-operators-tn4s8\" (UID: \"ba9a02be-b313-47fe-8cc7-c68c8c310a37\") " pod="openshift-marketplace/certified-operators-tn4s8" Feb 27 18:22:47 crc kubenswrapper[4700]: I0227 18:22:47.449433 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wskt\" (UniqueName: \"kubernetes.io/projected/ba9a02be-b313-47fe-8cc7-c68c8c310a37-kube-api-access-4wskt\") pod \"certified-operators-tn4s8\" (UID: \"ba9a02be-b313-47fe-8cc7-c68c8c310a37\") " pod="openshift-marketplace/certified-operators-tn4s8" Feb 27 18:22:47 crc kubenswrapper[4700]: I0227 18:22:47.551340 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wskt\" (UniqueName: \"kubernetes.io/projected/ba9a02be-b313-47fe-8cc7-c68c8c310a37-kube-api-access-4wskt\") pod \"certified-operators-tn4s8\" (UID: \"ba9a02be-b313-47fe-8cc7-c68c8c310a37\") " pod="openshift-marketplace/certified-operators-tn4s8" Feb 27 18:22:47 crc kubenswrapper[4700]: I0227 18:22:47.551856 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba9a02be-b313-47fe-8cc7-c68c8c310a37-utilities\") pod \"certified-operators-tn4s8\" (UID: \"ba9a02be-b313-47fe-8cc7-c68c8c310a37\") " pod="openshift-marketplace/certified-operators-tn4s8" Feb 27 18:22:47 crc kubenswrapper[4700]: I0227 18:22:47.551987 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba9a02be-b313-47fe-8cc7-c68c8c310a37-catalog-content\") pod \"certified-operators-tn4s8\" (UID: \"ba9a02be-b313-47fe-8cc7-c68c8c310a37\") " pod="openshift-marketplace/certified-operators-tn4s8" Feb 27 18:22:47 crc kubenswrapper[4700]: I0227 18:22:47.552543 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba9a02be-b313-47fe-8cc7-c68c8c310a37-utilities\") pod \"certified-operators-tn4s8\" (UID: \"ba9a02be-b313-47fe-8cc7-c68c8c310a37\") " pod="openshift-marketplace/certified-operators-tn4s8" Feb 27 18:22:47 crc kubenswrapper[4700]: I0227 18:22:47.552684 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba9a02be-b313-47fe-8cc7-c68c8c310a37-catalog-content\") pod \"certified-operators-tn4s8\" (UID: \"ba9a02be-b313-47fe-8cc7-c68c8c310a37\") " pod="openshift-marketplace/certified-operators-tn4s8" Feb 27 18:22:47 crc kubenswrapper[4700]: I0227 18:22:47.571218 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wskt\" (UniqueName: \"kubernetes.io/projected/ba9a02be-b313-47fe-8cc7-c68c8c310a37-kube-api-access-4wskt\") pod \"certified-operators-tn4s8\" (UID: \"ba9a02be-b313-47fe-8cc7-c68c8c310a37\") " pod="openshift-marketplace/certified-operators-tn4s8" Feb 27 18:22:47 crc kubenswrapper[4700]: I0227 18:22:47.655112 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tn4s8" Feb 27 18:22:48 crc kubenswrapper[4700]: I0227 18:22:48.312562 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tn4s8"] Feb 27 18:22:48 crc kubenswrapper[4700]: E0227 18:22:48.986169 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:22:48 crc kubenswrapper[4700]: I0227 18:22:48.993611 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="243afaab-17c1-4631-974c-f7b693236b08" path="/var/lib/kubelet/pods/243afaab-17c1-4631-974c-f7b693236b08/volumes" Feb 27 18:22:49 crc kubenswrapper[4700]: I0227 18:22:49.190752 4700 generic.go:334] "Generic (PLEG): container finished" podID="ba9a02be-b313-47fe-8cc7-c68c8c310a37" containerID="7d9114be79b5a139774ea0b12a26f95eaf3e060da02e001e785c1646d37043cf" exitCode=0 Feb 27 18:22:49 crc kubenswrapper[4700]: I0227 18:22:49.190823 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tn4s8" event={"ID":"ba9a02be-b313-47fe-8cc7-c68c8c310a37","Type":"ContainerDied","Data":"7d9114be79b5a139774ea0b12a26f95eaf3e060da02e001e785c1646d37043cf"} Feb 27 18:22:49 crc kubenswrapper[4700]: I0227 18:22:49.190862 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tn4s8" event={"ID":"ba9a02be-b313-47fe-8cc7-c68c8c310a37","Type":"ContainerStarted","Data":"0a3299d5137d29c793fc014aefd0a0d498ad71f4333031f477c0e2c994a34c2c"} Feb 27 18:22:49 crc kubenswrapper[4700]: E0227 18:22:49.906610 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 27 18:22:49 crc kubenswrapper[4700]: E0227 18:22:49.908309 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4wskt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-tn4s8_openshift-marketplace(ba9a02be-b313-47fe-8cc7-c68c8c310a37): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 18:22:49 crc kubenswrapper[4700]: E0227 18:22:49.909768 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/certified-operators-tn4s8" podUID="ba9a02be-b313-47fe-8cc7-c68c8c310a37" Feb 27 18:22:50 crc kubenswrapper[4700]: E0227 18:22:50.206723 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-tn4s8" podUID="ba9a02be-b313-47fe-8cc7-c68c8c310a37" Feb 27 18:22:53 crc kubenswrapper[4700]: E0227 18:22:53.995493 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:22:55 crc kubenswrapper[4700]: E0227 18:22:55.988866 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:22:59 crc kubenswrapper[4700]: E0227 18:22:59.984162 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:23:04 crc kubenswrapper[4700]: E0227 18:23:04.590036 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 27 18:23:04 crc kubenswrapper[4700]: E0227 18:23:04.590841 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4wskt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-tn4s8_openshift-marketplace(ba9a02be-b313-47fe-8cc7-c68c8c310a37): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 18:23:04 crc kubenswrapper[4700]: E0227 18:23:04.592879 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/certified-operators-tn4s8" podUID="ba9a02be-b313-47fe-8cc7-c68c8c310a37" Feb 27 18:23:06 crc kubenswrapper[4700]: E0227 18:23:06.984599 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:23:09 crc kubenswrapper[4700]: E0227 18:23:09.987440 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:23:11 crc kubenswrapper[4700]: E0227 18:23:11.984900 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:23:17 crc kubenswrapper[4700]: E0227 18:23:17.984703 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:23:19 crc kubenswrapper[4700]: E0227 18:23:19.984926 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-tn4s8" podUID="ba9a02be-b313-47fe-8cc7-c68c8c310a37" Feb 27 18:23:23 crc kubenswrapper[4700]: E0227 18:23:23.985445 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:23:24 crc kubenswrapper[4700]: E0227 18:23:24.984247 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:23:32 crc kubenswrapper[4700]: E0227 18:23:32.985869 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:23:35 crc kubenswrapper[4700]: E0227 18:23:35.035452 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 27 18:23:35 crc kubenswrapper[4700]: E0227 18:23:35.036045 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4wskt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-tn4s8_openshift-marketplace(ba9a02be-b313-47fe-8cc7-c68c8c310a37): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 18:23:35 crc kubenswrapper[4700]: E0227 18:23:35.037206 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/certified-operator-index@sha256=625372062485d8ed1e4e84c388a7d036cb39c1b93d8c56dd3418fce0c028b62b/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/certified-operators-tn4s8" podUID="ba9a02be-b313-47fe-8cc7-c68c8c310a37" Feb 27 18:23:36 crc kubenswrapper[4700]: E0227 18:23:36.988929 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:23:38 crc kubenswrapper[4700]: E0227 18:23:38.983818 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:23:43 crc kubenswrapper[4700]: E0227 18:23:43.984955 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:23:45 crc kubenswrapper[4700]: E0227 18:23:45.984435 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-tn4s8" podUID="ba9a02be-b313-47fe-8cc7-c68c8c310a37" Feb 27 18:23:48 crc kubenswrapper[4700]: E0227 18:23:48.985160 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:23:50 crc kubenswrapper[4700]: E0227 18:23:50.994033 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:23:55 crc kubenswrapper[4700]: E0227 18:23:55.984367 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:23:58 crc kubenswrapper[4700]: E0227 18:23:58.985892 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-tn4s8" podUID="ba9a02be-b313-47fe-8cc7-c68c8c310a37" Feb 27 18:24:00 crc kubenswrapper[4700]: I0227 18:24:00.171551 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536944-clxj9"] Feb 27 18:24:00 crc kubenswrapper[4700]: I0227 18:24:00.175970 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536944-clxj9" Feb 27 18:24:00 crc kubenswrapper[4700]: I0227 18:24:00.197430 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536944-clxj9"] Feb 27 18:24:00 crc kubenswrapper[4700]: I0227 18:24:00.344603 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbdvs\" (UniqueName: \"kubernetes.io/projected/9f757b70-0256-4500-8d34-368ff32dc545-kube-api-access-sbdvs\") pod \"auto-csr-approver-29536944-clxj9\" (UID: \"9f757b70-0256-4500-8d34-368ff32dc545\") " pod="openshift-infra/auto-csr-approver-29536944-clxj9" Feb 27 18:24:00 crc kubenswrapper[4700]: I0227 18:24:00.447346 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbdvs\" (UniqueName: \"kubernetes.io/projected/9f757b70-0256-4500-8d34-368ff32dc545-kube-api-access-sbdvs\") pod \"auto-csr-approver-29536944-clxj9\" (UID: \"9f757b70-0256-4500-8d34-368ff32dc545\") " pod="openshift-infra/auto-csr-approver-29536944-clxj9" Feb 27 18:24:00 crc kubenswrapper[4700]: I0227 18:24:00.483651 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbdvs\" (UniqueName: \"kubernetes.io/projected/9f757b70-0256-4500-8d34-368ff32dc545-kube-api-access-sbdvs\") pod \"auto-csr-approver-29536944-clxj9\" (UID: \"9f757b70-0256-4500-8d34-368ff32dc545\") " pod="openshift-infra/auto-csr-approver-29536944-clxj9" Feb 27 18:24:00 crc kubenswrapper[4700]: I0227 18:24:00.502153 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536944-clxj9" Feb 27 18:24:01 crc kubenswrapper[4700]: E0227 18:24:01.049212 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:24:01 crc kubenswrapper[4700]: W0227 18:24:01.052811 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f757b70_0256_4500_8d34_368ff32dc545.slice/crio-c4ea7dc2f18e376d4cd44760a7ea8a0d5794352db53c23581809cb405af98893 WatchSource:0}: Error finding container c4ea7dc2f18e376d4cd44760a7ea8a0d5794352db53c23581809cb405af98893: Status 404 returned error can't find the container with id c4ea7dc2f18e376d4cd44760a7ea8a0d5794352db53c23581809cb405af98893 Feb 27 18:24:01 crc kubenswrapper[4700]: I0227 18:24:01.057677 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536944-clxj9"] Feb 27 18:24:01 crc kubenswrapper[4700]: I0227 18:24:01.124766 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536944-clxj9" event={"ID":"9f757b70-0256-4500-8d34-368ff32dc545","Type":"ContainerStarted","Data":"c4ea7dc2f18e376d4cd44760a7ea8a0d5794352db53c23581809cb405af98893"} Feb 27 18:24:03 crc kubenswrapper[4700]: I0227 18:24:03.156341 4700 generic.go:334] "Generic (PLEG): container finished" podID="9f757b70-0256-4500-8d34-368ff32dc545" containerID="472b00220db6e77f0fc32d292654f8847eab10cc1d098ab50e4aaa0293fb1147" exitCode=0 Feb 27 18:24:03 crc kubenswrapper[4700]: I0227 18:24:03.156419 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536944-clxj9" event={"ID":"9f757b70-0256-4500-8d34-368ff32dc545","Type":"ContainerDied","Data":"472b00220db6e77f0fc32d292654f8847eab10cc1d098ab50e4aaa0293fb1147"} Feb 27 18:24:03 crc kubenswrapper[4700]: E0227 18:24:03.985038 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:24:04 crc kubenswrapper[4700]: I0227 18:24:04.780974 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536944-clxj9" Feb 27 18:24:04 crc kubenswrapper[4700]: I0227 18:24:04.968312 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbdvs\" (UniqueName: \"kubernetes.io/projected/9f757b70-0256-4500-8d34-368ff32dc545-kube-api-access-sbdvs\") pod \"9f757b70-0256-4500-8d34-368ff32dc545\" (UID: \"9f757b70-0256-4500-8d34-368ff32dc545\") " Feb 27 18:24:04 crc kubenswrapper[4700]: I0227 18:24:04.987707 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f757b70-0256-4500-8d34-368ff32dc545-kube-api-access-sbdvs" (OuterVolumeSpecName: "kube-api-access-sbdvs") pod "9f757b70-0256-4500-8d34-368ff32dc545" (UID: "9f757b70-0256-4500-8d34-368ff32dc545"). InnerVolumeSpecName "kube-api-access-sbdvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:24:05 crc kubenswrapper[4700]: I0227 18:24:05.071789 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbdvs\" (UniqueName: \"kubernetes.io/projected/9f757b70-0256-4500-8d34-368ff32dc545-kube-api-access-sbdvs\") on node \"crc\" DevicePath \"\"" Feb 27 18:24:05 crc kubenswrapper[4700]: I0227 18:24:05.180450 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536944-clxj9" event={"ID":"9f757b70-0256-4500-8d34-368ff32dc545","Type":"ContainerDied","Data":"c4ea7dc2f18e376d4cd44760a7ea8a0d5794352db53c23581809cb405af98893"} Feb 27 18:24:05 crc kubenswrapper[4700]: I0227 18:24:05.180532 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536944-clxj9" Feb 27 18:24:05 crc kubenswrapper[4700]: I0227 18:24:05.180552 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c4ea7dc2f18e376d4cd44760a7ea8a0d5794352db53c23581809cb405af98893" Feb 27 18:24:05 crc kubenswrapper[4700]: I0227 18:24:05.919957 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536938-bppsz"] Feb 27 18:24:05 crc kubenswrapper[4700]: I0227 18:24:05.939710 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536938-bppsz"] Feb 27 18:24:07 crc kubenswrapper[4700]: I0227 18:24:07.001936 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90f1566e-7eeb-4455-9bcd-a9eed39c5eb1" path="/var/lib/kubelet/pods/90f1566e-7eeb-4455-9bcd-a9eed39c5eb1/volumes" Feb 27 18:24:08 crc kubenswrapper[4700]: E0227 18:24:08.983918 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:24:13 crc kubenswrapper[4700]: E0227 18:24:13.984566 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-tn4s8" podUID="ba9a02be-b313-47fe-8cc7-c68c8c310a37" Feb 27 18:24:14 crc kubenswrapper[4700]: E0227 18:24:14.985448 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:24:17 crc kubenswrapper[4700]: E0227 18:24:17.984230 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:24:19 crc kubenswrapper[4700]: E0227 18:24:19.985210 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:24:29 crc kubenswrapper[4700]: E0227 18:24:29.983849 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:24:30 crc kubenswrapper[4700]: I0227 18:24:30.492742 4700 generic.go:334] "Generic (PLEG): container finished" podID="ba9a02be-b313-47fe-8cc7-c68c8c310a37" containerID="83eb4a92e23da2eaad1b67bf2c7613f5fa8002a629bdd3e797ee981b29730085" exitCode=0 Feb 27 18:24:30 crc kubenswrapper[4700]: I0227 18:24:30.492861 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tn4s8" event={"ID":"ba9a02be-b313-47fe-8cc7-c68c8c310a37","Type":"ContainerDied","Data":"83eb4a92e23da2eaad1b67bf2c7613f5fa8002a629bdd3e797ee981b29730085"} Feb 27 18:24:30 crc kubenswrapper[4700]: E0227 18:24:30.998696 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:24:31 crc kubenswrapper[4700]: I0227 18:24:31.513828 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tn4s8" event={"ID":"ba9a02be-b313-47fe-8cc7-c68c8c310a37","Type":"ContainerStarted","Data":"7575125667f7281fb61ee0c94b49c4cd55eb613f2e35cab65f5966ab1b6e3031"} Feb 27 18:24:31 crc kubenswrapper[4700]: I0227 18:24:31.540329 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tn4s8" podStartSLOduration=2.579611279 podStartE2EDuration="1m44.540297771s" podCreationTimestamp="2026-02-27 18:22:47 +0000 UTC" firstStartedPulling="2026-02-27 18:22:49.194180698 +0000 UTC m=+4929.179493445" lastFinishedPulling="2026-02-27 18:24:31.15486718 +0000 UTC m=+5031.140179937" observedRunningTime="2026-02-27 18:24:31.533204846 +0000 UTC m=+5031.518517593" watchObservedRunningTime="2026-02-27 18:24:31.540297771 +0000 UTC m=+5031.525610558" Feb 27 18:24:31 crc kubenswrapper[4700]: E0227 18:24:31.984713 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:24:37 crc kubenswrapper[4700]: I0227 18:24:37.656260 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tn4s8" Feb 27 18:24:37 crc kubenswrapper[4700]: I0227 18:24:37.656686 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tn4s8" Feb 27 18:24:37 crc kubenswrapper[4700]: I0227 18:24:37.738432 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tn4s8" Feb 27 18:24:39 crc kubenswrapper[4700]: I0227 18:24:39.120665 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tn4s8" Feb 27 18:24:39 crc kubenswrapper[4700]: I0227 18:24:39.197314 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tn4s8"] Feb 27 18:24:40 crc kubenswrapper[4700]: I0227 18:24:40.660209 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tn4s8" podUID="ba9a02be-b313-47fe-8cc7-c68c8c310a37" containerName="registry-server" containerID="cri-o://7575125667f7281fb61ee0c94b49c4cd55eb613f2e35cab65f5966ab1b6e3031" gracePeriod=2 Feb 27 18:24:41 crc kubenswrapper[4700]: E0227 18:24:41.005670 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:24:41 crc kubenswrapper[4700]: I0227 18:24:41.711179 4700 generic.go:334] "Generic (PLEG): container finished" podID="ba9a02be-b313-47fe-8cc7-c68c8c310a37" containerID="7575125667f7281fb61ee0c94b49c4cd55eb613f2e35cab65f5966ab1b6e3031" exitCode=0 Feb 27 18:24:41 crc kubenswrapper[4700]: I0227 18:24:41.711618 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tn4s8" event={"ID":"ba9a02be-b313-47fe-8cc7-c68c8c310a37","Type":"ContainerDied","Data":"7575125667f7281fb61ee0c94b49c4cd55eb613f2e35cab65f5966ab1b6e3031"} Feb 27 18:24:41 crc kubenswrapper[4700]: I0227 18:24:41.838674 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tn4s8" Feb 27 18:24:41 crc kubenswrapper[4700]: I0227 18:24:41.936413 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba9a02be-b313-47fe-8cc7-c68c8c310a37-utilities\") pod \"ba9a02be-b313-47fe-8cc7-c68c8c310a37\" (UID: \"ba9a02be-b313-47fe-8cc7-c68c8c310a37\") " Feb 27 18:24:41 crc kubenswrapper[4700]: I0227 18:24:41.936646 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba9a02be-b313-47fe-8cc7-c68c8c310a37-catalog-content\") pod \"ba9a02be-b313-47fe-8cc7-c68c8c310a37\" (UID: \"ba9a02be-b313-47fe-8cc7-c68c8c310a37\") " Feb 27 18:24:41 crc kubenswrapper[4700]: I0227 18:24:41.936762 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wskt\" (UniqueName: \"kubernetes.io/projected/ba9a02be-b313-47fe-8cc7-c68c8c310a37-kube-api-access-4wskt\") pod \"ba9a02be-b313-47fe-8cc7-c68c8c310a37\" (UID: \"ba9a02be-b313-47fe-8cc7-c68c8c310a37\") " Feb 27 18:24:41 crc kubenswrapper[4700]: I0227 18:24:41.939220 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba9a02be-b313-47fe-8cc7-c68c8c310a37-utilities" (OuterVolumeSpecName: "utilities") pod "ba9a02be-b313-47fe-8cc7-c68c8c310a37" (UID: "ba9a02be-b313-47fe-8cc7-c68c8c310a37"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 18:24:41 crc kubenswrapper[4700]: I0227 18:24:41.956318 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba9a02be-b313-47fe-8cc7-c68c8c310a37-kube-api-access-4wskt" (OuterVolumeSpecName: "kube-api-access-4wskt") pod "ba9a02be-b313-47fe-8cc7-c68c8c310a37" (UID: "ba9a02be-b313-47fe-8cc7-c68c8c310a37"). InnerVolumeSpecName "kube-api-access-4wskt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:24:42 crc kubenswrapper[4700]: I0227 18:24:42.018670 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba9a02be-b313-47fe-8cc7-c68c8c310a37-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ba9a02be-b313-47fe-8cc7-c68c8c310a37" (UID: "ba9a02be-b313-47fe-8cc7-c68c8c310a37"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 18:24:42 crc kubenswrapper[4700]: I0227 18:24:42.040884 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba9a02be-b313-47fe-8cc7-c68c8c310a37-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 18:24:42 crc kubenswrapper[4700]: I0227 18:24:42.040925 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba9a02be-b313-47fe-8cc7-c68c8c310a37-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 18:24:42 crc kubenswrapper[4700]: I0227 18:24:42.040941 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wskt\" (UniqueName: \"kubernetes.io/projected/ba9a02be-b313-47fe-8cc7-c68c8c310a37-kube-api-access-4wskt\") on node \"crc\" DevicePath \"\"" Feb 27 18:24:42 crc kubenswrapper[4700]: I0227 18:24:42.728309 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tn4s8" event={"ID":"ba9a02be-b313-47fe-8cc7-c68c8c310a37","Type":"ContainerDied","Data":"0a3299d5137d29c793fc014aefd0a0d498ad71f4333031f477c0e2c994a34c2c"} Feb 27 18:24:42 crc kubenswrapper[4700]: I0227 18:24:42.728435 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tn4s8" Feb 27 18:24:42 crc kubenswrapper[4700]: I0227 18:24:42.728813 4700 scope.go:117] "RemoveContainer" containerID="7575125667f7281fb61ee0c94b49c4cd55eb613f2e35cab65f5966ab1b6e3031" Feb 27 18:24:42 crc kubenswrapper[4700]: I0227 18:24:42.766439 4700 scope.go:117] "RemoveContainer" containerID="83eb4a92e23da2eaad1b67bf2c7613f5fa8002a629bdd3e797ee981b29730085" Feb 27 18:24:42 crc kubenswrapper[4700]: I0227 18:24:42.790318 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tn4s8"] Feb 27 18:24:42 crc kubenswrapper[4700]: I0227 18:24:42.810251 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tn4s8"] Feb 27 18:24:42 crc kubenswrapper[4700]: I0227 18:24:42.824173 4700 scope.go:117] "RemoveContainer" containerID="7d9114be79b5a139774ea0b12a26f95eaf3e060da02e001e785c1646d37043cf" Feb 27 18:24:43 crc kubenswrapper[4700]: I0227 18:24:43.000909 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba9a02be-b313-47fe-8cc7-c68c8c310a37" path="/var/lib/kubelet/pods/ba9a02be-b313-47fe-8cc7-c68c8c310a37/volumes" Feb 27 18:24:43 crc kubenswrapper[4700]: E0227 18:24:43.985543 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:24:43 crc kubenswrapper[4700]: E0227 18:24:43.985629 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:24:54 crc kubenswrapper[4700]: E0227 18:24:54.986504 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:24:55 crc kubenswrapper[4700]: E0227 18:24:55.985281 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:24:59 crc kubenswrapper[4700]: E0227 18:24:59.919358 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 18:24:59 crc kubenswrapper[4700]: E0227 18:24:59.920295 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 18:24:59 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 18:24:59 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qkjfz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536934-rw77f_openshift-infra(caa4e57d-b02d-4d65-944d-451b8294f884): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 18:24:59 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 18:24:59 crc kubenswrapper[4700]: E0227 18:24:59.922402 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:25:05 crc kubenswrapper[4700]: E0227 18:25:05.988001 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:25:06 crc kubenswrapper[4700]: I0227 18:25:06.410280 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 18:25:06 crc kubenswrapper[4700]: I0227 18:25:06.410354 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 18:25:07 crc kubenswrapper[4700]: E0227 18:25:07.983979 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:25:13 crc kubenswrapper[4700]: I0227 18:25:13.460809 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-p5l2d"] Feb 27 18:25:13 crc kubenswrapper[4700]: E0227 18:25:13.464240 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f757b70-0256-4500-8d34-368ff32dc545" containerName="oc" Feb 27 18:25:13 crc kubenswrapper[4700]: I0227 18:25:13.464598 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f757b70-0256-4500-8d34-368ff32dc545" containerName="oc" Feb 27 18:25:13 crc kubenswrapper[4700]: E0227 18:25:13.464824 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba9a02be-b313-47fe-8cc7-c68c8c310a37" containerName="registry-server" Feb 27 18:25:13 crc kubenswrapper[4700]: I0227 18:25:13.464967 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba9a02be-b313-47fe-8cc7-c68c8c310a37" containerName="registry-server" Feb 27 18:25:13 crc kubenswrapper[4700]: E0227 18:25:13.465112 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba9a02be-b313-47fe-8cc7-c68c8c310a37" containerName="extract-utilities" Feb 27 18:25:13 crc kubenswrapper[4700]: I0227 18:25:13.465249 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba9a02be-b313-47fe-8cc7-c68c8c310a37" containerName="extract-utilities" Feb 27 18:25:13 crc kubenswrapper[4700]: E0227 18:25:13.465380 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba9a02be-b313-47fe-8cc7-c68c8c310a37" containerName="extract-content" Feb 27 18:25:13 crc kubenswrapper[4700]: I0227 18:25:13.465551 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba9a02be-b313-47fe-8cc7-c68c8c310a37" containerName="extract-content" Feb 27 18:25:13 crc kubenswrapper[4700]: I0227 18:25:13.468604 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba9a02be-b313-47fe-8cc7-c68c8c310a37" containerName="registry-server" Feb 27 18:25:13 crc kubenswrapper[4700]: I0227 18:25:13.468783 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f757b70-0256-4500-8d34-368ff32dc545" containerName="oc" Feb 27 18:25:13 crc kubenswrapper[4700]: I0227 18:25:13.472874 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p5l2d" Feb 27 18:25:13 crc kubenswrapper[4700]: I0227 18:25:13.490405 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p5l2d"] Feb 27 18:25:13 crc kubenswrapper[4700]: I0227 18:25:13.668186 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5stmq\" (UniqueName: \"kubernetes.io/projected/f3de1a57-09ac-4688-b26b-08846e57a0f4-kube-api-access-5stmq\") pod \"redhat-marketplace-p5l2d\" (UID: \"f3de1a57-09ac-4688-b26b-08846e57a0f4\") " pod="openshift-marketplace/redhat-marketplace-p5l2d" Feb 27 18:25:13 crc kubenswrapper[4700]: I0227 18:25:13.668250 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3de1a57-09ac-4688-b26b-08846e57a0f4-catalog-content\") pod \"redhat-marketplace-p5l2d\" (UID: \"f3de1a57-09ac-4688-b26b-08846e57a0f4\") " pod="openshift-marketplace/redhat-marketplace-p5l2d" Feb 27 18:25:13 crc kubenswrapper[4700]: I0227 18:25:13.668300 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3de1a57-09ac-4688-b26b-08846e57a0f4-utilities\") pod \"redhat-marketplace-p5l2d\" (UID: \"f3de1a57-09ac-4688-b26b-08846e57a0f4\") " pod="openshift-marketplace/redhat-marketplace-p5l2d" Feb 27 18:25:13 crc kubenswrapper[4700]: I0227 18:25:13.770555 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3de1a57-09ac-4688-b26b-08846e57a0f4-utilities\") pod \"redhat-marketplace-p5l2d\" (UID: \"f3de1a57-09ac-4688-b26b-08846e57a0f4\") " pod="openshift-marketplace/redhat-marketplace-p5l2d" Feb 27 18:25:13 crc kubenswrapper[4700]: I0227 18:25:13.771000 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5stmq\" (UniqueName: \"kubernetes.io/projected/f3de1a57-09ac-4688-b26b-08846e57a0f4-kube-api-access-5stmq\") pod \"redhat-marketplace-p5l2d\" (UID: \"f3de1a57-09ac-4688-b26b-08846e57a0f4\") " pod="openshift-marketplace/redhat-marketplace-p5l2d" Feb 27 18:25:13 crc kubenswrapper[4700]: I0227 18:25:13.771049 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3de1a57-09ac-4688-b26b-08846e57a0f4-catalog-content\") pod \"redhat-marketplace-p5l2d\" (UID: \"f3de1a57-09ac-4688-b26b-08846e57a0f4\") " pod="openshift-marketplace/redhat-marketplace-p5l2d" Feb 27 18:25:13 crc kubenswrapper[4700]: I0227 18:25:13.771157 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3de1a57-09ac-4688-b26b-08846e57a0f4-utilities\") pod \"redhat-marketplace-p5l2d\" (UID: \"f3de1a57-09ac-4688-b26b-08846e57a0f4\") " pod="openshift-marketplace/redhat-marketplace-p5l2d" Feb 27 18:25:13 crc kubenswrapper[4700]: I0227 18:25:13.771519 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3de1a57-09ac-4688-b26b-08846e57a0f4-catalog-content\") pod \"redhat-marketplace-p5l2d\" (UID: \"f3de1a57-09ac-4688-b26b-08846e57a0f4\") " pod="openshift-marketplace/redhat-marketplace-p5l2d" Feb 27 18:25:13 crc kubenswrapper[4700]: I0227 18:25:13.792268 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5stmq\" (UniqueName: \"kubernetes.io/projected/f3de1a57-09ac-4688-b26b-08846e57a0f4-kube-api-access-5stmq\") pod \"redhat-marketplace-p5l2d\" (UID: \"f3de1a57-09ac-4688-b26b-08846e57a0f4\") " pod="openshift-marketplace/redhat-marketplace-p5l2d" Feb 27 18:25:13 crc kubenswrapper[4700]: I0227 18:25:13.806337 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p5l2d" Feb 27 18:25:13 crc kubenswrapper[4700]: E0227 18:25:13.983025 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:25:14 crc kubenswrapper[4700]: I0227 18:25:14.312819 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p5l2d"] Feb 27 18:25:15 crc kubenswrapper[4700]: I0227 18:25:15.175882 4700 generic.go:334] "Generic (PLEG): container finished" podID="f3de1a57-09ac-4688-b26b-08846e57a0f4" containerID="010e1df5f1bfb26b948296dabd1fb50446e9beb26511d73c86fcca0c9b4fd5ad" exitCode=0 Feb 27 18:25:15 crc kubenswrapper[4700]: I0227 18:25:15.175932 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p5l2d" event={"ID":"f3de1a57-09ac-4688-b26b-08846e57a0f4","Type":"ContainerDied","Data":"010e1df5f1bfb26b948296dabd1fb50446e9beb26511d73c86fcca0c9b4fd5ad"} Feb 27 18:25:15 crc kubenswrapper[4700]: I0227 18:25:15.176222 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p5l2d" event={"ID":"f3de1a57-09ac-4688-b26b-08846e57a0f4","Type":"ContainerStarted","Data":"59206bdb2a5c8924d41672246ab19431a311b91c334731ee1452cdb79168c5bf"} Feb 27 18:25:15 crc kubenswrapper[4700]: E0227 18:25:15.856854 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 18:25:15 crc kubenswrapper[4700]: E0227 18:25:15.857315 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5stmq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-p5l2d_openshift-marketplace(f3de1a57-09ac-4688-b26b-08846e57a0f4): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 18:25:15 crc kubenswrapper[4700]: E0227 18:25:15.858597 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/redhat-marketplace-index@sha256=e848a00af7690cfa41500b98e0e7a0b9738ce0af7b6b4fee3ea20e0838523c30/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/redhat-marketplace-p5l2d" podUID="f3de1a57-09ac-4688-b26b-08846e57a0f4" Feb 27 18:25:16 crc kubenswrapper[4700]: E0227 18:25:16.190123 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-p5l2d" podUID="f3de1a57-09ac-4688-b26b-08846e57a0f4" Feb 27 18:25:20 crc kubenswrapper[4700]: E0227 18:25:20.997681 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:25:21 crc kubenswrapper[4700]: E0227 18:25:21.989267 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:25:28 crc kubenswrapper[4700]: E0227 18:25:28.985258 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:25:29 crc kubenswrapper[4700]: E0227 18:25:29.318578 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing image from source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: unexpected end of JSON input" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 18:25:29 crc kubenswrapper[4700]: E0227 18:25:29.319089 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5stmq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-p5l2d_openshift-marketplace(f3de1a57-09ac-4688-b26b-08846e57a0f4): ErrImagePull: initializing image from source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: unexpected end of JSON input" logger="UnhandledError" Feb 27 18:25:29 crc kubenswrapper[4700]: E0227 18:25:29.320399 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing image from source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: unexpected end of JSON input\"" pod="openshift-marketplace/redhat-marketplace-p5l2d" podUID="f3de1a57-09ac-4688-b26b-08846e57a0f4" Feb 27 18:25:32 crc kubenswrapper[4700]: E0227 18:25:32.988797 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:25:33 crc kubenswrapper[4700]: E0227 18:25:33.983591 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:25:36 crc kubenswrapper[4700]: I0227 18:25:36.410869 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 18:25:36 crc kubenswrapper[4700]: I0227 18:25:36.411232 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 18:25:42 crc kubenswrapper[4700]: E0227 18:25:42.060187 4700 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.222:35182->38.102.83.222:41369: write tcp 38.102.83.222:35182->38.102.83.222:41369: write: broken pipe Feb 27 18:25:42 crc kubenswrapper[4700]: E0227 18:25:42.986072 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-p5l2d" podUID="f3de1a57-09ac-4688-b26b-08846e57a0f4" Feb 27 18:25:44 crc kubenswrapper[4700]: E0227 18:25:44.711138 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:25:47 crc kubenswrapper[4700]: E0227 18:25:47.984787 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:25:48 crc kubenswrapper[4700]: E0227 18:25:48.982258 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:25:57 crc kubenswrapper[4700]: I0227 18:25:57.905456 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p5l2d" event={"ID":"f3de1a57-09ac-4688-b26b-08846e57a0f4","Type":"ContainerStarted","Data":"a2f86837814c4e87f4616f34bad9abca63539723e13545d403fbb6e2dc1d3ab4"} Feb 27 18:25:58 crc kubenswrapper[4700]: I0227 18:25:58.919795 4700 generic.go:334] "Generic (PLEG): container finished" podID="f3de1a57-09ac-4688-b26b-08846e57a0f4" containerID="a2f86837814c4e87f4616f34bad9abca63539723e13545d403fbb6e2dc1d3ab4" exitCode=0 Feb 27 18:25:58 crc kubenswrapper[4700]: I0227 18:25:58.919847 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p5l2d" event={"ID":"f3de1a57-09ac-4688-b26b-08846e57a0f4","Type":"ContainerDied","Data":"a2f86837814c4e87f4616f34bad9abca63539723e13545d403fbb6e2dc1d3ab4"} Feb 27 18:25:58 crc kubenswrapper[4700]: E0227 18:25:58.984737 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:25:59 crc kubenswrapper[4700]: I0227 18:25:59.935107 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p5l2d" event={"ID":"f3de1a57-09ac-4688-b26b-08846e57a0f4","Type":"ContainerStarted","Data":"b65cc67776622bcd04871bf989047ce6771e8c7ec828be933a8f5feb5a086d73"} Feb 27 18:25:59 crc kubenswrapper[4700]: I0227 18:25:59.962339 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-p5l2d" podStartSLOduration=2.510069079 podStartE2EDuration="46.962308408s" podCreationTimestamp="2026-02-27 18:25:13 +0000 UTC" firstStartedPulling="2026-02-27 18:25:15.178679806 +0000 UTC m=+5075.163992593" lastFinishedPulling="2026-02-27 18:25:59.630919145 +0000 UTC m=+5119.616231922" observedRunningTime="2026-02-27 18:25:59.953700313 +0000 UTC m=+5119.939013070" watchObservedRunningTime="2026-02-27 18:25:59.962308408 +0000 UTC m=+5119.947621195" Feb 27 18:26:00 crc kubenswrapper[4700]: I0227 18:26:00.157231 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536946-rbnt5"] Feb 27 18:26:00 crc kubenswrapper[4700]: I0227 18:26:00.159680 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536946-rbnt5" Feb 27 18:26:00 crc kubenswrapper[4700]: I0227 18:26:00.166020 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536946-rbnt5"] Feb 27 18:26:00 crc kubenswrapper[4700]: I0227 18:26:00.245235 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zd4lm\" (UniqueName: \"kubernetes.io/projected/c3c40c88-7b53-43dd-8357-17b25bdcc22d-kube-api-access-zd4lm\") pod \"auto-csr-approver-29536946-rbnt5\" (UID: \"c3c40c88-7b53-43dd-8357-17b25bdcc22d\") " pod="openshift-infra/auto-csr-approver-29536946-rbnt5" Feb 27 18:26:00 crc kubenswrapper[4700]: I0227 18:26:00.347533 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zd4lm\" (UniqueName: \"kubernetes.io/projected/c3c40c88-7b53-43dd-8357-17b25bdcc22d-kube-api-access-zd4lm\") pod \"auto-csr-approver-29536946-rbnt5\" (UID: \"c3c40c88-7b53-43dd-8357-17b25bdcc22d\") " pod="openshift-infra/auto-csr-approver-29536946-rbnt5" Feb 27 18:26:00 crc kubenswrapper[4700]: I0227 18:26:00.384975 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zd4lm\" (UniqueName: \"kubernetes.io/projected/c3c40c88-7b53-43dd-8357-17b25bdcc22d-kube-api-access-zd4lm\") pod \"auto-csr-approver-29536946-rbnt5\" (UID: \"c3c40c88-7b53-43dd-8357-17b25bdcc22d\") " pod="openshift-infra/auto-csr-approver-29536946-rbnt5" Feb 27 18:26:00 crc kubenswrapper[4700]: I0227 18:26:00.484374 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536946-rbnt5" Feb 27 18:26:01 crc kubenswrapper[4700]: I0227 18:26:01.040071 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536946-rbnt5"] Feb 27 18:26:01 crc kubenswrapper[4700]: E0227 18:26:01.984826 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:26:02 crc kubenswrapper[4700]: I0227 18:26:02.002360 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536946-rbnt5" event={"ID":"c3c40c88-7b53-43dd-8357-17b25bdcc22d","Type":"ContainerStarted","Data":"848cd3888523a8335a609ea2fc24e41430c1b5aee58fdb0d07080dc159724665"} Feb 27 18:26:02 crc kubenswrapper[4700]: E0227 18:26:02.229112 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 18:26:02 crc kubenswrapper[4700]: E0227 18:26:02.229322 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 18:26:02 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 18:26:02 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zd4lm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536946-rbnt5_openshift-infra(c3c40c88-7b53-43dd-8357-17b25bdcc22d): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 18:26:02 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 18:26:02 crc kubenswrapper[4700]: E0227 18:26:02.231379 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536946-rbnt5" podUID="c3c40c88-7b53-43dd-8357-17b25bdcc22d" Feb 27 18:26:02 crc kubenswrapper[4700]: E0227 18:26:02.984046 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:26:03 crc kubenswrapper[4700]: E0227 18:26:03.019021 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536946-rbnt5" podUID="c3c40c88-7b53-43dd-8357-17b25bdcc22d" Feb 27 18:26:03 crc kubenswrapper[4700]: I0227 18:26:03.806659 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-p5l2d" Feb 27 18:26:03 crc kubenswrapper[4700]: I0227 18:26:03.806943 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-p5l2d" Feb 27 18:26:03 crc kubenswrapper[4700]: I0227 18:26:03.885661 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-p5l2d" Feb 27 18:26:04 crc kubenswrapper[4700]: I0227 18:26:04.127556 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-p5l2d" Feb 27 18:26:04 crc kubenswrapper[4700]: I0227 18:26:04.207494 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p5l2d"] Feb 27 18:26:06 crc kubenswrapper[4700]: I0227 18:26:06.053630 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-p5l2d" podUID="f3de1a57-09ac-4688-b26b-08846e57a0f4" containerName="registry-server" containerID="cri-o://b65cc67776622bcd04871bf989047ce6771e8c7ec828be933a8f5feb5a086d73" gracePeriod=2 Feb 27 18:26:06 crc kubenswrapper[4700]: I0227 18:26:06.410320 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 18:26:06 crc kubenswrapper[4700]: I0227 18:26:06.410833 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 18:26:06 crc kubenswrapper[4700]: I0227 18:26:06.410876 4700 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" Feb 27 18:26:06 crc kubenswrapper[4700]: I0227 18:26:06.411725 4700 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b3fdfb06db2c06428e3b3e96ef9804be97292a50a8f04f891caec5e729731adf"} pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 18:26:06 crc kubenswrapper[4700]: I0227 18:26:06.411783 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" containerID="cri-o://b3fdfb06db2c06428e3b3e96ef9804be97292a50a8f04f891caec5e729731adf" gracePeriod=600 Feb 27 18:26:06 crc kubenswrapper[4700]: E0227 18:26:06.543065 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:26:06 crc kubenswrapper[4700]: I0227 18:26:06.588807 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p5l2d" Feb 27 18:26:06 crc kubenswrapper[4700]: I0227 18:26:06.729562 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3de1a57-09ac-4688-b26b-08846e57a0f4-utilities\") pod \"f3de1a57-09ac-4688-b26b-08846e57a0f4\" (UID: \"f3de1a57-09ac-4688-b26b-08846e57a0f4\") " Feb 27 18:26:06 crc kubenswrapper[4700]: I0227 18:26:06.729738 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3de1a57-09ac-4688-b26b-08846e57a0f4-catalog-content\") pod \"f3de1a57-09ac-4688-b26b-08846e57a0f4\" (UID: \"f3de1a57-09ac-4688-b26b-08846e57a0f4\") " Feb 27 18:26:06 crc kubenswrapper[4700]: I0227 18:26:06.729924 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5stmq\" (UniqueName: \"kubernetes.io/projected/f3de1a57-09ac-4688-b26b-08846e57a0f4-kube-api-access-5stmq\") pod \"f3de1a57-09ac-4688-b26b-08846e57a0f4\" (UID: \"f3de1a57-09ac-4688-b26b-08846e57a0f4\") " Feb 27 18:26:06 crc kubenswrapper[4700]: I0227 18:26:06.731115 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3de1a57-09ac-4688-b26b-08846e57a0f4-utilities" (OuterVolumeSpecName: "utilities") pod "f3de1a57-09ac-4688-b26b-08846e57a0f4" (UID: "f3de1a57-09ac-4688-b26b-08846e57a0f4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 18:26:06 crc kubenswrapper[4700]: I0227 18:26:06.738676 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3de1a57-09ac-4688-b26b-08846e57a0f4-kube-api-access-5stmq" (OuterVolumeSpecName: "kube-api-access-5stmq") pod "f3de1a57-09ac-4688-b26b-08846e57a0f4" (UID: "f3de1a57-09ac-4688-b26b-08846e57a0f4"). InnerVolumeSpecName "kube-api-access-5stmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:26:06 crc kubenswrapper[4700]: I0227 18:26:06.770075 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3de1a57-09ac-4688-b26b-08846e57a0f4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f3de1a57-09ac-4688-b26b-08846e57a0f4" (UID: "f3de1a57-09ac-4688-b26b-08846e57a0f4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 18:26:06 crc kubenswrapper[4700]: I0227 18:26:06.832376 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3de1a57-09ac-4688-b26b-08846e57a0f4-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 18:26:06 crc kubenswrapper[4700]: I0227 18:26:06.832433 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3de1a57-09ac-4688-b26b-08846e57a0f4-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 18:26:06 crc kubenswrapper[4700]: I0227 18:26:06.832449 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5stmq\" (UniqueName: \"kubernetes.io/projected/f3de1a57-09ac-4688-b26b-08846e57a0f4-kube-api-access-5stmq\") on node \"crc\" DevicePath \"\"" Feb 27 18:26:07 crc kubenswrapper[4700]: I0227 18:26:07.069841 4700 generic.go:334] "Generic (PLEG): container finished" podID="84b45511-b94c-479f-98db-fd2c4eceec46" containerID="b3fdfb06db2c06428e3b3e96ef9804be97292a50a8f04f891caec5e729731adf" exitCode=0 Feb 27 18:26:07 crc kubenswrapper[4700]: I0227 18:26:07.069996 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerDied","Data":"b3fdfb06db2c06428e3b3e96ef9804be97292a50a8f04f891caec5e729731adf"} Feb 27 18:26:07 crc kubenswrapper[4700]: I0227 18:26:07.070044 4700 scope.go:117] "RemoveContainer" containerID="d616594fff3f3a84ae3299a2a56f70ba23d3b8c11784f6e69c706e5bb75192a5" Feb 27 18:26:07 crc kubenswrapper[4700]: I0227 18:26:07.071128 4700 scope.go:117] "RemoveContainer" containerID="b3fdfb06db2c06428e3b3e96ef9804be97292a50a8f04f891caec5e729731adf" Feb 27 18:26:07 crc kubenswrapper[4700]: E0227 18:26:07.071609 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:26:07 crc kubenswrapper[4700]: I0227 18:26:07.075910 4700 generic.go:334] "Generic (PLEG): container finished" podID="f3de1a57-09ac-4688-b26b-08846e57a0f4" containerID="b65cc67776622bcd04871bf989047ce6771e8c7ec828be933a8f5feb5a086d73" exitCode=0 Feb 27 18:26:07 crc kubenswrapper[4700]: I0227 18:26:07.075954 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p5l2d" event={"ID":"f3de1a57-09ac-4688-b26b-08846e57a0f4","Type":"ContainerDied","Data":"b65cc67776622bcd04871bf989047ce6771e8c7ec828be933a8f5feb5a086d73"} Feb 27 18:26:07 crc kubenswrapper[4700]: I0227 18:26:07.075991 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p5l2d" event={"ID":"f3de1a57-09ac-4688-b26b-08846e57a0f4","Type":"ContainerDied","Data":"59206bdb2a5c8924d41672246ab19431a311b91c334731ee1452cdb79168c5bf"} Feb 27 18:26:07 crc kubenswrapper[4700]: I0227 18:26:07.076058 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p5l2d" Feb 27 18:26:07 crc kubenswrapper[4700]: I0227 18:26:07.105058 4700 scope.go:117] "RemoveContainer" containerID="b65cc67776622bcd04871bf989047ce6771e8c7ec828be933a8f5feb5a086d73" Feb 27 18:26:07 crc kubenswrapper[4700]: I0227 18:26:07.137076 4700 scope.go:117] "RemoveContainer" containerID="a2f86837814c4e87f4616f34bad9abca63539723e13545d403fbb6e2dc1d3ab4" Feb 27 18:26:07 crc kubenswrapper[4700]: I0227 18:26:07.140504 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p5l2d"] Feb 27 18:26:07 crc kubenswrapper[4700]: I0227 18:26:07.151765 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-p5l2d"] Feb 27 18:26:07 crc kubenswrapper[4700]: I0227 18:26:07.466421 4700 scope.go:117] "RemoveContainer" containerID="010e1df5f1bfb26b948296dabd1fb50446e9beb26511d73c86fcca0c9b4fd5ad" Feb 27 18:26:07 crc kubenswrapper[4700]: I0227 18:26:07.539788 4700 scope.go:117] "RemoveContainer" containerID="b65cc67776622bcd04871bf989047ce6771e8c7ec828be933a8f5feb5a086d73" Feb 27 18:26:07 crc kubenswrapper[4700]: E0227 18:26:07.540213 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b65cc67776622bcd04871bf989047ce6771e8c7ec828be933a8f5feb5a086d73\": container with ID starting with b65cc67776622bcd04871bf989047ce6771e8c7ec828be933a8f5feb5a086d73 not found: ID does not exist" containerID="b65cc67776622bcd04871bf989047ce6771e8c7ec828be933a8f5feb5a086d73" Feb 27 18:26:07 crc kubenswrapper[4700]: I0227 18:26:07.540272 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b65cc67776622bcd04871bf989047ce6771e8c7ec828be933a8f5feb5a086d73"} err="failed to get container status \"b65cc67776622bcd04871bf989047ce6771e8c7ec828be933a8f5feb5a086d73\": rpc error: code = NotFound desc = could not find container \"b65cc67776622bcd04871bf989047ce6771e8c7ec828be933a8f5feb5a086d73\": container with ID starting with b65cc67776622bcd04871bf989047ce6771e8c7ec828be933a8f5feb5a086d73 not found: ID does not exist" Feb 27 18:26:07 crc kubenswrapper[4700]: I0227 18:26:07.540300 4700 scope.go:117] "RemoveContainer" containerID="a2f86837814c4e87f4616f34bad9abca63539723e13545d403fbb6e2dc1d3ab4" Feb 27 18:26:07 crc kubenswrapper[4700]: E0227 18:26:07.540790 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2f86837814c4e87f4616f34bad9abca63539723e13545d403fbb6e2dc1d3ab4\": container with ID starting with a2f86837814c4e87f4616f34bad9abca63539723e13545d403fbb6e2dc1d3ab4 not found: ID does not exist" containerID="a2f86837814c4e87f4616f34bad9abca63539723e13545d403fbb6e2dc1d3ab4" Feb 27 18:26:07 crc kubenswrapper[4700]: I0227 18:26:07.540820 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2f86837814c4e87f4616f34bad9abca63539723e13545d403fbb6e2dc1d3ab4"} err="failed to get container status \"a2f86837814c4e87f4616f34bad9abca63539723e13545d403fbb6e2dc1d3ab4\": rpc error: code = NotFound desc = could not find container \"a2f86837814c4e87f4616f34bad9abca63539723e13545d403fbb6e2dc1d3ab4\": container with ID starting with a2f86837814c4e87f4616f34bad9abca63539723e13545d403fbb6e2dc1d3ab4 not found: ID does not exist" Feb 27 18:26:07 crc kubenswrapper[4700]: I0227 18:26:07.540840 4700 scope.go:117] "RemoveContainer" containerID="010e1df5f1bfb26b948296dabd1fb50446e9beb26511d73c86fcca0c9b4fd5ad" Feb 27 18:26:07 crc kubenswrapper[4700]: E0227 18:26:07.541107 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"010e1df5f1bfb26b948296dabd1fb50446e9beb26511d73c86fcca0c9b4fd5ad\": container with ID starting with 010e1df5f1bfb26b948296dabd1fb50446e9beb26511d73c86fcca0c9b4fd5ad not found: ID does not exist" containerID="010e1df5f1bfb26b948296dabd1fb50446e9beb26511d73c86fcca0c9b4fd5ad" Feb 27 18:26:07 crc kubenswrapper[4700]: I0227 18:26:07.541135 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"010e1df5f1bfb26b948296dabd1fb50446e9beb26511d73c86fcca0c9b4fd5ad"} err="failed to get container status \"010e1df5f1bfb26b948296dabd1fb50446e9beb26511d73c86fcca0c9b4fd5ad\": rpc error: code = NotFound desc = could not find container \"010e1df5f1bfb26b948296dabd1fb50446e9beb26511d73c86fcca0c9b4fd5ad\": container with ID starting with 010e1df5f1bfb26b948296dabd1fb50446e9beb26511d73c86fcca0c9b4fd5ad not found: ID does not exist" Feb 27 18:26:09 crc kubenswrapper[4700]: I0227 18:26:09.012822 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3de1a57-09ac-4688-b26b-08846e57a0f4" path="/var/lib/kubelet/pods/f3de1a57-09ac-4688-b26b-08846e57a0f4/volumes" Feb 27 18:26:09 crc kubenswrapper[4700]: I0227 18:26:09.750840 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="f2325e29-9f4c-438f-9400-e643e2494808" containerName="galera" probeResult="failure" output="command timed out" Feb 27 18:26:09 crc kubenswrapper[4700]: I0227 18:26:09.750893 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="f2325e29-9f4c-438f-9400-e643e2494808" containerName="galera" probeResult="failure" output="command timed out" Feb 27 18:26:12 crc kubenswrapper[4700]: E0227 18:26:12.985457 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:26:13 crc kubenswrapper[4700]: I0227 18:26:13.098047 4700 scope.go:117] "RemoveContainer" containerID="74145e784d29840559779fda9c00434464fd79d87683569916a30512b3a7987b" Feb 27 18:26:16 crc kubenswrapper[4700]: E0227 18:26:16.983343 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:26:16 crc kubenswrapper[4700]: E0227 18:26:16.988253 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:26:18 crc kubenswrapper[4700]: I0227 18:26:18.216387 4700 generic.go:334] "Generic (PLEG): container finished" podID="c3c40c88-7b53-43dd-8357-17b25bdcc22d" containerID="eb2f510fecd019a1d6c0a1d4ab0815196430e59068d0c0d5be165340ebef96e9" exitCode=0 Feb 27 18:26:18 crc kubenswrapper[4700]: I0227 18:26:18.216531 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536946-rbnt5" event={"ID":"c3c40c88-7b53-43dd-8357-17b25bdcc22d","Type":"ContainerDied","Data":"eb2f510fecd019a1d6c0a1d4ab0815196430e59068d0c0d5be165340ebef96e9"} Feb 27 18:26:19 crc kubenswrapper[4700]: I0227 18:26:19.737601 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536946-rbnt5" Feb 27 18:26:19 crc kubenswrapper[4700]: I0227 18:26:19.797629 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zd4lm\" (UniqueName: \"kubernetes.io/projected/c3c40c88-7b53-43dd-8357-17b25bdcc22d-kube-api-access-zd4lm\") pod \"c3c40c88-7b53-43dd-8357-17b25bdcc22d\" (UID: \"c3c40c88-7b53-43dd-8357-17b25bdcc22d\") " Feb 27 18:26:19 crc kubenswrapper[4700]: I0227 18:26:19.808965 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3c40c88-7b53-43dd-8357-17b25bdcc22d-kube-api-access-zd4lm" (OuterVolumeSpecName: "kube-api-access-zd4lm") pod "c3c40c88-7b53-43dd-8357-17b25bdcc22d" (UID: "c3c40c88-7b53-43dd-8357-17b25bdcc22d"). InnerVolumeSpecName "kube-api-access-zd4lm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:26:19 crc kubenswrapper[4700]: I0227 18:26:19.902517 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zd4lm\" (UniqueName: \"kubernetes.io/projected/c3c40c88-7b53-43dd-8357-17b25bdcc22d-kube-api-access-zd4lm\") on node \"crc\" DevicePath \"\"" Feb 27 18:26:20 crc kubenswrapper[4700]: I0227 18:26:20.247875 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536946-rbnt5" event={"ID":"c3c40c88-7b53-43dd-8357-17b25bdcc22d","Type":"ContainerDied","Data":"848cd3888523a8335a609ea2fc24e41430c1b5aee58fdb0d07080dc159724665"} Feb 27 18:26:20 crc kubenswrapper[4700]: I0227 18:26:20.248270 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="848cd3888523a8335a609ea2fc24e41430c1b5aee58fdb0d07080dc159724665" Feb 27 18:26:20 crc kubenswrapper[4700]: I0227 18:26:20.248004 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536946-rbnt5" Feb 27 18:26:20 crc kubenswrapper[4700]: I0227 18:26:20.851869 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536940-2ffgp"] Feb 27 18:26:20 crc kubenswrapper[4700]: I0227 18:26:20.869402 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536940-2ffgp"] Feb 27 18:26:21 crc kubenswrapper[4700]: I0227 18:26:21.000204 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdeb63e9-124e-40bb-a865-2463bba1aef2" path="/var/lib/kubelet/pods/fdeb63e9-124e-40bb-a865-2463bba1aef2/volumes" Feb 27 18:26:22 crc kubenswrapper[4700]: I0227 18:26:22.981856 4700 scope.go:117] "RemoveContainer" containerID="b3fdfb06db2c06428e3b3e96ef9804be97292a50a8f04f891caec5e729731adf" Feb 27 18:26:22 crc kubenswrapper[4700]: E0227 18:26:22.982772 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:26:26 crc kubenswrapper[4700]: E0227 18:26:26.985345 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:26:29 crc kubenswrapper[4700]: E0227 18:26:29.983485 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:26:29 crc kubenswrapper[4700]: E0227 18:26:29.985153 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:26:37 crc kubenswrapper[4700]: I0227 18:26:37.981149 4700 scope.go:117] "RemoveContainer" containerID="b3fdfb06db2c06428e3b3e96ef9804be97292a50a8f04f891caec5e729731adf" Feb 27 18:26:37 crc kubenswrapper[4700]: E0227 18:26:37.982825 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:26:41 crc kubenswrapper[4700]: E0227 18:26:41.008989 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:26:43 crc kubenswrapper[4700]: E0227 18:26:43.984046 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:26:44 crc kubenswrapper[4700]: E0227 18:26:44.998686 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:26:51 crc kubenswrapper[4700]: I0227 18:26:51.981884 4700 scope.go:117] "RemoveContainer" containerID="b3fdfb06db2c06428e3b3e96ef9804be97292a50a8f04f891caec5e729731adf" Feb 27 18:26:51 crc kubenswrapper[4700]: E0227 18:26:51.983923 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:26:51 crc kubenswrapper[4700]: E0227 18:26:51.984541 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:26:54 crc kubenswrapper[4700]: E0227 18:26:54.984444 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:26:58 crc kubenswrapper[4700]: E0227 18:26:58.984335 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:27:02 crc kubenswrapper[4700]: E0227 18:27:02.984710 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:27:05 crc kubenswrapper[4700]: I0227 18:27:05.982614 4700 scope.go:117] "RemoveContainer" containerID="b3fdfb06db2c06428e3b3e96ef9804be97292a50a8f04f891caec5e729731adf" Feb 27 18:27:05 crc kubenswrapper[4700]: E0227 18:27:05.983332 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:27:05 crc kubenswrapper[4700]: E0227 18:27:05.984707 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:27:12 crc kubenswrapper[4700]: E0227 18:27:12.986034 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:27:13 crc kubenswrapper[4700]: I0227 18:27:13.234012 4700 scope.go:117] "RemoveContainer" containerID="626b83a35710ba8583357907106bf3ee7bf7f041e6cfd87b21a5822ae18d3c2e" Feb 27 18:27:13 crc kubenswrapper[4700]: E0227 18:27:13.984994 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:27:16 crc kubenswrapper[4700]: I0227 18:27:16.981424 4700 scope.go:117] "RemoveContainer" containerID="b3fdfb06db2c06428e3b3e96ef9804be97292a50a8f04f891caec5e729731adf" Feb 27 18:27:16 crc kubenswrapper[4700]: E0227 18:27:16.982292 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:27:20 crc kubenswrapper[4700]: E0227 18:27:20.997730 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" Feb 27 18:27:24 crc kubenswrapper[4700]: E0227 18:27:24.982902 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:27:24 crc kubenswrapper[4700]: E0227 18:27:24.985218 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" Feb 27 18:27:28 crc kubenswrapper[4700]: I0227 18:27:28.981893 4700 scope.go:117] "RemoveContainer" containerID="b3fdfb06db2c06428e3b3e96ef9804be97292a50a8f04f891caec5e729731adf" Feb 27 18:27:28 crc kubenswrapper[4700]: E0227 18:27:28.982741 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:27:32 crc kubenswrapper[4700]: I0227 18:27:32.984846 4700 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 18:27:35 crc kubenswrapper[4700]: I0227 18:27:35.168729 4700 generic.go:334] "Generic (PLEG): container finished" podID="d37cbaaf-1dee-4efd-9327-7bee80986a35" containerID="99203e60bd2f4332e38617f1a8b60b443dcded5a1ec8b88d77f76aeeb4c0d145" exitCode=0 Feb 27 18:27:35 crc kubenswrapper[4700]: I0227 18:27:35.168778 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" event={"ID":"d37cbaaf-1dee-4efd-9327-7bee80986a35","Type":"ContainerDied","Data":"99203e60bd2f4332e38617f1a8b60b443dcded5a1ec8b88d77f76aeeb4c0d145"} Feb 27 18:27:36 crc kubenswrapper[4700]: I0227 18:27:36.600539 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" Feb 27 18:27:36 crc kubenswrapper[4700]: I0227 18:27:36.679742 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjvdh\" (UniqueName: \"kubernetes.io/projected/d37cbaaf-1dee-4efd-9327-7bee80986a35-kube-api-access-hjvdh\") pod \"d37cbaaf-1dee-4efd-9327-7bee80986a35\" (UID: \"d37cbaaf-1dee-4efd-9327-7bee80986a35\") " Feb 27 18:27:36 crc kubenswrapper[4700]: I0227 18:27:36.692904 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d37cbaaf-1dee-4efd-9327-7bee80986a35-kube-api-access-hjvdh" (OuterVolumeSpecName: "kube-api-access-hjvdh") pod "d37cbaaf-1dee-4efd-9327-7bee80986a35" (UID: "d37cbaaf-1dee-4efd-9327-7bee80986a35"). InnerVolumeSpecName "kube-api-access-hjvdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:27:36 crc kubenswrapper[4700]: I0227 18:27:36.783320 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjvdh\" (UniqueName: \"kubernetes.io/projected/d37cbaaf-1dee-4efd-9327-7bee80986a35-kube-api-access-hjvdh\") on node \"crc\" DevicePath \"\"" Feb 27 18:27:37 crc kubenswrapper[4700]: I0227 18:27:37.205570 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" event={"ID":"d37cbaaf-1dee-4efd-9327-7bee80986a35","Type":"ContainerDied","Data":"b52f8a05c71006850ae39fde28800be4e5a2a2fa515d775f7cac159f03db22dd"} Feb 27 18:27:37 crc kubenswrapper[4700]: I0227 18:27:37.205997 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b52f8a05c71006850ae39fde28800be4e5a2a2fa515d775f7cac159f03db22dd" Feb 27 18:27:37 crc kubenswrapper[4700]: I0227 18:27:37.205694 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536926-s6ft8" Feb 27 18:27:37 crc kubenswrapper[4700]: I0227 18:27:37.687709 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536926-s6ft8"] Feb 27 18:27:37 crc kubenswrapper[4700]: I0227 18:27:37.697094 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536926-s6ft8"] Feb 27 18:27:38 crc kubenswrapper[4700]: E0227 18:27:38.984531 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:27:38 crc kubenswrapper[4700]: I0227 18:27:38.994940 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" path="/var/lib/kubelet/pods/d37cbaaf-1dee-4efd-9327-7bee80986a35/volumes" Feb 27 18:27:40 crc kubenswrapper[4700]: I0227 18:27:40.254885 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" event={"ID":"3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5","Type":"ContainerStarted","Data":"eddf47b752176a5dca7ef89b2e9da9f47d2a28693a50fab9e03d35eee675a117"} Feb 27 18:27:40 crc kubenswrapper[4700]: I0227 18:27:40.290449 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" podStartSLOduration=1.565124227 podStartE2EDuration="17m40.290411843s" podCreationTimestamp="2026-02-27 18:10:00 +0000 UTC" firstStartedPulling="2026-02-27 18:10:01.073034917 +0000 UTC m=+4161.058347674" lastFinishedPulling="2026-02-27 18:27:39.798322533 +0000 UTC m=+5219.783635290" observedRunningTime="2026-02-27 18:27:40.277304961 +0000 UTC m=+5220.262617708" watchObservedRunningTime="2026-02-27 18:27:40.290411843 +0000 UTC m=+5220.275724630" Feb 27 18:27:41 crc kubenswrapper[4700]: I0227 18:27:41.263138 4700 generic.go:334] "Generic (PLEG): container finished" podID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" containerID="eddf47b752176a5dca7ef89b2e9da9f47d2a28693a50fab9e03d35eee675a117" exitCode=0 Feb 27 18:27:41 crc kubenswrapper[4700]: I0227 18:27:41.263214 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" event={"ID":"3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5","Type":"ContainerDied","Data":"eddf47b752176a5dca7ef89b2e9da9f47d2a28693a50fab9e03d35eee675a117"} Feb 27 18:27:41 crc kubenswrapper[4700]: I0227 18:27:41.982438 4700 scope.go:117] "RemoveContainer" containerID="b3fdfb06db2c06428e3b3e96ef9804be97292a50a8f04f891caec5e729731adf" Feb 27 18:27:41 crc kubenswrapper[4700]: E0227 18:27:41.982941 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:27:42 crc kubenswrapper[4700]: I0227 18:27:42.815741 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" Feb 27 18:27:42 crc kubenswrapper[4700]: I0227 18:27:42.946818 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtm8j\" (UniqueName: \"kubernetes.io/projected/3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5-kube-api-access-gtm8j\") pod \"3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5\" (UID: \"3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5\") " Feb 27 18:27:42 crc kubenswrapper[4700]: I0227 18:27:42.956887 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5-kube-api-access-gtm8j" (OuterVolumeSpecName: "kube-api-access-gtm8j") pod "3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" (UID: "3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5"). InnerVolumeSpecName "kube-api-access-gtm8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:27:43 crc kubenswrapper[4700]: I0227 18:27:43.050784 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtm8j\" (UniqueName: \"kubernetes.io/projected/3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5-kube-api-access-gtm8j\") on node \"crc\" DevicePath \"\"" Feb 27 18:27:43 crc kubenswrapper[4700]: I0227 18:27:43.290018 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" event={"ID":"3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5","Type":"ContainerDied","Data":"e20847a166818207490b96d0764bc0a4028117728ae13315b61ff1d7d2895c27"} Feb 27 18:27:43 crc kubenswrapper[4700]: I0227 18:27:43.290061 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e20847a166818207490b96d0764bc0a4028117728ae13315b61ff1d7d2895c27" Feb 27 18:27:43 crc kubenswrapper[4700]: I0227 18:27:43.290271 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536930-xbg7f" Feb 27 18:27:43 crc kubenswrapper[4700]: I0227 18:27:43.350437 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536930-xbg7f"] Feb 27 18:27:43 crc kubenswrapper[4700]: I0227 18:27:43.375107 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536930-xbg7f"] Feb 27 18:27:44 crc kubenswrapper[4700]: I0227 18:27:44.994844 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" path="/var/lib/kubelet/pods/3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5/volumes" Feb 27 18:27:52 crc kubenswrapper[4700]: I0227 18:27:52.982249 4700 scope.go:117] "RemoveContainer" containerID="b3fdfb06db2c06428e3b3e96ef9804be97292a50a8f04f891caec5e729731adf" Feb 27 18:27:52 crc kubenswrapper[4700]: E0227 18:27:52.983631 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:27:53 crc kubenswrapper[4700]: E0227 18:27:53.985220 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:28:00 crc kubenswrapper[4700]: I0227 18:28:00.169494 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536948-zt6lq"] Feb 27 18:28:00 crc kubenswrapper[4700]: E0227 18:28:00.171044 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3de1a57-09ac-4688-b26b-08846e57a0f4" containerName="extract-content" Feb 27 18:28:00 crc kubenswrapper[4700]: I0227 18:28:00.171068 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3de1a57-09ac-4688-b26b-08846e57a0f4" containerName="extract-content" Feb 27 18:28:00 crc kubenswrapper[4700]: E0227 18:28:00.171098 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" containerName="oc" Feb 27 18:28:00 crc kubenswrapper[4700]: I0227 18:28:00.171110 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" containerName="oc" Feb 27 18:28:00 crc kubenswrapper[4700]: E0227 18:28:00.171135 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3de1a57-09ac-4688-b26b-08846e57a0f4" containerName="extract-utilities" Feb 27 18:28:00 crc kubenswrapper[4700]: I0227 18:28:00.171149 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3de1a57-09ac-4688-b26b-08846e57a0f4" containerName="extract-utilities" Feb 27 18:28:00 crc kubenswrapper[4700]: E0227 18:28:00.171179 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3c40c88-7b53-43dd-8357-17b25bdcc22d" containerName="oc" Feb 27 18:28:00 crc kubenswrapper[4700]: I0227 18:28:00.171190 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3c40c88-7b53-43dd-8357-17b25bdcc22d" containerName="oc" Feb 27 18:28:00 crc kubenswrapper[4700]: E0227 18:28:00.171215 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" containerName="oc" Feb 27 18:28:00 crc kubenswrapper[4700]: I0227 18:28:00.171227 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" containerName="oc" Feb 27 18:28:00 crc kubenswrapper[4700]: E0227 18:28:00.171259 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3de1a57-09ac-4688-b26b-08846e57a0f4" containerName="registry-server" Feb 27 18:28:00 crc kubenswrapper[4700]: I0227 18:28:00.171270 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3de1a57-09ac-4688-b26b-08846e57a0f4" containerName="registry-server" Feb 27 18:28:00 crc kubenswrapper[4700]: I0227 18:28:00.171646 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cd6cbf8-4159-4eb9-84c2-b55c1cd85dc5" containerName="oc" Feb 27 18:28:00 crc kubenswrapper[4700]: I0227 18:28:00.171682 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="d37cbaaf-1dee-4efd-9327-7bee80986a35" containerName="oc" Feb 27 18:28:00 crc kubenswrapper[4700]: I0227 18:28:00.171709 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3de1a57-09ac-4688-b26b-08846e57a0f4" containerName="registry-server" Feb 27 18:28:00 crc kubenswrapper[4700]: I0227 18:28:00.171736 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3c40c88-7b53-43dd-8357-17b25bdcc22d" containerName="oc" Feb 27 18:28:00 crc kubenswrapper[4700]: I0227 18:28:00.173081 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536948-zt6lq" Feb 27 18:28:00 crc kubenswrapper[4700]: I0227 18:28:00.197044 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536948-zt6lq"] Feb 27 18:28:00 crc kubenswrapper[4700]: I0227 18:28:00.275625 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vshs\" (UniqueName: \"kubernetes.io/projected/4c99a671-4160-4665-bb56-5630aad6ad58-kube-api-access-9vshs\") pod \"auto-csr-approver-29536948-zt6lq\" (UID: \"4c99a671-4160-4665-bb56-5630aad6ad58\") " pod="openshift-infra/auto-csr-approver-29536948-zt6lq" Feb 27 18:28:00 crc kubenswrapper[4700]: I0227 18:28:00.379197 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vshs\" (UniqueName: \"kubernetes.io/projected/4c99a671-4160-4665-bb56-5630aad6ad58-kube-api-access-9vshs\") pod \"auto-csr-approver-29536948-zt6lq\" (UID: \"4c99a671-4160-4665-bb56-5630aad6ad58\") " pod="openshift-infra/auto-csr-approver-29536948-zt6lq" Feb 27 18:28:00 crc kubenswrapper[4700]: I0227 18:28:00.417744 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vshs\" (UniqueName: \"kubernetes.io/projected/4c99a671-4160-4665-bb56-5630aad6ad58-kube-api-access-9vshs\") pod \"auto-csr-approver-29536948-zt6lq\" (UID: \"4c99a671-4160-4665-bb56-5630aad6ad58\") " pod="openshift-infra/auto-csr-approver-29536948-zt6lq" Feb 27 18:28:00 crc kubenswrapper[4700]: I0227 18:28:00.500982 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536948-zt6lq" Feb 27 18:28:01 crc kubenswrapper[4700]: I0227 18:28:01.013072 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536948-zt6lq"] Feb 27 18:28:01 crc kubenswrapper[4700]: I0227 18:28:01.543609 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536948-zt6lq" event={"ID":"4c99a671-4160-4665-bb56-5630aad6ad58","Type":"ContainerStarted","Data":"71d602180bb4f8fa52ce8d9c397d6c280b123f9a1a82a4cca76abd5fbd601056"} Feb 27 18:28:02 crc kubenswrapper[4700]: I0227 18:28:02.557581 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536948-zt6lq" event={"ID":"4c99a671-4160-4665-bb56-5630aad6ad58","Type":"ContainerStarted","Data":"bef2cb63a3b1ee53e15e8d40529d1fc26948d802f5478c8b2e25db61e155c8e1"} Feb 27 18:28:02 crc kubenswrapper[4700]: I0227 18:28:02.585076 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536948-zt6lq" podStartSLOduration=1.445379632 podStartE2EDuration="2.585054201s" podCreationTimestamp="2026-02-27 18:28:00 +0000 UTC" firstStartedPulling="2026-02-27 18:28:01.01421549 +0000 UTC m=+5240.999528247" lastFinishedPulling="2026-02-27 18:28:02.153890059 +0000 UTC m=+5242.139202816" observedRunningTime="2026-02-27 18:28:02.5796124 +0000 UTC m=+5242.564925177" watchObservedRunningTime="2026-02-27 18:28:02.585054201 +0000 UTC m=+5242.570366958" Feb 27 18:28:03 crc kubenswrapper[4700]: I0227 18:28:03.572120 4700 generic.go:334] "Generic (PLEG): container finished" podID="4c99a671-4160-4665-bb56-5630aad6ad58" containerID="bef2cb63a3b1ee53e15e8d40529d1fc26948d802f5478c8b2e25db61e155c8e1" exitCode=0 Feb 27 18:28:03 crc kubenswrapper[4700]: I0227 18:28:03.572325 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536948-zt6lq" event={"ID":"4c99a671-4160-4665-bb56-5630aad6ad58","Type":"ContainerDied","Data":"bef2cb63a3b1ee53e15e8d40529d1fc26948d802f5478c8b2e25db61e155c8e1"} Feb 27 18:28:05 crc kubenswrapper[4700]: I0227 18:28:05.075794 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536948-zt6lq" Feb 27 18:28:05 crc kubenswrapper[4700]: I0227 18:28:05.225828 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vshs\" (UniqueName: \"kubernetes.io/projected/4c99a671-4160-4665-bb56-5630aad6ad58-kube-api-access-9vshs\") pod \"4c99a671-4160-4665-bb56-5630aad6ad58\" (UID: \"4c99a671-4160-4665-bb56-5630aad6ad58\") " Feb 27 18:28:05 crc kubenswrapper[4700]: I0227 18:28:05.234542 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c99a671-4160-4665-bb56-5630aad6ad58-kube-api-access-9vshs" (OuterVolumeSpecName: "kube-api-access-9vshs") pod "4c99a671-4160-4665-bb56-5630aad6ad58" (UID: "4c99a671-4160-4665-bb56-5630aad6ad58"). InnerVolumeSpecName "kube-api-access-9vshs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:28:05 crc kubenswrapper[4700]: I0227 18:28:05.328572 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vshs\" (UniqueName: \"kubernetes.io/projected/4c99a671-4160-4665-bb56-5630aad6ad58-kube-api-access-9vshs\") on node \"crc\" DevicePath \"\"" Feb 27 18:28:05 crc kubenswrapper[4700]: I0227 18:28:05.613442 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536948-zt6lq" event={"ID":"4c99a671-4160-4665-bb56-5630aad6ad58","Type":"ContainerDied","Data":"71d602180bb4f8fa52ce8d9c397d6c280b123f9a1a82a4cca76abd5fbd601056"} Feb 27 18:28:05 crc kubenswrapper[4700]: I0227 18:28:05.613503 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="71d602180bb4f8fa52ce8d9c397d6c280b123f9a1a82a4cca76abd5fbd601056" Feb 27 18:28:05 crc kubenswrapper[4700]: I0227 18:28:05.613966 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536948-zt6lq" Feb 27 18:28:05 crc kubenswrapper[4700]: I0227 18:28:05.686963 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536942-st4kg"] Feb 27 18:28:05 crc kubenswrapper[4700]: I0227 18:28:05.698084 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536942-st4kg"] Feb 27 18:28:06 crc kubenswrapper[4700]: I0227 18:28:06.983455 4700 scope.go:117] "RemoveContainer" containerID="b3fdfb06db2c06428e3b3e96ef9804be97292a50a8f04f891caec5e729731adf" Feb 27 18:28:06 crc kubenswrapper[4700]: E0227 18:28:06.984278 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:28:06 crc kubenswrapper[4700]: E0227 18:28:06.984300 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:28:07 crc kubenswrapper[4700]: I0227 18:28:07.001875 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="237d2183-87cf-4196-944e-ddbda807b62b" path="/var/lib/kubelet/pods/237d2183-87cf-4196-944e-ddbda807b62b/volumes" Feb 27 18:28:13 crc kubenswrapper[4700]: I0227 18:28:13.361562 4700 scope.go:117] "RemoveContainer" containerID="cdff47536c8d552a075a8e0a9cf511f2b71c81f2d61d3cb94c490b7d1f5176ee" Feb 27 18:28:19 crc kubenswrapper[4700]: I0227 18:28:19.981696 4700 scope.go:117] "RemoveContainer" containerID="b3fdfb06db2c06428e3b3e96ef9804be97292a50a8f04f891caec5e729731adf" Feb 27 18:28:19 crc kubenswrapper[4700]: E0227 18:28:19.982428 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:28:21 crc kubenswrapper[4700]: E0227 18:28:21.983491 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:28:33 crc kubenswrapper[4700]: E0227 18:28:33.983780 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:28:34 crc kubenswrapper[4700]: I0227 18:28:34.982728 4700 scope.go:117] "RemoveContainer" containerID="b3fdfb06db2c06428e3b3e96ef9804be97292a50a8f04f891caec5e729731adf" Feb 27 18:28:34 crc kubenswrapper[4700]: E0227 18:28:34.983038 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:28:45 crc kubenswrapper[4700]: I0227 18:28:45.982899 4700 scope.go:117] "RemoveContainer" containerID="b3fdfb06db2c06428e3b3e96ef9804be97292a50a8f04f891caec5e729731adf" Feb 27 18:28:45 crc kubenswrapper[4700]: E0227 18:28:45.984094 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:28:47 crc kubenswrapper[4700]: E0227 18:28:47.985532 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:28:58 crc kubenswrapper[4700]: E0227 18:28:58.984783 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:28:59 crc kubenswrapper[4700]: I0227 18:28:59.981448 4700 scope.go:117] "RemoveContainer" containerID="b3fdfb06db2c06428e3b3e96ef9804be97292a50a8f04f891caec5e729731adf" Feb 27 18:28:59 crc kubenswrapper[4700]: E0227 18:28:59.982491 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:29:09 crc kubenswrapper[4700]: E0227 18:29:09.988358 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:29:14 crc kubenswrapper[4700]: I0227 18:29:14.984669 4700 scope.go:117] "RemoveContainer" containerID="b3fdfb06db2c06428e3b3e96ef9804be97292a50a8f04f891caec5e729731adf" Feb 27 18:29:14 crc kubenswrapper[4700]: E0227 18:29:14.985813 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:29:23 crc kubenswrapper[4700]: E0227 18:29:23.983691 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:29:27 crc kubenswrapper[4700]: I0227 18:29:27.983018 4700 scope.go:117] "RemoveContainer" containerID="b3fdfb06db2c06428e3b3e96ef9804be97292a50a8f04f891caec5e729731adf" Feb 27 18:29:27 crc kubenswrapper[4700]: E0227 18:29:27.984134 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:29:37 crc kubenswrapper[4700]: E0227 18:29:37.983647 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:29:40 crc kubenswrapper[4700]: I0227 18:29:40.991258 4700 scope.go:117] "RemoveContainer" containerID="b3fdfb06db2c06428e3b3e96ef9804be97292a50a8f04f891caec5e729731adf" Feb 27 18:29:40 crc kubenswrapper[4700]: E0227 18:29:40.992559 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:29:46 crc kubenswrapper[4700]: I0227 18:29:46.297496 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lrt8r"] Feb 27 18:29:46 crc kubenswrapper[4700]: E0227 18:29:46.298687 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c99a671-4160-4665-bb56-5630aad6ad58" containerName="oc" Feb 27 18:29:46 crc kubenswrapper[4700]: I0227 18:29:46.298702 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c99a671-4160-4665-bb56-5630aad6ad58" containerName="oc" Feb 27 18:29:46 crc kubenswrapper[4700]: I0227 18:29:46.298982 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c99a671-4160-4665-bb56-5630aad6ad58" containerName="oc" Feb 27 18:29:46 crc kubenswrapper[4700]: I0227 18:29:46.300933 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lrt8r" Feb 27 18:29:46 crc kubenswrapper[4700]: I0227 18:29:46.319989 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lrt8r"] Feb 27 18:29:46 crc kubenswrapper[4700]: I0227 18:29:46.466192 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98c1faf7-b2cc-47c3-89e3-20e563b9f507-catalog-content\") pod \"redhat-operators-lrt8r\" (UID: \"98c1faf7-b2cc-47c3-89e3-20e563b9f507\") " pod="openshift-marketplace/redhat-operators-lrt8r" Feb 27 18:29:46 crc kubenswrapper[4700]: I0227 18:29:46.466249 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98c1faf7-b2cc-47c3-89e3-20e563b9f507-utilities\") pod \"redhat-operators-lrt8r\" (UID: \"98c1faf7-b2cc-47c3-89e3-20e563b9f507\") " pod="openshift-marketplace/redhat-operators-lrt8r" Feb 27 18:29:46 crc kubenswrapper[4700]: I0227 18:29:46.466296 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nwmt\" (UniqueName: \"kubernetes.io/projected/98c1faf7-b2cc-47c3-89e3-20e563b9f507-kube-api-access-6nwmt\") pod \"redhat-operators-lrt8r\" (UID: \"98c1faf7-b2cc-47c3-89e3-20e563b9f507\") " pod="openshift-marketplace/redhat-operators-lrt8r" Feb 27 18:29:46 crc kubenswrapper[4700]: I0227 18:29:46.568442 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98c1faf7-b2cc-47c3-89e3-20e563b9f507-catalog-content\") pod \"redhat-operators-lrt8r\" (UID: \"98c1faf7-b2cc-47c3-89e3-20e563b9f507\") " pod="openshift-marketplace/redhat-operators-lrt8r" Feb 27 18:29:46 crc kubenswrapper[4700]: I0227 18:29:46.568507 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98c1faf7-b2cc-47c3-89e3-20e563b9f507-utilities\") pod \"redhat-operators-lrt8r\" (UID: \"98c1faf7-b2cc-47c3-89e3-20e563b9f507\") " pod="openshift-marketplace/redhat-operators-lrt8r" Feb 27 18:29:46 crc kubenswrapper[4700]: I0227 18:29:46.568554 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nwmt\" (UniqueName: \"kubernetes.io/projected/98c1faf7-b2cc-47c3-89e3-20e563b9f507-kube-api-access-6nwmt\") pod \"redhat-operators-lrt8r\" (UID: \"98c1faf7-b2cc-47c3-89e3-20e563b9f507\") " pod="openshift-marketplace/redhat-operators-lrt8r" Feb 27 18:29:46 crc kubenswrapper[4700]: I0227 18:29:46.568866 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98c1faf7-b2cc-47c3-89e3-20e563b9f507-catalog-content\") pod \"redhat-operators-lrt8r\" (UID: \"98c1faf7-b2cc-47c3-89e3-20e563b9f507\") " pod="openshift-marketplace/redhat-operators-lrt8r" Feb 27 18:29:46 crc kubenswrapper[4700]: I0227 18:29:46.569366 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98c1faf7-b2cc-47c3-89e3-20e563b9f507-utilities\") pod \"redhat-operators-lrt8r\" (UID: \"98c1faf7-b2cc-47c3-89e3-20e563b9f507\") " pod="openshift-marketplace/redhat-operators-lrt8r" Feb 27 18:29:46 crc kubenswrapper[4700]: I0227 18:29:46.591323 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nwmt\" (UniqueName: \"kubernetes.io/projected/98c1faf7-b2cc-47c3-89e3-20e563b9f507-kube-api-access-6nwmt\") pod \"redhat-operators-lrt8r\" (UID: \"98c1faf7-b2cc-47c3-89e3-20e563b9f507\") " pod="openshift-marketplace/redhat-operators-lrt8r" Feb 27 18:29:46 crc kubenswrapper[4700]: I0227 18:29:46.618704 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lrt8r" Feb 27 18:29:47 crc kubenswrapper[4700]: I0227 18:29:47.132683 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lrt8r"] Feb 27 18:29:47 crc kubenswrapper[4700]: I0227 18:29:47.886098 4700 generic.go:334] "Generic (PLEG): container finished" podID="98c1faf7-b2cc-47c3-89e3-20e563b9f507" containerID="2872e40bc161207c25f7614f9528cd7c88871edac6c66c0f2de1f6cb2a3ae789" exitCode=0 Feb 27 18:29:47 crc kubenswrapper[4700]: I0227 18:29:47.886288 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrt8r" event={"ID":"98c1faf7-b2cc-47c3-89e3-20e563b9f507","Type":"ContainerDied","Data":"2872e40bc161207c25f7614f9528cd7c88871edac6c66c0f2de1f6cb2a3ae789"} Feb 27 18:29:47 crc kubenswrapper[4700]: I0227 18:29:47.886363 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrt8r" event={"ID":"98c1faf7-b2cc-47c3-89e3-20e563b9f507","Type":"ContainerStarted","Data":"d0111d8397a2c72d15144e0180d3ac0d423eaafafe1d957985a397bf6212f384"} Feb 27 18:29:48 crc kubenswrapper[4700]: I0227 18:29:48.896140 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrt8r" event={"ID":"98c1faf7-b2cc-47c3-89e3-20e563b9f507","Type":"ContainerStarted","Data":"2437c89416ad507ad89800fdeef354efa0de5dd6a65d8879874af48e5b58f3a1"} Feb 27 18:29:49 crc kubenswrapper[4700]: E0227 18:29:49.985061 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:29:52 crc kubenswrapper[4700]: I0227 18:29:52.982007 4700 scope.go:117] "RemoveContainer" containerID="b3fdfb06db2c06428e3b3e96ef9804be97292a50a8f04f891caec5e729731adf" Feb 27 18:29:52 crc kubenswrapper[4700]: E0227 18:29:52.984289 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:29:53 crc kubenswrapper[4700]: I0227 18:29:53.958611 4700 generic.go:334] "Generic (PLEG): container finished" podID="98c1faf7-b2cc-47c3-89e3-20e563b9f507" containerID="2437c89416ad507ad89800fdeef354efa0de5dd6a65d8879874af48e5b58f3a1" exitCode=0 Feb 27 18:29:53 crc kubenswrapper[4700]: I0227 18:29:53.958682 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrt8r" event={"ID":"98c1faf7-b2cc-47c3-89e3-20e563b9f507","Type":"ContainerDied","Data":"2437c89416ad507ad89800fdeef354efa0de5dd6a65d8879874af48e5b58f3a1"} Feb 27 18:29:54 crc kubenswrapper[4700]: I0227 18:29:54.969088 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrt8r" event={"ID":"98c1faf7-b2cc-47c3-89e3-20e563b9f507","Type":"ContainerStarted","Data":"c63aeace47d8f492efd296c51962ea3bd9afdfc0a9cc49085c24e07c443f768c"} Feb 27 18:29:55 crc kubenswrapper[4700]: I0227 18:29:55.008783 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lrt8r" podStartSLOduration=2.51984136 podStartE2EDuration="9.008758767s" podCreationTimestamp="2026-02-27 18:29:46 +0000 UTC" firstStartedPulling="2026-02-27 18:29:47.888633126 +0000 UTC m=+5347.873945873" lastFinishedPulling="2026-02-27 18:29:54.377550503 +0000 UTC m=+5354.362863280" observedRunningTime="2026-02-27 18:29:54.999642359 +0000 UTC m=+5354.984955126" watchObservedRunningTime="2026-02-27 18:29:55.008758767 +0000 UTC m=+5354.994071514" Feb 27 18:29:56 crc kubenswrapper[4700]: I0227 18:29:56.619713 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lrt8r" Feb 27 18:29:56 crc kubenswrapper[4700]: I0227 18:29:56.620101 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lrt8r" Feb 27 18:29:57 crc kubenswrapper[4700]: I0227 18:29:57.716254 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lrt8r" podUID="98c1faf7-b2cc-47c3-89e3-20e563b9f507" containerName="registry-server" probeResult="failure" output=< Feb 27 18:29:57 crc kubenswrapper[4700]: timeout: failed to connect service ":50051" within 1s Feb 27 18:29:57 crc kubenswrapper[4700]: > Feb 27 18:29:58 crc kubenswrapper[4700]: I0227 18:29:58.493445 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-h89qg"] Feb 27 18:29:58 crc kubenswrapper[4700]: I0227 18:29:58.496318 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h89qg" Feb 27 18:29:58 crc kubenswrapper[4700]: I0227 18:29:58.538074 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h89qg"] Feb 27 18:29:58 crc kubenswrapper[4700]: I0227 18:29:58.613770 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plrxr\" (UniqueName: \"kubernetes.io/projected/b0fd7b6d-8f71-4bff-b903-cdb5a0954f87-kube-api-access-plrxr\") pod \"community-operators-h89qg\" (UID: \"b0fd7b6d-8f71-4bff-b903-cdb5a0954f87\") " pod="openshift-marketplace/community-operators-h89qg" Feb 27 18:29:58 crc kubenswrapper[4700]: I0227 18:29:58.613834 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0fd7b6d-8f71-4bff-b903-cdb5a0954f87-catalog-content\") pod \"community-operators-h89qg\" (UID: \"b0fd7b6d-8f71-4bff-b903-cdb5a0954f87\") " pod="openshift-marketplace/community-operators-h89qg" Feb 27 18:29:58 crc kubenswrapper[4700]: I0227 18:29:58.615070 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0fd7b6d-8f71-4bff-b903-cdb5a0954f87-utilities\") pod \"community-operators-h89qg\" (UID: \"b0fd7b6d-8f71-4bff-b903-cdb5a0954f87\") " pod="openshift-marketplace/community-operators-h89qg" Feb 27 18:29:58 crc kubenswrapper[4700]: I0227 18:29:58.717070 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0fd7b6d-8f71-4bff-b903-cdb5a0954f87-utilities\") pod \"community-operators-h89qg\" (UID: \"b0fd7b6d-8f71-4bff-b903-cdb5a0954f87\") " pod="openshift-marketplace/community-operators-h89qg" Feb 27 18:29:58 crc kubenswrapper[4700]: I0227 18:29:58.717270 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plrxr\" (UniqueName: \"kubernetes.io/projected/b0fd7b6d-8f71-4bff-b903-cdb5a0954f87-kube-api-access-plrxr\") pod \"community-operators-h89qg\" (UID: \"b0fd7b6d-8f71-4bff-b903-cdb5a0954f87\") " pod="openshift-marketplace/community-operators-h89qg" Feb 27 18:29:58 crc kubenswrapper[4700]: I0227 18:29:58.717333 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0fd7b6d-8f71-4bff-b903-cdb5a0954f87-catalog-content\") pod \"community-operators-h89qg\" (UID: \"b0fd7b6d-8f71-4bff-b903-cdb5a0954f87\") " pod="openshift-marketplace/community-operators-h89qg" Feb 27 18:29:58 crc kubenswrapper[4700]: I0227 18:29:58.717604 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0fd7b6d-8f71-4bff-b903-cdb5a0954f87-utilities\") pod \"community-operators-h89qg\" (UID: \"b0fd7b6d-8f71-4bff-b903-cdb5a0954f87\") " pod="openshift-marketplace/community-operators-h89qg" Feb 27 18:29:58 crc kubenswrapper[4700]: I0227 18:29:58.717661 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0fd7b6d-8f71-4bff-b903-cdb5a0954f87-catalog-content\") pod \"community-operators-h89qg\" (UID: \"b0fd7b6d-8f71-4bff-b903-cdb5a0954f87\") " pod="openshift-marketplace/community-operators-h89qg" Feb 27 18:29:58 crc kubenswrapper[4700]: I0227 18:29:58.751510 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plrxr\" (UniqueName: \"kubernetes.io/projected/b0fd7b6d-8f71-4bff-b903-cdb5a0954f87-kube-api-access-plrxr\") pod \"community-operators-h89qg\" (UID: \"b0fd7b6d-8f71-4bff-b903-cdb5a0954f87\") " pod="openshift-marketplace/community-operators-h89qg" Feb 27 18:29:58 crc kubenswrapper[4700]: I0227 18:29:58.829509 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h89qg" Feb 27 18:29:59 crc kubenswrapper[4700]: I0227 18:29:59.451217 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h89qg"] Feb 27 18:29:59 crc kubenswrapper[4700]: W0227 18:29:59.466003 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb0fd7b6d_8f71_4bff_b903_cdb5a0954f87.slice/crio-91f4ab7e9143d9c6e25fbca874f1eb387cb0d59c6d36e92fc5ee17a09681ea82 WatchSource:0}: Error finding container 91f4ab7e9143d9c6e25fbca874f1eb387cb0d59c6d36e92fc5ee17a09681ea82: Status 404 returned error can't find the container with id 91f4ab7e9143d9c6e25fbca874f1eb387cb0d59c6d36e92fc5ee17a09681ea82 Feb 27 18:30:00 crc kubenswrapper[4700]: I0227 18:30:00.052300 4700 generic.go:334] "Generic (PLEG): container finished" podID="b0fd7b6d-8f71-4bff-b903-cdb5a0954f87" containerID="4b65a79cf2eb03a8dd97bcdd54a620d09775cde1106b8a7b7fff34d53100a0b9" exitCode=0 Feb 27 18:30:00 crc kubenswrapper[4700]: I0227 18:30:00.052355 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h89qg" event={"ID":"b0fd7b6d-8f71-4bff-b903-cdb5a0954f87","Type":"ContainerDied","Data":"4b65a79cf2eb03a8dd97bcdd54a620d09775cde1106b8a7b7fff34d53100a0b9"} Feb 27 18:30:00 crc kubenswrapper[4700]: I0227 18:30:00.052384 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h89qg" event={"ID":"b0fd7b6d-8f71-4bff-b903-cdb5a0954f87","Type":"ContainerStarted","Data":"91f4ab7e9143d9c6e25fbca874f1eb387cb0d59c6d36e92fc5ee17a09681ea82"} Feb 27 18:30:00 crc kubenswrapper[4700]: I0227 18:30:00.176785 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536950-ln6wp"] Feb 27 18:30:00 crc kubenswrapper[4700]: I0227 18:30:00.178908 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536950-ln6wp" Feb 27 18:30:00 crc kubenswrapper[4700]: I0227 18:30:00.180370 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 27 18:30:00 crc kubenswrapper[4700]: I0227 18:30:00.182047 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 27 18:30:00 crc kubenswrapper[4700]: I0227 18:30:00.191062 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536950-458j5"] Feb 27 18:30:00 crc kubenswrapper[4700]: I0227 18:30:00.193444 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536950-458j5" Feb 27 18:30:00 crc kubenswrapper[4700]: I0227 18:30:00.205079 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536950-ln6wp"] Feb 27 18:30:00 crc kubenswrapper[4700]: I0227 18:30:00.215040 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536950-458j5"] Feb 27 18:30:00 crc kubenswrapper[4700]: I0227 18:30:00.274966 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftprg\" (UniqueName: \"kubernetes.io/projected/050b88a9-b342-4990-9bac-17c98b014f49-kube-api-access-ftprg\") pod \"collect-profiles-29536950-ln6wp\" (UID: \"050b88a9-b342-4990-9bac-17c98b014f49\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536950-ln6wp" Feb 27 18:30:00 crc kubenswrapper[4700]: I0227 18:30:00.275225 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/050b88a9-b342-4990-9bac-17c98b014f49-config-volume\") pod \"collect-profiles-29536950-ln6wp\" (UID: \"050b88a9-b342-4990-9bac-17c98b014f49\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536950-ln6wp" Feb 27 18:30:00 crc kubenswrapper[4700]: I0227 18:30:00.275450 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/050b88a9-b342-4990-9bac-17c98b014f49-secret-volume\") pod \"collect-profiles-29536950-ln6wp\" (UID: \"050b88a9-b342-4990-9bac-17c98b014f49\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536950-ln6wp" Feb 27 18:30:00 crc kubenswrapper[4700]: I0227 18:30:00.377268 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftprg\" (UniqueName: \"kubernetes.io/projected/050b88a9-b342-4990-9bac-17c98b014f49-kube-api-access-ftprg\") pod \"collect-profiles-29536950-ln6wp\" (UID: \"050b88a9-b342-4990-9bac-17c98b014f49\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536950-ln6wp" Feb 27 18:30:00 crc kubenswrapper[4700]: I0227 18:30:00.377331 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/050b88a9-b342-4990-9bac-17c98b014f49-config-volume\") pod \"collect-profiles-29536950-ln6wp\" (UID: \"050b88a9-b342-4990-9bac-17c98b014f49\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536950-ln6wp" Feb 27 18:30:00 crc kubenswrapper[4700]: I0227 18:30:00.377413 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgn64\" (UniqueName: \"kubernetes.io/projected/74489fab-6aca-4a3f-9303-04a2b82f1fde-kube-api-access-hgn64\") pod \"auto-csr-approver-29536950-458j5\" (UID: \"74489fab-6aca-4a3f-9303-04a2b82f1fde\") " pod="openshift-infra/auto-csr-approver-29536950-458j5" Feb 27 18:30:00 crc kubenswrapper[4700]: I0227 18:30:00.377443 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/050b88a9-b342-4990-9bac-17c98b014f49-secret-volume\") pod \"collect-profiles-29536950-ln6wp\" (UID: \"050b88a9-b342-4990-9bac-17c98b014f49\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536950-ln6wp" Feb 27 18:30:00 crc kubenswrapper[4700]: I0227 18:30:00.378571 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/050b88a9-b342-4990-9bac-17c98b014f49-config-volume\") pod \"collect-profiles-29536950-ln6wp\" (UID: \"050b88a9-b342-4990-9bac-17c98b014f49\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536950-ln6wp" Feb 27 18:30:00 crc kubenswrapper[4700]: I0227 18:30:00.385022 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/050b88a9-b342-4990-9bac-17c98b014f49-secret-volume\") pod \"collect-profiles-29536950-ln6wp\" (UID: \"050b88a9-b342-4990-9bac-17c98b014f49\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536950-ln6wp" Feb 27 18:30:00 crc kubenswrapper[4700]: I0227 18:30:00.400934 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftprg\" (UniqueName: \"kubernetes.io/projected/050b88a9-b342-4990-9bac-17c98b014f49-kube-api-access-ftprg\") pod \"collect-profiles-29536950-ln6wp\" (UID: \"050b88a9-b342-4990-9bac-17c98b014f49\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536950-ln6wp" Feb 27 18:30:00 crc kubenswrapper[4700]: I0227 18:30:00.480182 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgn64\" (UniqueName: \"kubernetes.io/projected/74489fab-6aca-4a3f-9303-04a2b82f1fde-kube-api-access-hgn64\") pod \"auto-csr-approver-29536950-458j5\" (UID: \"74489fab-6aca-4a3f-9303-04a2b82f1fde\") " pod="openshift-infra/auto-csr-approver-29536950-458j5" Feb 27 18:30:00 crc kubenswrapper[4700]: I0227 18:30:00.495963 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgn64\" (UniqueName: \"kubernetes.io/projected/74489fab-6aca-4a3f-9303-04a2b82f1fde-kube-api-access-hgn64\") pod \"auto-csr-approver-29536950-458j5\" (UID: \"74489fab-6aca-4a3f-9303-04a2b82f1fde\") " pod="openshift-infra/auto-csr-approver-29536950-458j5" Feb 27 18:30:00 crc kubenswrapper[4700]: I0227 18:30:00.527012 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536950-ln6wp" Feb 27 18:30:00 crc kubenswrapper[4700]: I0227 18:30:00.540650 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536950-458j5" Feb 27 18:30:00 crc kubenswrapper[4700]: E0227 18:30:00.724924 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 27 18:30:00 crc kubenswrapper[4700]: E0227 18:30:00.725332 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-plrxr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-h89qg_openshift-marketplace(b0fd7b6d-8f71-4bff-b903-cdb5a0954f87): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 18:30:00 crc kubenswrapper[4700]: E0227 18:30:00.726557 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/community-operators-h89qg" podUID="b0fd7b6d-8f71-4bff-b903-cdb5a0954f87" Feb 27 18:30:01 crc kubenswrapper[4700]: I0227 18:30:01.037912 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536950-458j5"] Feb 27 18:30:01 crc kubenswrapper[4700]: W0227 18:30:01.040769 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod74489fab_6aca_4a3f_9303_04a2b82f1fde.slice/crio-374b2b0ff5c6a0f7e6b8cc63d2329fce6c7fd391ab2adf0893e3afa3704a7d5c WatchSource:0}: Error finding container 374b2b0ff5c6a0f7e6b8cc63d2329fce6c7fd391ab2adf0893e3afa3704a7d5c: Status 404 returned error can't find the container with id 374b2b0ff5c6a0f7e6b8cc63d2329fce6c7fd391ab2adf0893e3afa3704a7d5c Feb 27 18:30:01 crc kubenswrapper[4700]: I0227 18:30:01.065488 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536950-458j5" event={"ID":"74489fab-6aca-4a3f-9303-04a2b82f1fde","Type":"ContainerStarted","Data":"374b2b0ff5c6a0f7e6b8cc63d2329fce6c7fd391ab2adf0893e3afa3704a7d5c"} Feb 27 18:30:01 crc kubenswrapper[4700]: E0227 18:30:01.068893 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-h89qg" podUID="b0fd7b6d-8f71-4bff-b903-cdb5a0954f87" Feb 27 18:30:01 crc kubenswrapper[4700]: I0227 18:30:01.114441 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536950-ln6wp"] Feb 27 18:30:01 crc kubenswrapper[4700]: E0227 18:30:01.921844 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 18:30:01 crc kubenswrapper[4700]: E0227 18:30:01.922779 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 18:30:01 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 18:30:01 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hgn64,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536950-458j5_openshift-infra(74489fab-6aca-4a3f-9303-04a2b82f1fde): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 18:30:01 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 18:30:01 crc kubenswrapper[4700]: E0227 18:30:01.924388 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536950-458j5" podUID="74489fab-6aca-4a3f-9303-04a2b82f1fde" Feb 27 18:30:02 crc kubenswrapper[4700]: I0227 18:30:02.086657 4700 generic.go:334] "Generic (PLEG): container finished" podID="050b88a9-b342-4990-9bac-17c98b014f49" containerID="ddbe09800d68adf7131a739be50875492d0351336a8bbf13d88ca9f58b872261" exitCode=0 Feb 27 18:30:02 crc kubenswrapper[4700]: I0227 18:30:02.086738 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536950-ln6wp" event={"ID":"050b88a9-b342-4990-9bac-17c98b014f49","Type":"ContainerDied","Data":"ddbe09800d68adf7131a739be50875492d0351336a8bbf13d88ca9f58b872261"} Feb 27 18:30:02 crc kubenswrapper[4700]: I0227 18:30:02.086826 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536950-ln6wp" event={"ID":"050b88a9-b342-4990-9bac-17c98b014f49","Type":"ContainerStarted","Data":"a4e802a3bf666a1e4c3c3b7f1d79e2faead6a57c78917cb456efeabea0c8fe87"} Feb 27 18:30:02 crc kubenswrapper[4700]: E0227 18:30:02.093134 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536950-458j5" podUID="74489fab-6aca-4a3f-9303-04a2b82f1fde" Feb 27 18:30:03 crc kubenswrapper[4700]: I0227 18:30:03.666509 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536950-ln6wp" Feb 27 18:30:03 crc kubenswrapper[4700]: I0227 18:30:03.705817 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/050b88a9-b342-4990-9bac-17c98b014f49-config-volume\") pod \"050b88a9-b342-4990-9bac-17c98b014f49\" (UID: \"050b88a9-b342-4990-9bac-17c98b014f49\") " Feb 27 18:30:03 crc kubenswrapper[4700]: I0227 18:30:03.706279 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/050b88a9-b342-4990-9bac-17c98b014f49-secret-volume\") pod \"050b88a9-b342-4990-9bac-17c98b014f49\" (UID: \"050b88a9-b342-4990-9bac-17c98b014f49\") " Feb 27 18:30:03 crc kubenswrapper[4700]: I0227 18:30:03.706410 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftprg\" (UniqueName: \"kubernetes.io/projected/050b88a9-b342-4990-9bac-17c98b014f49-kube-api-access-ftprg\") pod \"050b88a9-b342-4990-9bac-17c98b014f49\" (UID: \"050b88a9-b342-4990-9bac-17c98b014f49\") " Feb 27 18:30:03 crc kubenswrapper[4700]: I0227 18:30:03.707906 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/050b88a9-b342-4990-9bac-17c98b014f49-config-volume" (OuterVolumeSpecName: "config-volume") pod "050b88a9-b342-4990-9bac-17c98b014f49" (UID: "050b88a9-b342-4990-9bac-17c98b014f49"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 18:30:03 crc kubenswrapper[4700]: I0227 18:30:03.717759 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/050b88a9-b342-4990-9bac-17c98b014f49-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "050b88a9-b342-4990-9bac-17c98b014f49" (UID: "050b88a9-b342-4990-9bac-17c98b014f49"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 18:30:03 crc kubenswrapper[4700]: I0227 18:30:03.723734 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/050b88a9-b342-4990-9bac-17c98b014f49-kube-api-access-ftprg" (OuterVolumeSpecName: "kube-api-access-ftprg") pod "050b88a9-b342-4990-9bac-17c98b014f49" (UID: "050b88a9-b342-4990-9bac-17c98b014f49"). InnerVolumeSpecName "kube-api-access-ftprg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:30:03 crc kubenswrapper[4700]: I0227 18:30:03.811110 4700 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/050b88a9-b342-4990-9bac-17c98b014f49-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 27 18:30:03 crc kubenswrapper[4700]: I0227 18:30:03.811183 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ftprg\" (UniqueName: \"kubernetes.io/projected/050b88a9-b342-4990-9bac-17c98b014f49-kube-api-access-ftprg\") on node \"crc\" DevicePath \"\"" Feb 27 18:30:03 crc kubenswrapper[4700]: I0227 18:30:03.811207 4700 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/050b88a9-b342-4990-9bac-17c98b014f49-config-volume\") on node \"crc\" DevicePath \"\"" Feb 27 18:30:04 crc kubenswrapper[4700]: I0227 18:30:04.116130 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536950-ln6wp" event={"ID":"050b88a9-b342-4990-9bac-17c98b014f49","Type":"ContainerDied","Data":"a4e802a3bf666a1e4c3c3b7f1d79e2faead6a57c78917cb456efeabea0c8fe87"} Feb 27 18:30:04 crc kubenswrapper[4700]: I0227 18:30:04.116195 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4e802a3bf666a1e4c3c3b7f1d79e2faead6a57c78917cb456efeabea0c8fe87" Feb 27 18:30:04 crc kubenswrapper[4700]: I0227 18:30:04.116212 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536950-ln6wp" Feb 27 18:30:04 crc kubenswrapper[4700]: I0227 18:30:04.807713 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536905-cchlk"] Feb 27 18:30:04 crc kubenswrapper[4700]: I0227 18:30:04.819204 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536905-cchlk"] Feb 27 18:30:05 crc kubenswrapper[4700]: I0227 18:30:05.002178 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86fc5416-1e17-405a-a34d-49965c235b33" path="/var/lib/kubelet/pods/86fc5416-1e17-405a-a34d-49965c235b33/volumes" Feb 27 18:30:06 crc kubenswrapper[4700]: E0227 18:30:06.156218 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 18:30:06 crc kubenswrapper[4700]: E0227 18:30:06.156440 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 18:30:06 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 18:30:06 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qkjfz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536934-rw77f_openshift-infra(caa4e57d-b02d-4d65-944d-451b8294f884): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 18:30:06 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 18:30:06 crc kubenswrapper[4700]: E0227 18:30:06.157838 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:30:07 crc kubenswrapper[4700]: I0227 18:30:07.692808 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lrt8r" podUID="98c1faf7-b2cc-47c3-89e3-20e563b9f507" containerName="registry-server" probeResult="failure" output=< Feb 27 18:30:07 crc kubenswrapper[4700]: timeout: failed to connect service ":50051" within 1s Feb 27 18:30:07 crc kubenswrapper[4700]: > Feb 27 18:30:07 crc kubenswrapper[4700]: I0227 18:30:07.983814 4700 scope.go:117] "RemoveContainer" containerID="b3fdfb06db2c06428e3b3e96ef9804be97292a50a8f04f891caec5e729731adf" Feb 27 18:30:07 crc kubenswrapper[4700]: E0227 18:30:07.984514 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:30:13 crc kubenswrapper[4700]: I0227 18:30:13.503822 4700 scope.go:117] "RemoveContainer" containerID="371595c7840fd50eadd739ae642603d40a0efa929815ad8c6a687fc0d21b0dc7" Feb 27 18:30:15 crc kubenswrapper[4700]: I0227 18:30:15.263134 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536950-458j5" event={"ID":"74489fab-6aca-4a3f-9303-04a2b82f1fde","Type":"ContainerStarted","Data":"4a54bd2797b5714d8a57bcd263accf252512d3e44013a6579b8785a2cc66e4d6"} Feb 27 18:30:15 crc kubenswrapper[4700]: I0227 18:30:15.291836 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536950-458j5" podStartSLOduration=1.456864511 podStartE2EDuration="15.291814623s" podCreationTimestamp="2026-02-27 18:30:00 +0000 UTC" firstStartedPulling="2026-02-27 18:30:01.04417332 +0000 UTC m=+5361.029486077" lastFinishedPulling="2026-02-27 18:30:14.879123432 +0000 UTC m=+5374.864436189" observedRunningTime="2026-02-27 18:30:15.281041482 +0000 UTC m=+5375.266354229" watchObservedRunningTime="2026-02-27 18:30:15.291814623 +0000 UTC m=+5375.277127380" Feb 27 18:30:16 crc kubenswrapper[4700]: I0227 18:30:16.276971 4700 generic.go:334] "Generic (PLEG): container finished" podID="74489fab-6aca-4a3f-9303-04a2b82f1fde" containerID="4a54bd2797b5714d8a57bcd263accf252512d3e44013a6579b8785a2cc66e4d6" exitCode=0 Feb 27 18:30:16 crc kubenswrapper[4700]: I0227 18:30:16.277045 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536950-458j5" event={"ID":"74489fab-6aca-4a3f-9303-04a2b82f1fde","Type":"ContainerDied","Data":"4a54bd2797b5714d8a57bcd263accf252512d3e44013a6579b8785a2cc66e4d6"} Feb 27 18:30:16 crc kubenswrapper[4700]: E0227 18:30:16.828270 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 27 18:30:16 crc kubenswrapper[4700]: E0227 18:30:16.829155 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-plrxr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-h89qg_openshift-marketplace(b0fd7b6d-8f71-4bff-b903-cdb5a0954f87): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 18:30:16 crc kubenswrapper[4700]: E0227 18:30:16.830526 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/community-operators-h89qg" podUID="b0fd7b6d-8f71-4bff-b903-cdb5a0954f87" Feb 27 18:30:17 crc kubenswrapper[4700]: I0227 18:30:17.697688 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lrt8r" podUID="98c1faf7-b2cc-47c3-89e3-20e563b9f507" containerName="registry-server" probeResult="failure" output=< Feb 27 18:30:17 crc kubenswrapper[4700]: timeout: failed to connect service ":50051" within 1s Feb 27 18:30:17 crc kubenswrapper[4700]: > Feb 27 18:30:17 crc kubenswrapper[4700]: I0227 18:30:17.705714 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536950-458j5" Feb 27 18:30:17 crc kubenswrapper[4700]: I0227 18:30:17.748073 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgn64\" (UniqueName: \"kubernetes.io/projected/74489fab-6aca-4a3f-9303-04a2b82f1fde-kube-api-access-hgn64\") pod \"74489fab-6aca-4a3f-9303-04a2b82f1fde\" (UID: \"74489fab-6aca-4a3f-9303-04a2b82f1fde\") " Feb 27 18:30:17 crc kubenswrapper[4700]: I0227 18:30:17.755002 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74489fab-6aca-4a3f-9303-04a2b82f1fde-kube-api-access-hgn64" (OuterVolumeSpecName: "kube-api-access-hgn64") pod "74489fab-6aca-4a3f-9303-04a2b82f1fde" (UID: "74489fab-6aca-4a3f-9303-04a2b82f1fde"). InnerVolumeSpecName "kube-api-access-hgn64". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:30:17 crc kubenswrapper[4700]: I0227 18:30:17.850299 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgn64\" (UniqueName: \"kubernetes.io/projected/74489fab-6aca-4a3f-9303-04a2b82f1fde-kube-api-access-hgn64\") on node \"crc\" DevicePath \"\"" Feb 27 18:30:17 crc kubenswrapper[4700]: E0227 18:30:17.983857 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:30:18 crc kubenswrapper[4700]: I0227 18:30:18.305655 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536950-458j5" event={"ID":"74489fab-6aca-4a3f-9303-04a2b82f1fde","Type":"ContainerDied","Data":"374b2b0ff5c6a0f7e6b8cc63d2329fce6c7fd391ab2adf0893e3afa3704a7d5c"} Feb 27 18:30:18 crc kubenswrapper[4700]: I0227 18:30:18.305698 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="374b2b0ff5c6a0f7e6b8cc63d2329fce6c7fd391ab2adf0893e3afa3704a7d5c" Feb 27 18:30:18 crc kubenswrapper[4700]: I0227 18:30:18.306122 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536950-458j5" Feb 27 18:30:18 crc kubenswrapper[4700]: I0227 18:30:18.383916 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536944-clxj9"] Feb 27 18:30:18 crc kubenswrapper[4700]: I0227 18:30:18.398304 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536944-clxj9"] Feb 27 18:30:18 crc kubenswrapper[4700]: I0227 18:30:18.991174 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f757b70-0256-4500-8d34-368ff32dc545" path="/var/lib/kubelet/pods/9f757b70-0256-4500-8d34-368ff32dc545/volumes" Feb 27 18:30:19 crc kubenswrapper[4700]: I0227 18:30:19.982444 4700 scope.go:117] "RemoveContainer" containerID="b3fdfb06db2c06428e3b3e96ef9804be97292a50a8f04f891caec5e729731adf" Feb 27 18:30:19 crc kubenswrapper[4700]: E0227 18:30:19.983317 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:30:26 crc kubenswrapper[4700]: I0227 18:30:26.692804 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lrt8r" Feb 27 18:30:26 crc kubenswrapper[4700]: I0227 18:30:26.784413 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lrt8r" Feb 27 18:30:26 crc kubenswrapper[4700]: I0227 18:30:26.942899 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lrt8r"] Feb 27 18:30:28 crc kubenswrapper[4700]: I0227 18:30:28.430014 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lrt8r" podUID="98c1faf7-b2cc-47c3-89e3-20e563b9f507" containerName="registry-server" containerID="cri-o://c63aeace47d8f492efd296c51962ea3bd9afdfc0a9cc49085c24e07c443f768c" gracePeriod=2 Feb 27 18:30:29 crc kubenswrapper[4700]: I0227 18:30:29.049284 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lrt8r" Feb 27 18:30:29 crc kubenswrapper[4700]: I0227 18:30:29.187935 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98c1faf7-b2cc-47c3-89e3-20e563b9f507-catalog-content\") pod \"98c1faf7-b2cc-47c3-89e3-20e563b9f507\" (UID: \"98c1faf7-b2cc-47c3-89e3-20e563b9f507\") " Feb 27 18:30:29 crc kubenswrapper[4700]: I0227 18:30:29.188701 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98c1faf7-b2cc-47c3-89e3-20e563b9f507-utilities\") pod \"98c1faf7-b2cc-47c3-89e3-20e563b9f507\" (UID: \"98c1faf7-b2cc-47c3-89e3-20e563b9f507\") " Feb 27 18:30:29 crc kubenswrapper[4700]: I0227 18:30:29.188954 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6nwmt\" (UniqueName: \"kubernetes.io/projected/98c1faf7-b2cc-47c3-89e3-20e563b9f507-kube-api-access-6nwmt\") pod \"98c1faf7-b2cc-47c3-89e3-20e563b9f507\" (UID: \"98c1faf7-b2cc-47c3-89e3-20e563b9f507\") " Feb 27 18:30:29 crc kubenswrapper[4700]: I0227 18:30:29.189791 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98c1faf7-b2cc-47c3-89e3-20e563b9f507-utilities" (OuterVolumeSpecName: "utilities") pod "98c1faf7-b2cc-47c3-89e3-20e563b9f507" (UID: "98c1faf7-b2cc-47c3-89e3-20e563b9f507"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 18:30:29 crc kubenswrapper[4700]: I0227 18:30:29.199766 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98c1faf7-b2cc-47c3-89e3-20e563b9f507-kube-api-access-6nwmt" (OuterVolumeSpecName: "kube-api-access-6nwmt") pod "98c1faf7-b2cc-47c3-89e3-20e563b9f507" (UID: "98c1faf7-b2cc-47c3-89e3-20e563b9f507"). InnerVolumeSpecName "kube-api-access-6nwmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:30:29 crc kubenswrapper[4700]: I0227 18:30:29.291959 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98c1faf7-b2cc-47c3-89e3-20e563b9f507-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 18:30:29 crc kubenswrapper[4700]: I0227 18:30:29.291991 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6nwmt\" (UniqueName: \"kubernetes.io/projected/98c1faf7-b2cc-47c3-89e3-20e563b9f507-kube-api-access-6nwmt\") on node \"crc\" DevicePath \"\"" Feb 27 18:30:29 crc kubenswrapper[4700]: I0227 18:30:29.382328 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98c1faf7-b2cc-47c3-89e3-20e563b9f507-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "98c1faf7-b2cc-47c3-89e3-20e563b9f507" (UID: "98c1faf7-b2cc-47c3-89e3-20e563b9f507"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 18:30:29 crc kubenswrapper[4700]: I0227 18:30:29.393991 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98c1faf7-b2cc-47c3-89e3-20e563b9f507-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 18:30:29 crc kubenswrapper[4700]: I0227 18:30:29.441836 4700 generic.go:334] "Generic (PLEG): container finished" podID="98c1faf7-b2cc-47c3-89e3-20e563b9f507" containerID="c63aeace47d8f492efd296c51962ea3bd9afdfc0a9cc49085c24e07c443f768c" exitCode=0 Feb 27 18:30:29 crc kubenswrapper[4700]: I0227 18:30:29.441877 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrt8r" event={"ID":"98c1faf7-b2cc-47c3-89e3-20e563b9f507","Type":"ContainerDied","Data":"c63aeace47d8f492efd296c51962ea3bd9afdfc0a9cc49085c24e07c443f768c"} Feb 27 18:30:29 crc kubenswrapper[4700]: I0227 18:30:29.441903 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrt8r" event={"ID":"98c1faf7-b2cc-47c3-89e3-20e563b9f507","Type":"ContainerDied","Data":"d0111d8397a2c72d15144e0180d3ac0d423eaafafe1d957985a397bf6212f384"} Feb 27 18:30:29 crc kubenswrapper[4700]: I0227 18:30:29.441903 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lrt8r" Feb 27 18:30:29 crc kubenswrapper[4700]: I0227 18:30:29.441988 4700 scope.go:117] "RemoveContainer" containerID="c63aeace47d8f492efd296c51962ea3bd9afdfc0a9cc49085c24e07c443f768c" Feb 27 18:30:29 crc kubenswrapper[4700]: I0227 18:30:29.484752 4700 scope.go:117] "RemoveContainer" containerID="2437c89416ad507ad89800fdeef354efa0de5dd6a65d8879874af48e5b58f3a1" Feb 27 18:30:29 crc kubenswrapper[4700]: I0227 18:30:29.506524 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lrt8r"] Feb 27 18:30:29 crc kubenswrapper[4700]: I0227 18:30:29.519031 4700 scope.go:117] "RemoveContainer" containerID="2872e40bc161207c25f7614f9528cd7c88871edac6c66c0f2de1f6cb2a3ae789" Feb 27 18:30:29 crc kubenswrapper[4700]: I0227 18:30:29.529864 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lrt8r"] Feb 27 18:30:29 crc kubenswrapper[4700]: I0227 18:30:29.587165 4700 scope.go:117] "RemoveContainer" containerID="c63aeace47d8f492efd296c51962ea3bd9afdfc0a9cc49085c24e07c443f768c" Feb 27 18:30:29 crc kubenswrapper[4700]: E0227 18:30:29.587711 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c63aeace47d8f492efd296c51962ea3bd9afdfc0a9cc49085c24e07c443f768c\": container with ID starting with c63aeace47d8f492efd296c51962ea3bd9afdfc0a9cc49085c24e07c443f768c not found: ID does not exist" containerID="c63aeace47d8f492efd296c51962ea3bd9afdfc0a9cc49085c24e07c443f768c" Feb 27 18:30:29 crc kubenswrapper[4700]: I0227 18:30:29.587776 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c63aeace47d8f492efd296c51962ea3bd9afdfc0a9cc49085c24e07c443f768c"} err="failed to get container status \"c63aeace47d8f492efd296c51962ea3bd9afdfc0a9cc49085c24e07c443f768c\": rpc error: code = NotFound desc = could not find container \"c63aeace47d8f492efd296c51962ea3bd9afdfc0a9cc49085c24e07c443f768c\": container with ID starting with c63aeace47d8f492efd296c51962ea3bd9afdfc0a9cc49085c24e07c443f768c not found: ID does not exist" Feb 27 18:30:29 crc kubenswrapper[4700]: I0227 18:30:29.587817 4700 scope.go:117] "RemoveContainer" containerID="2437c89416ad507ad89800fdeef354efa0de5dd6a65d8879874af48e5b58f3a1" Feb 27 18:30:29 crc kubenswrapper[4700]: E0227 18:30:29.588352 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2437c89416ad507ad89800fdeef354efa0de5dd6a65d8879874af48e5b58f3a1\": container with ID starting with 2437c89416ad507ad89800fdeef354efa0de5dd6a65d8879874af48e5b58f3a1 not found: ID does not exist" containerID="2437c89416ad507ad89800fdeef354efa0de5dd6a65d8879874af48e5b58f3a1" Feb 27 18:30:29 crc kubenswrapper[4700]: I0227 18:30:29.588409 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2437c89416ad507ad89800fdeef354efa0de5dd6a65d8879874af48e5b58f3a1"} err="failed to get container status \"2437c89416ad507ad89800fdeef354efa0de5dd6a65d8879874af48e5b58f3a1\": rpc error: code = NotFound desc = could not find container \"2437c89416ad507ad89800fdeef354efa0de5dd6a65d8879874af48e5b58f3a1\": container with ID starting with 2437c89416ad507ad89800fdeef354efa0de5dd6a65d8879874af48e5b58f3a1 not found: ID does not exist" Feb 27 18:30:29 crc kubenswrapper[4700]: I0227 18:30:29.588457 4700 scope.go:117] "RemoveContainer" containerID="2872e40bc161207c25f7614f9528cd7c88871edac6c66c0f2de1f6cb2a3ae789" Feb 27 18:30:29 crc kubenswrapper[4700]: E0227 18:30:29.589266 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2872e40bc161207c25f7614f9528cd7c88871edac6c66c0f2de1f6cb2a3ae789\": container with ID starting with 2872e40bc161207c25f7614f9528cd7c88871edac6c66c0f2de1f6cb2a3ae789 not found: ID does not exist" containerID="2872e40bc161207c25f7614f9528cd7c88871edac6c66c0f2de1f6cb2a3ae789" Feb 27 18:30:29 crc kubenswrapper[4700]: I0227 18:30:29.589363 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2872e40bc161207c25f7614f9528cd7c88871edac6c66c0f2de1f6cb2a3ae789"} err="failed to get container status \"2872e40bc161207c25f7614f9528cd7c88871edac6c66c0f2de1f6cb2a3ae789\": rpc error: code = NotFound desc = could not find container \"2872e40bc161207c25f7614f9528cd7c88871edac6c66c0f2de1f6cb2a3ae789\": container with ID starting with 2872e40bc161207c25f7614f9528cd7c88871edac6c66c0f2de1f6cb2a3ae789 not found: ID does not exist" Feb 27 18:30:29 crc kubenswrapper[4700]: E0227 18:30:29.984635 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:30:31 crc kubenswrapper[4700]: E0227 18:30:31.001997 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-h89qg" podUID="b0fd7b6d-8f71-4bff-b903-cdb5a0954f87" Feb 27 18:30:31 crc kubenswrapper[4700]: I0227 18:30:31.011154 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98c1faf7-b2cc-47c3-89e3-20e563b9f507" path="/var/lib/kubelet/pods/98c1faf7-b2cc-47c3-89e3-20e563b9f507/volumes" Feb 27 18:30:33 crc kubenswrapper[4700]: I0227 18:30:33.981202 4700 scope.go:117] "RemoveContainer" containerID="b3fdfb06db2c06428e3b3e96ef9804be97292a50a8f04f891caec5e729731adf" Feb 27 18:30:33 crc kubenswrapper[4700]: E0227 18:30:33.982238 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:30:42 crc kubenswrapper[4700]: E0227 18:30:42.984540 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:30:44 crc kubenswrapper[4700]: E0227 18:30:44.891580 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 27 18:30:44 crc kubenswrapper[4700]: E0227 18:30:44.892082 4700 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-plrxr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-h89qg_openshift-marketplace(b0fd7b6d-8f71-4bff-b903-cdb5a0954f87): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)" logger="UnhandledError" Feb 27 18:30:44 crc kubenswrapper[4700]: E0227 18:30:44.893273 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/redhat/community-operator-index@sha256=886ecdbcb5b8f90338063f6476072fab73c2a9a65b9f2b3835b7bd01c69794c1/signature-2: status 500 (Internal Server Error)\"" pod="openshift-marketplace/community-operators-h89qg" podUID="b0fd7b6d-8f71-4bff-b903-cdb5a0954f87" Feb 27 18:30:47 crc kubenswrapper[4700]: I0227 18:30:47.981948 4700 scope.go:117] "RemoveContainer" containerID="b3fdfb06db2c06428e3b3e96ef9804be97292a50a8f04f891caec5e729731adf" Feb 27 18:30:47 crc kubenswrapper[4700]: E0227 18:30:47.983048 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:30:56 crc kubenswrapper[4700]: E0227 18:30:56.986379 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-h89qg" podUID="b0fd7b6d-8f71-4bff-b903-cdb5a0954f87" Feb 27 18:30:57 crc kubenswrapper[4700]: E0227 18:30:57.983733 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:30:58 crc kubenswrapper[4700]: I0227 18:30:58.980975 4700 scope.go:117] "RemoveContainer" containerID="b3fdfb06db2c06428e3b3e96ef9804be97292a50a8f04f891caec5e729731adf" Feb 27 18:30:58 crc kubenswrapper[4700]: E0227 18:30:58.981627 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:31:08 crc kubenswrapper[4700]: E0227 18:31:08.983942 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-h89qg" podUID="b0fd7b6d-8f71-4bff-b903-cdb5a0954f87" Feb 27 18:31:11 crc kubenswrapper[4700]: E0227 18:31:11.984850 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:31:13 crc kubenswrapper[4700]: I0227 18:31:13.595555 4700 scope.go:117] "RemoveContainer" containerID="472b00220db6e77f0fc32d292654f8847eab10cc1d098ab50e4aaa0293fb1147" Feb 27 18:31:13 crc kubenswrapper[4700]: I0227 18:31:13.981906 4700 scope.go:117] "RemoveContainer" containerID="b3fdfb06db2c06428e3b3e96ef9804be97292a50a8f04f891caec5e729731adf" Feb 27 18:31:15 crc kubenswrapper[4700]: I0227 18:31:15.090731 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerStarted","Data":"fc0ec093303e51bd2ed80860af9a81018e732a7449b44faf6a9d28fb0570f2b0"} Feb 27 18:31:21 crc kubenswrapper[4700]: E0227 18:31:21.052356 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-h89qg" podUID="b0fd7b6d-8f71-4bff-b903-cdb5a0954f87" Feb 27 18:31:26 crc kubenswrapper[4700]: E0227 18:31:26.985057 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:31:36 crc kubenswrapper[4700]: I0227 18:31:36.397610 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h89qg" event={"ID":"b0fd7b6d-8f71-4bff-b903-cdb5a0954f87","Type":"ContainerStarted","Data":"e97e47712fd288531d516f11e1fab47375c1ced8eed9dbaeced42db991965e94"} Feb 27 18:31:37 crc kubenswrapper[4700]: I0227 18:31:37.412907 4700 generic.go:334] "Generic (PLEG): container finished" podID="b0fd7b6d-8f71-4bff-b903-cdb5a0954f87" containerID="e97e47712fd288531d516f11e1fab47375c1ced8eed9dbaeced42db991965e94" exitCode=0 Feb 27 18:31:37 crc kubenswrapper[4700]: I0227 18:31:37.413023 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h89qg" event={"ID":"b0fd7b6d-8f71-4bff-b903-cdb5a0954f87","Type":"ContainerDied","Data":"e97e47712fd288531d516f11e1fab47375c1ced8eed9dbaeced42db991965e94"} Feb 27 18:31:38 crc kubenswrapper[4700]: I0227 18:31:38.429392 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h89qg" event={"ID":"b0fd7b6d-8f71-4bff-b903-cdb5a0954f87","Type":"ContainerStarted","Data":"faaf4cbdb4e82be8acc0dcc2d18d73b76850bb2af9b860be175b40c6dc293658"} Feb 27 18:31:38 crc kubenswrapper[4700]: I0227 18:31:38.466352 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-h89qg" podStartSLOduration=2.695903587 podStartE2EDuration="1m40.46631787s" podCreationTimestamp="2026-02-27 18:29:58 +0000 UTC" firstStartedPulling="2026-02-27 18:30:00.054802196 +0000 UTC m=+5360.040114943" lastFinishedPulling="2026-02-27 18:31:37.825216459 +0000 UTC m=+5457.810529226" observedRunningTime="2026-02-27 18:31:38.45361969 +0000 UTC m=+5458.438932447" watchObservedRunningTime="2026-02-27 18:31:38.46631787 +0000 UTC m=+5458.451630657" Feb 27 18:31:38 crc kubenswrapper[4700]: I0227 18:31:38.830479 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-h89qg" Feb 27 18:31:38 crc kubenswrapper[4700]: I0227 18:31:38.830572 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-h89qg" Feb 27 18:31:38 crc kubenswrapper[4700]: E0227 18:31:38.982674 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:31:39 crc kubenswrapper[4700]: I0227 18:31:39.917910 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-h89qg" podUID="b0fd7b6d-8f71-4bff-b903-cdb5a0954f87" containerName="registry-server" probeResult="failure" output=< Feb 27 18:31:39 crc kubenswrapper[4700]: timeout: failed to connect service ":50051" within 1s Feb 27 18:31:39 crc kubenswrapper[4700]: > Feb 27 18:31:48 crc kubenswrapper[4700]: I0227 18:31:48.897888 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-h89qg" Feb 27 18:31:48 crc kubenswrapper[4700]: I0227 18:31:48.976925 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-h89qg" Feb 27 18:31:49 crc kubenswrapper[4700]: I0227 18:31:49.158591 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h89qg"] Feb 27 18:31:50 crc kubenswrapper[4700]: I0227 18:31:50.574923 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-h89qg" podUID="b0fd7b6d-8f71-4bff-b903-cdb5a0954f87" containerName="registry-server" containerID="cri-o://faaf4cbdb4e82be8acc0dcc2d18d73b76850bb2af9b860be175b40c6dc293658" gracePeriod=2 Feb 27 18:31:50 crc kubenswrapper[4700]: E0227 18:31:50.991603 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:31:51 crc kubenswrapper[4700]: I0227 18:31:51.124676 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h89qg" Feb 27 18:31:51 crc kubenswrapper[4700]: I0227 18:31:51.227746 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0fd7b6d-8f71-4bff-b903-cdb5a0954f87-catalog-content\") pod \"b0fd7b6d-8f71-4bff-b903-cdb5a0954f87\" (UID: \"b0fd7b6d-8f71-4bff-b903-cdb5a0954f87\") " Feb 27 18:31:51 crc kubenswrapper[4700]: I0227 18:31:51.227980 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-plrxr\" (UniqueName: \"kubernetes.io/projected/b0fd7b6d-8f71-4bff-b903-cdb5a0954f87-kube-api-access-plrxr\") pod \"b0fd7b6d-8f71-4bff-b903-cdb5a0954f87\" (UID: \"b0fd7b6d-8f71-4bff-b903-cdb5a0954f87\") " Feb 27 18:31:51 crc kubenswrapper[4700]: I0227 18:31:51.228102 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0fd7b6d-8f71-4bff-b903-cdb5a0954f87-utilities\") pod \"b0fd7b6d-8f71-4bff-b903-cdb5a0954f87\" (UID: \"b0fd7b6d-8f71-4bff-b903-cdb5a0954f87\") " Feb 27 18:31:51 crc kubenswrapper[4700]: I0227 18:31:51.229525 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0fd7b6d-8f71-4bff-b903-cdb5a0954f87-utilities" (OuterVolumeSpecName: "utilities") pod "b0fd7b6d-8f71-4bff-b903-cdb5a0954f87" (UID: "b0fd7b6d-8f71-4bff-b903-cdb5a0954f87"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 18:31:51 crc kubenswrapper[4700]: I0227 18:31:51.239737 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0fd7b6d-8f71-4bff-b903-cdb5a0954f87-kube-api-access-plrxr" (OuterVolumeSpecName: "kube-api-access-plrxr") pod "b0fd7b6d-8f71-4bff-b903-cdb5a0954f87" (UID: "b0fd7b6d-8f71-4bff-b903-cdb5a0954f87"). InnerVolumeSpecName "kube-api-access-plrxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:31:51 crc kubenswrapper[4700]: I0227 18:31:51.292085 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0fd7b6d-8f71-4bff-b903-cdb5a0954f87-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b0fd7b6d-8f71-4bff-b903-cdb5a0954f87" (UID: "b0fd7b6d-8f71-4bff-b903-cdb5a0954f87"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 18:31:51 crc kubenswrapper[4700]: I0227 18:31:51.330734 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-plrxr\" (UniqueName: \"kubernetes.io/projected/b0fd7b6d-8f71-4bff-b903-cdb5a0954f87-kube-api-access-plrxr\") on node \"crc\" DevicePath \"\"" Feb 27 18:31:51 crc kubenswrapper[4700]: I0227 18:31:51.330778 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0fd7b6d-8f71-4bff-b903-cdb5a0954f87-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 18:31:51 crc kubenswrapper[4700]: I0227 18:31:51.330791 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0fd7b6d-8f71-4bff-b903-cdb5a0954f87-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 18:31:51 crc kubenswrapper[4700]: I0227 18:31:51.585824 4700 generic.go:334] "Generic (PLEG): container finished" podID="b0fd7b6d-8f71-4bff-b903-cdb5a0954f87" containerID="faaf4cbdb4e82be8acc0dcc2d18d73b76850bb2af9b860be175b40c6dc293658" exitCode=0 Feb 27 18:31:51 crc kubenswrapper[4700]: I0227 18:31:51.585875 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h89qg" event={"ID":"b0fd7b6d-8f71-4bff-b903-cdb5a0954f87","Type":"ContainerDied","Data":"faaf4cbdb4e82be8acc0dcc2d18d73b76850bb2af9b860be175b40c6dc293658"} Feb 27 18:31:51 crc kubenswrapper[4700]: I0227 18:31:51.585903 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h89qg" event={"ID":"b0fd7b6d-8f71-4bff-b903-cdb5a0954f87","Type":"ContainerDied","Data":"91f4ab7e9143d9c6e25fbca874f1eb387cb0d59c6d36e92fc5ee17a09681ea82"} Feb 27 18:31:51 crc kubenswrapper[4700]: I0227 18:31:51.585907 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h89qg" Feb 27 18:31:51 crc kubenswrapper[4700]: I0227 18:31:51.585923 4700 scope.go:117] "RemoveContainer" containerID="faaf4cbdb4e82be8acc0dcc2d18d73b76850bb2af9b860be175b40c6dc293658" Feb 27 18:31:51 crc kubenswrapper[4700]: I0227 18:31:51.632682 4700 scope.go:117] "RemoveContainer" containerID="e97e47712fd288531d516f11e1fab47375c1ced8eed9dbaeced42db991965e94" Feb 27 18:31:51 crc kubenswrapper[4700]: I0227 18:31:51.634544 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h89qg"] Feb 27 18:31:51 crc kubenswrapper[4700]: I0227 18:31:51.648330 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-h89qg"] Feb 27 18:31:51 crc kubenswrapper[4700]: I0227 18:31:51.679730 4700 scope.go:117] "RemoveContainer" containerID="4b65a79cf2eb03a8dd97bcdd54a620d09775cde1106b8a7b7fff34d53100a0b9" Feb 27 18:31:51 crc kubenswrapper[4700]: I0227 18:31:51.738943 4700 scope.go:117] "RemoveContainer" containerID="faaf4cbdb4e82be8acc0dcc2d18d73b76850bb2af9b860be175b40c6dc293658" Feb 27 18:31:51 crc kubenswrapper[4700]: E0227 18:31:51.739642 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"faaf4cbdb4e82be8acc0dcc2d18d73b76850bb2af9b860be175b40c6dc293658\": container with ID starting with faaf4cbdb4e82be8acc0dcc2d18d73b76850bb2af9b860be175b40c6dc293658 not found: ID does not exist" containerID="faaf4cbdb4e82be8acc0dcc2d18d73b76850bb2af9b860be175b40c6dc293658" Feb 27 18:31:51 crc kubenswrapper[4700]: I0227 18:31:51.739772 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"faaf4cbdb4e82be8acc0dcc2d18d73b76850bb2af9b860be175b40c6dc293658"} err="failed to get container status \"faaf4cbdb4e82be8acc0dcc2d18d73b76850bb2af9b860be175b40c6dc293658\": rpc error: code = NotFound desc = could not find container \"faaf4cbdb4e82be8acc0dcc2d18d73b76850bb2af9b860be175b40c6dc293658\": container with ID starting with faaf4cbdb4e82be8acc0dcc2d18d73b76850bb2af9b860be175b40c6dc293658 not found: ID does not exist" Feb 27 18:31:51 crc kubenswrapper[4700]: I0227 18:31:51.739885 4700 scope.go:117] "RemoveContainer" containerID="e97e47712fd288531d516f11e1fab47375c1ced8eed9dbaeced42db991965e94" Feb 27 18:31:51 crc kubenswrapper[4700]: E0227 18:31:51.740508 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e97e47712fd288531d516f11e1fab47375c1ced8eed9dbaeced42db991965e94\": container with ID starting with e97e47712fd288531d516f11e1fab47375c1ced8eed9dbaeced42db991965e94 not found: ID does not exist" containerID="e97e47712fd288531d516f11e1fab47375c1ced8eed9dbaeced42db991965e94" Feb 27 18:31:51 crc kubenswrapper[4700]: I0227 18:31:51.741109 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e97e47712fd288531d516f11e1fab47375c1ced8eed9dbaeced42db991965e94"} err="failed to get container status \"e97e47712fd288531d516f11e1fab47375c1ced8eed9dbaeced42db991965e94\": rpc error: code = NotFound desc = could not find container \"e97e47712fd288531d516f11e1fab47375c1ced8eed9dbaeced42db991965e94\": container with ID starting with e97e47712fd288531d516f11e1fab47375c1ced8eed9dbaeced42db991965e94 not found: ID does not exist" Feb 27 18:31:51 crc kubenswrapper[4700]: I0227 18:31:51.741216 4700 scope.go:117] "RemoveContainer" containerID="4b65a79cf2eb03a8dd97bcdd54a620d09775cde1106b8a7b7fff34d53100a0b9" Feb 27 18:31:51 crc kubenswrapper[4700]: E0227 18:31:51.741673 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b65a79cf2eb03a8dd97bcdd54a620d09775cde1106b8a7b7fff34d53100a0b9\": container with ID starting with 4b65a79cf2eb03a8dd97bcdd54a620d09775cde1106b8a7b7fff34d53100a0b9 not found: ID does not exist" containerID="4b65a79cf2eb03a8dd97bcdd54a620d09775cde1106b8a7b7fff34d53100a0b9" Feb 27 18:31:51 crc kubenswrapper[4700]: I0227 18:31:51.741721 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b65a79cf2eb03a8dd97bcdd54a620d09775cde1106b8a7b7fff34d53100a0b9"} err="failed to get container status \"4b65a79cf2eb03a8dd97bcdd54a620d09775cde1106b8a7b7fff34d53100a0b9\": rpc error: code = NotFound desc = could not find container \"4b65a79cf2eb03a8dd97bcdd54a620d09775cde1106b8a7b7fff34d53100a0b9\": container with ID starting with 4b65a79cf2eb03a8dd97bcdd54a620d09775cde1106b8a7b7fff34d53100a0b9 not found: ID does not exist" Feb 27 18:31:53 crc kubenswrapper[4700]: I0227 18:31:53.003536 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0fd7b6d-8f71-4bff-b903-cdb5a0954f87" path="/var/lib/kubelet/pods/b0fd7b6d-8f71-4bff-b903-cdb5a0954f87/volumes" Feb 27 18:32:00 crc kubenswrapper[4700]: I0227 18:32:00.152831 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536952-p4tmr"] Feb 27 18:32:00 crc kubenswrapper[4700]: E0227 18:32:00.154278 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98c1faf7-b2cc-47c3-89e3-20e563b9f507" containerName="extract-utilities" Feb 27 18:32:00 crc kubenswrapper[4700]: I0227 18:32:00.154326 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="98c1faf7-b2cc-47c3-89e3-20e563b9f507" containerName="extract-utilities" Feb 27 18:32:00 crc kubenswrapper[4700]: E0227 18:32:00.154353 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0fd7b6d-8f71-4bff-b903-cdb5a0954f87" containerName="extract-utilities" Feb 27 18:32:00 crc kubenswrapper[4700]: I0227 18:32:00.154365 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0fd7b6d-8f71-4bff-b903-cdb5a0954f87" containerName="extract-utilities" Feb 27 18:32:00 crc kubenswrapper[4700]: E0227 18:32:00.154389 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0fd7b6d-8f71-4bff-b903-cdb5a0954f87" containerName="extract-content" Feb 27 18:32:00 crc kubenswrapper[4700]: I0227 18:32:00.154401 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0fd7b6d-8f71-4bff-b903-cdb5a0954f87" containerName="extract-content" Feb 27 18:32:00 crc kubenswrapper[4700]: E0227 18:32:00.154420 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98c1faf7-b2cc-47c3-89e3-20e563b9f507" containerName="registry-server" Feb 27 18:32:00 crc kubenswrapper[4700]: I0227 18:32:00.154431 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="98c1faf7-b2cc-47c3-89e3-20e563b9f507" containerName="registry-server" Feb 27 18:32:00 crc kubenswrapper[4700]: E0227 18:32:00.154446 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="050b88a9-b342-4990-9bac-17c98b014f49" containerName="collect-profiles" Feb 27 18:32:00 crc kubenswrapper[4700]: I0227 18:32:00.154513 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="050b88a9-b342-4990-9bac-17c98b014f49" containerName="collect-profiles" Feb 27 18:32:00 crc kubenswrapper[4700]: E0227 18:32:00.154566 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98c1faf7-b2cc-47c3-89e3-20e563b9f507" containerName="extract-content" Feb 27 18:32:00 crc kubenswrapper[4700]: I0227 18:32:00.154582 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="98c1faf7-b2cc-47c3-89e3-20e563b9f507" containerName="extract-content" Feb 27 18:32:00 crc kubenswrapper[4700]: E0227 18:32:00.154610 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0fd7b6d-8f71-4bff-b903-cdb5a0954f87" containerName="registry-server" Feb 27 18:32:00 crc kubenswrapper[4700]: I0227 18:32:00.154620 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0fd7b6d-8f71-4bff-b903-cdb5a0954f87" containerName="registry-server" Feb 27 18:32:00 crc kubenswrapper[4700]: E0227 18:32:00.154646 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74489fab-6aca-4a3f-9303-04a2b82f1fde" containerName="oc" Feb 27 18:32:00 crc kubenswrapper[4700]: I0227 18:32:00.154656 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="74489fab-6aca-4a3f-9303-04a2b82f1fde" containerName="oc" Feb 27 18:32:00 crc kubenswrapper[4700]: I0227 18:32:00.154964 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="98c1faf7-b2cc-47c3-89e3-20e563b9f507" containerName="registry-server" Feb 27 18:32:00 crc kubenswrapper[4700]: I0227 18:32:00.155009 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="050b88a9-b342-4990-9bac-17c98b014f49" containerName="collect-profiles" Feb 27 18:32:00 crc kubenswrapper[4700]: I0227 18:32:00.155030 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0fd7b6d-8f71-4bff-b903-cdb5a0954f87" containerName="registry-server" Feb 27 18:32:00 crc kubenswrapper[4700]: I0227 18:32:00.155056 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="74489fab-6aca-4a3f-9303-04a2b82f1fde" containerName="oc" Feb 27 18:32:00 crc kubenswrapper[4700]: I0227 18:32:00.156095 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536952-p4tmr" Feb 27 18:32:00 crc kubenswrapper[4700]: I0227 18:32:00.163869 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536952-p4tmr"] Feb 27 18:32:00 crc kubenswrapper[4700]: I0227 18:32:00.165947 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmfp7\" (UniqueName: \"kubernetes.io/projected/38cf4801-ea2b-4538-8417-ae67e1668122-kube-api-access-xmfp7\") pod \"auto-csr-approver-29536952-p4tmr\" (UID: \"38cf4801-ea2b-4538-8417-ae67e1668122\") " pod="openshift-infra/auto-csr-approver-29536952-p4tmr" Feb 27 18:32:00 crc kubenswrapper[4700]: I0227 18:32:00.267950 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmfp7\" (UniqueName: \"kubernetes.io/projected/38cf4801-ea2b-4538-8417-ae67e1668122-kube-api-access-xmfp7\") pod \"auto-csr-approver-29536952-p4tmr\" (UID: \"38cf4801-ea2b-4538-8417-ae67e1668122\") " pod="openshift-infra/auto-csr-approver-29536952-p4tmr" Feb 27 18:32:00 crc kubenswrapper[4700]: I0227 18:32:00.298360 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmfp7\" (UniqueName: \"kubernetes.io/projected/38cf4801-ea2b-4538-8417-ae67e1668122-kube-api-access-xmfp7\") pod \"auto-csr-approver-29536952-p4tmr\" (UID: \"38cf4801-ea2b-4538-8417-ae67e1668122\") " pod="openshift-infra/auto-csr-approver-29536952-p4tmr" Feb 27 18:32:00 crc kubenswrapper[4700]: I0227 18:32:00.506103 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536952-p4tmr" Feb 27 18:32:00 crc kubenswrapper[4700]: W0227 18:32:00.986762 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod38cf4801_ea2b_4538_8417_ae67e1668122.slice/crio-7e01db1e41a75765b01945e537f579caac363b78bd149168f5b293df824e1187 WatchSource:0}: Error finding container 7e01db1e41a75765b01945e537f579caac363b78bd149168f5b293df824e1187: Status 404 returned error can't find the container with id 7e01db1e41a75765b01945e537f579caac363b78bd149168f5b293df824e1187 Feb 27 18:32:01 crc kubenswrapper[4700]: I0227 18:32:01.003856 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536952-p4tmr"] Feb 27 18:32:01 crc kubenswrapper[4700]: I0227 18:32:01.708947 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536952-p4tmr" event={"ID":"38cf4801-ea2b-4538-8417-ae67e1668122","Type":"ContainerStarted","Data":"7e01db1e41a75765b01945e537f579caac363b78bd149168f5b293df824e1187"} Feb 27 18:32:01 crc kubenswrapper[4700]: E0227 18:32:01.875446 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 18:32:01 crc kubenswrapper[4700]: E0227 18:32:01.875683 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 18:32:01 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 18:32:01 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xmfp7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536952-p4tmr_openshift-infra(38cf4801-ea2b-4538-8417-ae67e1668122): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 18:32:01 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 18:32:01 crc kubenswrapper[4700]: E0227 18:32:01.876868 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536952-p4tmr" podUID="38cf4801-ea2b-4538-8417-ae67e1668122" Feb 27 18:32:01 crc kubenswrapper[4700]: E0227 18:32:01.983235 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:32:02 crc kubenswrapper[4700]: E0227 18:32:02.729206 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536952-p4tmr" podUID="38cf4801-ea2b-4538-8417-ae67e1668122" Feb 27 18:32:16 crc kubenswrapper[4700]: E0227 18:32:16.984602 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:32:17 crc kubenswrapper[4700]: I0227 18:32:17.932061 4700 generic.go:334] "Generic (PLEG): container finished" podID="38cf4801-ea2b-4538-8417-ae67e1668122" containerID="b4e1fe01b7eeeb7aff8e5c7ec44de422fdc079aa27f51ebade305bc61c0305db" exitCode=0 Feb 27 18:32:17 crc kubenswrapper[4700]: I0227 18:32:17.932251 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536952-p4tmr" event={"ID":"38cf4801-ea2b-4538-8417-ae67e1668122","Type":"ContainerDied","Data":"b4e1fe01b7eeeb7aff8e5c7ec44de422fdc079aa27f51ebade305bc61c0305db"} Feb 27 18:32:19 crc kubenswrapper[4700]: I0227 18:32:19.328047 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536952-p4tmr" Feb 27 18:32:19 crc kubenswrapper[4700]: I0227 18:32:19.455546 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmfp7\" (UniqueName: \"kubernetes.io/projected/38cf4801-ea2b-4538-8417-ae67e1668122-kube-api-access-xmfp7\") pod \"38cf4801-ea2b-4538-8417-ae67e1668122\" (UID: \"38cf4801-ea2b-4538-8417-ae67e1668122\") " Feb 27 18:32:19 crc kubenswrapper[4700]: I0227 18:32:19.466058 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38cf4801-ea2b-4538-8417-ae67e1668122-kube-api-access-xmfp7" (OuterVolumeSpecName: "kube-api-access-xmfp7") pod "38cf4801-ea2b-4538-8417-ae67e1668122" (UID: "38cf4801-ea2b-4538-8417-ae67e1668122"). InnerVolumeSpecName "kube-api-access-xmfp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:32:19 crc kubenswrapper[4700]: I0227 18:32:19.559419 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmfp7\" (UniqueName: \"kubernetes.io/projected/38cf4801-ea2b-4538-8417-ae67e1668122-kube-api-access-xmfp7\") on node \"crc\" DevicePath \"\"" Feb 27 18:32:19 crc kubenswrapper[4700]: I0227 18:32:19.961367 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536952-p4tmr" event={"ID":"38cf4801-ea2b-4538-8417-ae67e1668122","Type":"ContainerDied","Data":"7e01db1e41a75765b01945e537f579caac363b78bd149168f5b293df824e1187"} Feb 27 18:32:19 crc kubenswrapper[4700]: I0227 18:32:19.961437 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e01db1e41a75765b01945e537f579caac363b78bd149168f5b293df824e1187" Feb 27 18:32:19 crc kubenswrapper[4700]: I0227 18:32:19.961548 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536952-p4tmr" Feb 27 18:32:20 crc kubenswrapper[4700]: I0227 18:32:20.473198 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536946-rbnt5"] Feb 27 18:32:20 crc kubenswrapper[4700]: I0227 18:32:20.487826 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536946-rbnt5"] Feb 27 18:32:21 crc kubenswrapper[4700]: I0227 18:32:21.001616 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3c40c88-7b53-43dd-8357-17b25bdcc22d" path="/var/lib/kubelet/pods/c3c40c88-7b53-43dd-8357-17b25bdcc22d/volumes" Feb 27 18:32:31 crc kubenswrapper[4700]: E0227 18:32:31.985908 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:32:44 crc kubenswrapper[4700]: E0227 18:32:44.984056 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:32:58 crc kubenswrapper[4700]: E0227 18:32:58.990930 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:33:11 crc kubenswrapper[4700]: E0227 18:33:11.985889 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:33:13 crc kubenswrapper[4700]: I0227 18:33:13.789217 4700 scope.go:117] "RemoveContainer" containerID="eb2f510fecd019a1d6c0a1d4ab0815196430e59068d0c0d5be165340ebef96e9" Feb 27 18:33:26 crc kubenswrapper[4700]: E0227 18:33:26.986702 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:33:30 crc kubenswrapper[4700]: I0227 18:33:30.825265 4700 generic.go:334] "Generic (PLEG): container finished" podID="805fcd3a-360c-445d-85d8-889bf15d310b" containerID="72ffb278d0378cf551b82b535304736b4dcda281beffa87353431f5b4d160ec5" exitCode=1 Feb 27 18:33:30 crc kubenswrapper[4700]: I0227 18:33:30.825383 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"805fcd3a-360c-445d-85d8-889bf15d310b","Type":"ContainerDied","Data":"72ffb278d0378cf551b82b535304736b4dcda281beffa87353431f5b4d160ec5"} Feb 27 18:33:32 crc kubenswrapper[4700]: I0227 18:33:32.340930 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 27 18:33:32 crc kubenswrapper[4700]: I0227 18:33:32.506054 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/805fcd3a-360c-445d-85d8-889bf15d310b-test-operator-ephemeral-temporary\") pod \"805fcd3a-360c-445d-85d8-889bf15d310b\" (UID: \"805fcd3a-360c-445d-85d8-889bf15d310b\") " Feb 27 18:33:32 crc kubenswrapper[4700]: I0227 18:33:32.506124 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/805fcd3a-360c-445d-85d8-889bf15d310b-ca-certs\") pod \"805fcd3a-360c-445d-85d8-889bf15d310b\" (UID: \"805fcd3a-360c-445d-85d8-889bf15d310b\") " Feb 27 18:33:32 crc kubenswrapper[4700]: I0227 18:33:32.506173 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5gwp\" (UniqueName: \"kubernetes.io/projected/805fcd3a-360c-445d-85d8-889bf15d310b-kube-api-access-k5gwp\") pod \"805fcd3a-360c-445d-85d8-889bf15d310b\" (UID: \"805fcd3a-360c-445d-85d8-889bf15d310b\") " Feb 27 18:33:32 crc kubenswrapper[4700]: I0227 18:33:32.506271 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/805fcd3a-360c-445d-85d8-889bf15d310b-openstack-config-secret\") pod \"805fcd3a-360c-445d-85d8-889bf15d310b\" (UID: \"805fcd3a-360c-445d-85d8-889bf15d310b\") " Feb 27 18:33:32 crc kubenswrapper[4700]: I0227 18:33:32.506337 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/805fcd3a-360c-445d-85d8-889bf15d310b-ssh-key\") pod \"805fcd3a-360c-445d-85d8-889bf15d310b\" (UID: \"805fcd3a-360c-445d-85d8-889bf15d310b\") " Feb 27 18:33:32 crc kubenswrapper[4700]: I0227 18:33:32.506375 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/805fcd3a-360c-445d-85d8-889bf15d310b-config-data\") pod \"805fcd3a-360c-445d-85d8-889bf15d310b\" (UID: \"805fcd3a-360c-445d-85d8-889bf15d310b\") " Feb 27 18:33:32 crc kubenswrapper[4700]: I0227 18:33:32.506413 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/805fcd3a-360c-445d-85d8-889bf15d310b-test-operator-ephemeral-workdir\") pod \"805fcd3a-360c-445d-85d8-889bf15d310b\" (UID: \"805fcd3a-360c-445d-85d8-889bf15d310b\") " Feb 27 18:33:32 crc kubenswrapper[4700]: I0227 18:33:32.506446 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"805fcd3a-360c-445d-85d8-889bf15d310b\" (UID: \"805fcd3a-360c-445d-85d8-889bf15d310b\") " Feb 27 18:33:32 crc kubenswrapper[4700]: I0227 18:33:32.506572 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/805fcd3a-360c-445d-85d8-889bf15d310b-openstack-config\") pod \"805fcd3a-360c-445d-85d8-889bf15d310b\" (UID: \"805fcd3a-360c-445d-85d8-889bf15d310b\") " Feb 27 18:33:32 crc kubenswrapper[4700]: I0227 18:33:32.507613 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/805fcd3a-360c-445d-85d8-889bf15d310b-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "805fcd3a-360c-445d-85d8-889bf15d310b" (UID: "805fcd3a-360c-445d-85d8-889bf15d310b"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 18:33:32 crc kubenswrapper[4700]: I0227 18:33:32.509851 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/805fcd3a-360c-445d-85d8-889bf15d310b-config-data" (OuterVolumeSpecName: "config-data") pod "805fcd3a-360c-445d-85d8-889bf15d310b" (UID: "805fcd3a-360c-445d-85d8-889bf15d310b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 18:33:32 crc kubenswrapper[4700]: I0227 18:33:32.514400 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/805fcd3a-360c-445d-85d8-889bf15d310b-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "805fcd3a-360c-445d-85d8-889bf15d310b" (UID: "805fcd3a-360c-445d-85d8-889bf15d310b"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 18:33:32 crc kubenswrapper[4700]: I0227 18:33:32.516007 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/805fcd3a-360c-445d-85d8-889bf15d310b-kube-api-access-k5gwp" (OuterVolumeSpecName: "kube-api-access-k5gwp") pod "805fcd3a-360c-445d-85d8-889bf15d310b" (UID: "805fcd3a-360c-445d-85d8-889bf15d310b"). InnerVolumeSpecName "kube-api-access-k5gwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:33:32 crc kubenswrapper[4700]: I0227 18:33:32.517728 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "test-operator-logs") pod "805fcd3a-360c-445d-85d8-889bf15d310b" (UID: "805fcd3a-360c-445d-85d8-889bf15d310b"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 27 18:33:32 crc kubenswrapper[4700]: I0227 18:33:32.547014 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/805fcd3a-360c-445d-85d8-889bf15d310b-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "805fcd3a-360c-445d-85d8-889bf15d310b" (UID: "805fcd3a-360c-445d-85d8-889bf15d310b"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 18:33:32 crc kubenswrapper[4700]: I0227 18:33:32.548131 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/805fcd3a-360c-445d-85d8-889bf15d310b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "805fcd3a-360c-445d-85d8-889bf15d310b" (UID: "805fcd3a-360c-445d-85d8-889bf15d310b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 18:33:32 crc kubenswrapper[4700]: I0227 18:33:32.563427 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/805fcd3a-360c-445d-85d8-889bf15d310b-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "805fcd3a-360c-445d-85d8-889bf15d310b" (UID: "805fcd3a-360c-445d-85d8-889bf15d310b"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 18:33:32 crc kubenswrapper[4700]: I0227 18:33:32.601235 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/805fcd3a-360c-445d-85d8-889bf15d310b-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "805fcd3a-360c-445d-85d8-889bf15d310b" (UID: "805fcd3a-360c-445d-85d8-889bf15d310b"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 18:33:32 crc kubenswrapper[4700]: I0227 18:33:32.609189 4700 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/805fcd3a-360c-445d-85d8-889bf15d310b-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 18:33:32 crc kubenswrapper[4700]: I0227 18:33:32.609229 4700 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/805fcd3a-360c-445d-85d8-889bf15d310b-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Feb 27 18:33:32 crc kubenswrapper[4700]: I0227 18:33:32.609278 4700 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Feb 27 18:33:32 crc kubenswrapper[4700]: I0227 18:33:32.609300 4700 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/805fcd3a-360c-445d-85d8-889bf15d310b-openstack-config\") on node \"crc\" DevicePath \"\"" Feb 27 18:33:32 crc kubenswrapper[4700]: I0227 18:33:32.609319 4700 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/805fcd3a-360c-445d-85d8-889bf15d310b-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Feb 27 18:33:32 crc kubenswrapper[4700]: I0227 18:33:32.609334 4700 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/805fcd3a-360c-445d-85d8-889bf15d310b-ca-certs\") on node \"crc\" DevicePath \"\"" Feb 27 18:33:32 crc kubenswrapper[4700]: I0227 18:33:32.609349 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5gwp\" (UniqueName: \"kubernetes.io/projected/805fcd3a-360c-445d-85d8-889bf15d310b-kube-api-access-k5gwp\") on node \"crc\" DevicePath \"\"" Feb 27 18:33:32 crc kubenswrapper[4700]: I0227 18:33:32.609364 4700 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/805fcd3a-360c-445d-85d8-889bf15d310b-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Feb 27 18:33:32 crc kubenswrapper[4700]: I0227 18:33:32.609379 4700 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/805fcd3a-360c-445d-85d8-889bf15d310b-ssh-key\") on node \"crc\" DevicePath \"\"" Feb 27 18:33:32 crc kubenswrapper[4700]: I0227 18:33:32.656875 4700 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Feb 27 18:33:32 crc kubenswrapper[4700]: I0227 18:33:32.712407 4700 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Feb 27 18:33:32 crc kubenswrapper[4700]: I0227 18:33:32.852448 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"805fcd3a-360c-445d-85d8-889bf15d310b","Type":"ContainerDied","Data":"c16b5ad8ef3e1c56a689da5097461cd9ac5d9508dd8e8a8766f2406d1017456f"} Feb 27 18:33:32 crc kubenswrapper[4700]: I0227 18:33:32.852530 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c16b5ad8ef3e1c56a689da5097461cd9ac5d9508dd8e8a8766f2406d1017456f" Feb 27 18:33:32 crc kubenswrapper[4700]: I0227 18:33:32.852611 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 27 18:33:35 crc kubenswrapper[4700]: I0227 18:33:35.995285 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 27 18:33:36 crc kubenswrapper[4700]: E0227 18:33:36.003112 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="805fcd3a-360c-445d-85d8-889bf15d310b" containerName="tempest-tests-tempest-tests-runner" Feb 27 18:33:36 crc kubenswrapper[4700]: I0227 18:33:36.003162 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="805fcd3a-360c-445d-85d8-889bf15d310b" containerName="tempest-tests-tempest-tests-runner" Feb 27 18:33:36 crc kubenswrapper[4700]: E0227 18:33:36.003247 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38cf4801-ea2b-4538-8417-ae67e1668122" containerName="oc" Feb 27 18:33:36 crc kubenswrapper[4700]: I0227 18:33:36.003258 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="38cf4801-ea2b-4538-8417-ae67e1668122" containerName="oc" Feb 27 18:33:36 crc kubenswrapper[4700]: I0227 18:33:36.004040 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="805fcd3a-360c-445d-85d8-889bf15d310b" containerName="tempest-tests-tempest-tests-runner" Feb 27 18:33:36 crc kubenswrapper[4700]: I0227 18:33:36.004063 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="38cf4801-ea2b-4538-8417-ae67e1668122" containerName="oc" Feb 27 18:33:36 crc kubenswrapper[4700]: I0227 18:33:36.007687 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 27 18:33:36 crc kubenswrapper[4700]: I0227 18:33:36.014890 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-nrmqz" Feb 27 18:33:36 crc kubenswrapper[4700]: I0227 18:33:36.046673 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 27 18:33:36 crc kubenswrapper[4700]: I0227 18:33:36.198769 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4n6h\" (UniqueName: \"kubernetes.io/projected/82d7f7cf-9ad7-4f87-b31c-a3c31b454289-kube-api-access-m4n6h\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"82d7f7cf-9ad7-4f87-b31c-a3c31b454289\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 27 18:33:36 crc kubenswrapper[4700]: I0227 18:33:36.198933 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"82d7f7cf-9ad7-4f87-b31c-a3c31b454289\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 27 18:33:36 crc kubenswrapper[4700]: I0227 18:33:36.300772 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4n6h\" (UniqueName: \"kubernetes.io/projected/82d7f7cf-9ad7-4f87-b31c-a3c31b454289-kube-api-access-m4n6h\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"82d7f7cf-9ad7-4f87-b31c-a3c31b454289\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 27 18:33:36 crc kubenswrapper[4700]: I0227 18:33:36.300879 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"82d7f7cf-9ad7-4f87-b31c-a3c31b454289\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 27 18:33:36 crc kubenswrapper[4700]: I0227 18:33:36.301421 4700 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"82d7f7cf-9ad7-4f87-b31c-a3c31b454289\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 27 18:33:36 crc kubenswrapper[4700]: I0227 18:33:36.328025 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4n6h\" (UniqueName: \"kubernetes.io/projected/82d7f7cf-9ad7-4f87-b31c-a3c31b454289-kube-api-access-m4n6h\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"82d7f7cf-9ad7-4f87-b31c-a3c31b454289\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 27 18:33:36 crc kubenswrapper[4700]: I0227 18:33:36.333383 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"82d7f7cf-9ad7-4f87-b31c-a3c31b454289\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 27 18:33:36 crc kubenswrapper[4700]: I0227 18:33:36.355449 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 27 18:33:36 crc kubenswrapper[4700]: I0227 18:33:36.410674 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 18:33:36 crc kubenswrapper[4700]: I0227 18:33:36.410865 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 18:33:36 crc kubenswrapper[4700]: I0227 18:33:36.850435 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 27 18:33:37 crc kubenswrapper[4700]: I0227 18:33:37.561808 4700 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 18:33:37 crc kubenswrapper[4700]: I0227 18:33:37.923285 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"82d7f7cf-9ad7-4f87-b31c-a3c31b454289","Type":"ContainerStarted","Data":"f6fbc97bedf4c470dc6351483d7179a4d033a9927bc1200d9e1458d2653cbcf6"} Feb 27 18:33:38 crc kubenswrapper[4700]: I0227 18:33:38.934839 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"82d7f7cf-9ad7-4f87-b31c-a3c31b454289","Type":"ContainerStarted","Data":"863c129423c78357400a989c6d4f63fdf977c5d2c8b0b6a2ae80662fdb1fd3e4"} Feb 27 18:33:38 crc kubenswrapper[4700]: I0227 18:33:38.962916 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.884647756 podStartE2EDuration="3.962887284s" podCreationTimestamp="2026-02-27 18:33:35 +0000 UTC" firstStartedPulling="2026-02-27 18:33:37.561412725 +0000 UTC m=+5577.546725512" lastFinishedPulling="2026-02-27 18:33:38.639652263 +0000 UTC m=+5578.624965040" observedRunningTime="2026-02-27 18:33:38.9539333 +0000 UTC m=+5578.939246067" watchObservedRunningTime="2026-02-27 18:33:38.962887284 +0000 UTC m=+5578.948200071" Feb 27 18:33:39 crc kubenswrapper[4700]: E0227 18:33:39.983744 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:33:51 crc kubenswrapper[4700]: E0227 18:33:51.016146 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:33:52 crc kubenswrapper[4700]: I0227 18:33:52.489191 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-g4j9g"] Feb 27 18:33:52 crc kubenswrapper[4700]: I0227 18:33:52.492007 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g4j9g" Feb 27 18:33:52 crc kubenswrapper[4700]: I0227 18:33:52.515927 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g4j9g"] Feb 27 18:33:52 crc kubenswrapper[4700]: I0227 18:33:52.536909 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4-catalog-content\") pod \"certified-operators-g4j9g\" (UID: \"d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4\") " pod="openshift-marketplace/certified-operators-g4j9g" Feb 27 18:33:52 crc kubenswrapper[4700]: I0227 18:33:52.536980 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4-utilities\") pod \"certified-operators-g4j9g\" (UID: \"d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4\") " pod="openshift-marketplace/certified-operators-g4j9g" Feb 27 18:33:52 crc kubenswrapper[4700]: I0227 18:33:52.537017 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szxxw\" (UniqueName: \"kubernetes.io/projected/d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4-kube-api-access-szxxw\") pod \"certified-operators-g4j9g\" (UID: \"d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4\") " pod="openshift-marketplace/certified-operators-g4j9g" Feb 27 18:33:52 crc kubenswrapper[4700]: I0227 18:33:52.638587 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4-catalog-content\") pod \"certified-operators-g4j9g\" (UID: \"d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4\") " pod="openshift-marketplace/certified-operators-g4j9g" Feb 27 18:33:52 crc kubenswrapper[4700]: I0227 18:33:52.638661 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4-utilities\") pod \"certified-operators-g4j9g\" (UID: \"d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4\") " pod="openshift-marketplace/certified-operators-g4j9g" Feb 27 18:33:52 crc kubenswrapper[4700]: I0227 18:33:52.638702 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szxxw\" (UniqueName: \"kubernetes.io/projected/d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4-kube-api-access-szxxw\") pod \"certified-operators-g4j9g\" (UID: \"d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4\") " pod="openshift-marketplace/certified-operators-g4j9g" Feb 27 18:33:52 crc kubenswrapper[4700]: I0227 18:33:52.639072 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4-catalog-content\") pod \"certified-operators-g4j9g\" (UID: \"d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4\") " pod="openshift-marketplace/certified-operators-g4j9g" Feb 27 18:33:52 crc kubenswrapper[4700]: I0227 18:33:52.639138 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4-utilities\") pod \"certified-operators-g4j9g\" (UID: \"d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4\") " pod="openshift-marketplace/certified-operators-g4j9g" Feb 27 18:33:52 crc kubenswrapper[4700]: I0227 18:33:52.666214 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szxxw\" (UniqueName: \"kubernetes.io/projected/d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4-kube-api-access-szxxw\") pod \"certified-operators-g4j9g\" (UID: \"d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4\") " pod="openshift-marketplace/certified-operators-g4j9g" Feb 27 18:33:52 crc kubenswrapper[4700]: I0227 18:33:52.814454 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g4j9g" Feb 27 18:33:53 crc kubenswrapper[4700]: I0227 18:33:53.311468 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g4j9g"] Feb 27 18:33:54 crc kubenswrapper[4700]: I0227 18:33:54.158269 4700 generic.go:334] "Generic (PLEG): container finished" podID="d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4" containerID="7b395c0ccc6f4e0f009f46b86e0a008d6d1e5684c909da03fc98d06f19190e62" exitCode=0 Feb 27 18:33:54 crc kubenswrapper[4700]: I0227 18:33:54.158352 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g4j9g" event={"ID":"d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4","Type":"ContainerDied","Data":"7b395c0ccc6f4e0f009f46b86e0a008d6d1e5684c909da03fc98d06f19190e62"} Feb 27 18:33:54 crc kubenswrapper[4700]: I0227 18:33:54.158760 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g4j9g" event={"ID":"d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4","Type":"ContainerStarted","Data":"f749db84c73eefcae7ee4966494c6544f0ad122500795beee0a15359ef9e5a73"} Feb 27 18:33:55 crc kubenswrapper[4700]: I0227 18:33:55.175370 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g4j9g" event={"ID":"d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4","Type":"ContainerStarted","Data":"f694b501c26543cefedbaf7636bf3c06a100bb9c5d74e1b84c52c3773b79685b"} Feb 27 18:33:57 crc kubenswrapper[4700]: I0227 18:33:57.206618 4700 generic.go:334] "Generic (PLEG): container finished" podID="d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4" containerID="f694b501c26543cefedbaf7636bf3c06a100bb9c5d74e1b84c52c3773b79685b" exitCode=0 Feb 27 18:33:57 crc kubenswrapper[4700]: I0227 18:33:57.206723 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g4j9g" event={"ID":"d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4","Type":"ContainerDied","Data":"f694b501c26543cefedbaf7636bf3c06a100bb9c5d74e1b84c52c3773b79685b"} Feb 27 18:33:59 crc kubenswrapper[4700]: I0227 18:33:59.254624 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g4j9g" event={"ID":"d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4","Type":"ContainerStarted","Data":"b7f2fe325bb8123a2228d411c37daaadfa601a9907bf81667dbcee1d84f6226f"} Feb 27 18:33:59 crc kubenswrapper[4700]: I0227 18:33:59.280418 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-g4j9g" podStartSLOduration=3.505686206 podStartE2EDuration="7.280399898s" podCreationTimestamp="2026-02-27 18:33:52 +0000 UTC" firstStartedPulling="2026-02-27 18:33:54.161394947 +0000 UTC m=+5594.146707704" lastFinishedPulling="2026-02-27 18:33:57.936108609 +0000 UTC m=+5597.921421396" observedRunningTime="2026-02-27 18:33:59.276536127 +0000 UTC m=+5599.261848874" watchObservedRunningTime="2026-02-27 18:33:59.280399898 +0000 UTC m=+5599.265712645" Feb 27 18:34:00 crc kubenswrapper[4700]: I0227 18:34:00.166051 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536954-6cc9p"] Feb 27 18:34:00 crc kubenswrapper[4700]: I0227 18:34:00.168711 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536954-6cc9p" Feb 27 18:34:00 crc kubenswrapper[4700]: I0227 18:34:00.186667 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536954-6cc9p"] Feb 27 18:34:00 crc kubenswrapper[4700]: I0227 18:34:00.253058 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkx26\" (UniqueName: \"kubernetes.io/projected/3dc264f3-4cbc-4c56-b9f8-4090451d6f8d-kube-api-access-fkx26\") pod \"auto-csr-approver-29536954-6cc9p\" (UID: \"3dc264f3-4cbc-4c56-b9f8-4090451d6f8d\") " pod="openshift-infra/auto-csr-approver-29536954-6cc9p" Feb 27 18:34:00 crc kubenswrapper[4700]: I0227 18:34:00.355278 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkx26\" (UniqueName: \"kubernetes.io/projected/3dc264f3-4cbc-4c56-b9f8-4090451d6f8d-kube-api-access-fkx26\") pod \"auto-csr-approver-29536954-6cc9p\" (UID: \"3dc264f3-4cbc-4c56-b9f8-4090451d6f8d\") " pod="openshift-infra/auto-csr-approver-29536954-6cc9p" Feb 27 18:34:00 crc kubenswrapper[4700]: I0227 18:34:00.382516 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkx26\" (UniqueName: \"kubernetes.io/projected/3dc264f3-4cbc-4c56-b9f8-4090451d6f8d-kube-api-access-fkx26\") pod \"auto-csr-approver-29536954-6cc9p\" (UID: \"3dc264f3-4cbc-4c56-b9f8-4090451d6f8d\") " pod="openshift-infra/auto-csr-approver-29536954-6cc9p" Feb 27 18:34:00 crc kubenswrapper[4700]: I0227 18:34:00.505533 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536954-6cc9p" Feb 27 18:34:01 crc kubenswrapper[4700]: I0227 18:34:01.007826 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536954-6cc9p"] Feb 27 18:34:01 crc kubenswrapper[4700]: W0227 18:34:01.015125 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3dc264f3_4cbc_4c56_b9f8_4090451d6f8d.slice/crio-84643ed6ef69419f2a0e8afac8ea262aee8c0104bc6d8a3240147e9abb679ab0 WatchSource:0}: Error finding container 84643ed6ef69419f2a0e8afac8ea262aee8c0104bc6d8a3240147e9abb679ab0: Status 404 returned error can't find the container with id 84643ed6ef69419f2a0e8afac8ea262aee8c0104bc6d8a3240147e9abb679ab0 Feb 27 18:34:01 crc kubenswrapper[4700]: I0227 18:34:01.296687 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536954-6cc9p" event={"ID":"3dc264f3-4cbc-4c56-b9f8-4090451d6f8d","Type":"ContainerStarted","Data":"84643ed6ef69419f2a0e8afac8ea262aee8c0104bc6d8a3240147e9abb679ab0"} Feb 27 18:34:02 crc kubenswrapper[4700]: I0227 18:34:02.814612 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-g4j9g" Feb 27 18:34:02 crc kubenswrapper[4700]: I0227 18:34:02.815111 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-g4j9g" Feb 27 18:34:02 crc kubenswrapper[4700]: I0227 18:34:02.906024 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-g4j9g" Feb 27 18:34:03 crc kubenswrapper[4700]: I0227 18:34:03.345870 4700 generic.go:334] "Generic (PLEG): container finished" podID="3dc264f3-4cbc-4c56-b9f8-4090451d6f8d" containerID="a20d6b3024c7096e4d6dbd8f38beca19c244a4c825e28672f18614b6a3a89ef5" exitCode=0 Feb 27 18:34:03 crc kubenswrapper[4700]: I0227 18:34:03.346538 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536954-6cc9p" event={"ID":"3dc264f3-4cbc-4c56-b9f8-4090451d6f8d","Type":"ContainerDied","Data":"a20d6b3024c7096e4d6dbd8f38beca19c244a4c825e28672f18614b6a3a89ef5"} Feb 27 18:34:03 crc kubenswrapper[4700]: I0227 18:34:03.477202 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-g4j9g" Feb 27 18:34:04 crc kubenswrapper[4700]: I0227 18:34:04.890067 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536954-6cc9p" Feb 27 18:34:04 crc kubenswrapper[4700]: I0227 18:34:04.992674 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fkx26\" (UniqueName: \"kubernetes.io/projected/3dc264f3-4cbc-4c56-b9f8-4090451d6f8d-kube-api-access-fkx26\") pod \"3dc264f3-4cbc-4c56-b9f8-4090451d6f8d\" (UID: \"3dc264f3-4cbc-4c56-b9f8-4090451d6f8d\") " Feb 27 18:34:05 crc kubenswrapper[4700]: I0227 18:34:05.005168 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dc264f3-4cbc-4c56-b9f8-4090451d6f8d-kube-api-access-fkx26" (OuterVolumeSpecName: "kube-api-access-fkx26") pod "3dc264f3-4cbc-4c56-b9f8-4090451d6f8d" (UID: "3dc264f3-4cbc-4c56-b9f8-4090451d6f8d"). InnerVolumeSpecName "kube-api-access-fkx26". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:34:05 crc kubenswrapper[4700]: I0227 18:34:05.096940 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fkx26\" (UniqueName: \"kubernetes.io/projected/3dc264f3-4cbc-4c56-b9f8-4090451d6f8d-kube-api-access-fkx26\") on node \"crc\" DevicePath \"\"" Feb 27 18:34:05 crc kubenswrapper[4700]: I0227 18:34:05.373787 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536954-6cc9p" event={"ID":"3dc264f3-4cbc-4c56-b9f8-4090451d6f8d","Type":"ContainerDied","Data":"84643ed6ef69419f2a0e8afac8ea262aee8c0104bc6d8a3240147e9abb679ab0"} Feb 27 18:34:05 crc kubenswrapper[4700]: I0227 18:34:05.374031 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="84643ed6ef69419f2a0e8afac8ea262aee8c0104bc6d8a3240147e9abb679ab0" Feb 27 18:34:05 crc kubenswrapper[4700]: I0227 18:34:05.374101 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536954-6cc9p" Feb 27 18:34:05 crc kubenswrapper[4700]: E0227 18:34:05.988275 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:34:06 crc kubenswrapper[4700]: I0227 18:34:06.003481 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536948-zt6lq"] Feb 27 18:34:06 crc kubenswrapper[4700]: I0227 18:34:06.020502 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536948-zt6lq"] Feb 27 18:34:06 crc kubenswrapper[4700]: I0227 18:34:06.416777 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 18:34:06 crc kubenswrapper[4700]: I0227 18:34:06.416897 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 18:34:07 crc kubenswrapper[4700]: I0227 18:34:07.002263 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c99a671-4160-4665-bb56-5630aad6ad58" path="/var/lib/kubelet/pods/4c99a671-4160-4665-bb56-5630aad6ad58/volumes" Feb 27 18:34:07 crc kubenswrapper[4700]: I0227 18:34:07.361743 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-g4j9g"] Feb 27 18:34:07 crc kubenswrapper[4700]: I0227 18:34:07.362667 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-g4j9g" podUID="d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4" containerName="registry-server" containerID="cri-o://b7f2fe325bb8123a2228d411c37daaadfa601a9907bf81667dbcee1d84f6226f" gracePeriod=2 Feb 27 18:34:08 crc kubenswrapper[4700]: I0227 18:34:08.006300 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g4j9g" Feb 27 18:34:08 crc kubenswrapper[4700]: I0227 18:34:08.084894 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szxxw\" (UniqueName: \"kubernetes.io/projected/d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4-kube-api-access-szxxw\") pod \"d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4\" (UID: \"d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4\") " Feb 27 18:34:08 crc kubenswrapper[4700]: I0227 18:34:08.085014 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4-catalog-content\") pod \"d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4\" (UID: \"d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4\") " Feb 27 18:34:08 crc kubenswrapper[4700]: I0227 18:34:08.085184 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4-utilities\") pod \"d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4\" (UID: \"d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4\") " Feb 27 18:34:08 crc kubenswrapper[4700]: I0227 18:34:08.086840 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4-utilities" (OuterVolumeSpecName: "utilities") pod "d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4" (UID: "d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 18:34:08 crc kubenswrapper[4700]: I0227 18:34:08.113822 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4-kube-api-access-szxxw" (OuterVolumeSpecName: "kube-api-access-szxxw") pod "d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4" (UID: "d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4"). InnerVolumeSpecName "kube-api-access-szxxw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:34:08 crc kubenswrapper[4700]: I0227 18:34:08.149646 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4" (UID: "d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 18:34:08 crc kubenswrapper[4700]: I0227 18:34:08.189264 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 18:34:08 crc kubenswrapper[4700]: I0227 18:34:08.189322 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szxxw\" (UniqueName: \"kubernetes.io/projected/d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4-kube-api-access-szxxw\") on node \"crc\" DevicePath \"\"" Feb 27 18:34:08 crc kubenswrapper[4700]: I0227 18:34:08.189340 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 18:34:08 crc kubenswrapper[4700]: I0227 18:34:08.467234 4700 generic.go:334] "Generic (PLEG): container finished" podID="d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4" containerID="b7f2fe325bb8123a2228d411c37daaadfa601a9907bf81667dbcee1d84f6226f" exitCode=0 Feb 27 18:34:08 crc kubenswrapper[4700]: I0227 18:34:08.467300 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g4j9g" event={"ID":"d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4","Type":"ContainerDied","Data":"b7f2fe325bb8123a2228d411c37daaadfa601a9907bf81667dbcee1d84f6226f"} Feb 27 18:34:08 crc kubenswrapper[4700]: I0227 18:34:08.467343 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g4j9g" event={"ID":"d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4","Type":"ContainerDied","Data":"f749db84c73eefcae7ee4966494c6544f0ad122500795beee0a15359ef9e5a73"} Feb 27 18:34:08 crc kubenswrapper[4700]: I0227 18:34:08.467372 4700 scope.go:117] "RemoveContainer" containerID="b7f2fe325bb8123a2228d411c37daaadfa601a9907bf81667dbcee1d84f6226f" Feb 27 18:34:08 crc kubenswrapper[4700]: I0227 18:34:08.467619 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g4j9g" Feb 27 18:34:08 crc kubenswrapper[4700]: I0227 18:34:08.522068 4700 scope.go:117] "RemoveContainer" containerID="f694b501c26543cefedbaf7636bf3c06a100bb9c5d74e1b84c52c3773b79685b" Feb 27 18:34:08 crc kubenswrapper[4700]: I0227 18:34:08.533808 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-g4j9g"] Feb 27 18:34:08 crc kubenswrapper[4700]: I0227 18:34:08.546530 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-g4j9g"] Feb 27 18:34:08 crc kubenswrapper[4700]: I0227 18:34:08.556209 4700 scope.go:117] "RemoveContainer" containerID="7b395c0ccc6f4e0f009f46b86e0a008d6d1e5684c909da03fc98d06f19190e62" Feb 27 18:34:08 crc kubenswrapper[4700]: I0227 18:34:08.634092 4700 scope.go:117] "RemoveContainer" containerID="b7f2fe325bb8123a2228d411c37daaadfa601a9907bf81667dbcee1d84f6226f" Feb 27 18:34:08 crc kubenswrapper[4700]: E0227 18:34:08.634888 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7f2fe325bb8123a2228d411c37daaadfa601a9907bf81667dbcee1d84f6226f\": container with ID starting with b7f2fe325bb8123a2228d411c37daaadfa601a9907bf81667dbcee1d84f6226f not found: ID does not exist" containerID="b7f2fe325bb8123a2228d411c37daaadfa601a9907bf81667dbcee1d84f6226f" Feb 27 18:34:08 crc kubenswrapper[4700]: I0227 18:34:08.635038 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7f2fe325bb8123a2228d411c37daaadfa601a9907bf81667dbcee1d84f6226f"} err="failed to get container status \"b7f2fe325bb8123a2228d411c37daaadfa601a9907bf81667dbcee1d84f6226f\": rpc error: code = NotFound desc = could not find container \"b7f2fe325bb8123a2228d411c37daaadfa601a9907bf81667dbcee1d84f6226f\": container with ID starting with b7f2fe325bb8123a2228d411c37daaadfa601a9907bf81667dbcee1d84f6226f not found: ID does not exist" Feb 27 18:34:08 crc kubenswrapper[4700]: I0227 18:34:08.635152 4700 scope.go:117] "RemoveContainer" containerID="f694b501c26543cefedbaf7636bf3c06a100bb9c5d74e1b84c52c3773b79685b" Feb 27 18:34:08 crc kubenswrapper[4700]: E0227 18:34:08.635877 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f694b501c26543cefedbaf7636bf3c06a100bb9c5d74e1b84c52c3773b79685b\": container with ID starting with f694b501c26543cefedbaf7636bf3c06a100bb9c5d74e1b84c52c3773b79685b not found: ID does not exist" containerID="f694b501c26543cefedbaf7636bf3c06a100bb9c5d74e1b84c52c3773b79685b" Feb 27 18:34:08 crc kubenswrapper[4700]: I0227 18:34:08.636016 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f694b501c26543cefedbaf7636bf3c06a100bb9c5d74e1b84c52c3773b79685b"} err="failed to get container status \"f694b501c26543cefedbaf7636bf3c06a100bb9c5d74e1b84c52c3773b79685b\": rpc error: code = NotFound desc = could not find container \"f694b501c26543cefedbaf7636bf3c06a100bb9c5d74e1b84c52c3773b79685b\": container with ID starting with f694b501c26543cefedbaf7636bf3c06a100bb9c5d74e1b84c52c3773b79685b not found: ID does not exist" Feb 27 18:34:08 crc kubenswrapper[4700]: I0227 18:34:08.636111 4700 scope.go:117] "RemoveContainer" containerID="7b395c0ccc6f4e0f009f46b86e0a008d6d1e5684c909da03fc98d06f19190e62" Feb 27 18:34:08 crc kubenswrapper[4700]: E0227 18:34:08.636638 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b395c0ccc6f4e0f009f46b86e0a008d6d1e5684c909da03fc98d06f19190e62\": container with ID starting with 7b395c0ccc6f4e0f009f46b86e0a008d6d1e5684c909da03fc98d06f19190e62 not found: ID does not exist" containerID="7b395c0ccc6f4e0f009f46b86e0a008d6d1e5684c909da03fc98d06f19190e62" Feb 27 18:34:08 crc kubenswrapper[4700]: I0227 18:34:08.636682 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b395c0ccc6f4e0f009f46b86e0a008d6d1e5684c909da03fc98d06f19190e62"} err="failed to get container status \"7b395c0ccc6f4e0f009f46b86e0a008d6d1e5684c909da03fc98d06f19190e62\": rpc error: code = NotFound desc = could not find container \"7b395c0ccc6f4e0f009f46b86e0a008d6d1e5684c909da03fc98d06f19190e62\": container with ID starting with 7b395c0ccc6f4e0f009f46b86e0a008d6d1e5684c909da03fc98d06f19190e62 not found: ID does not exist" Feb 27 18:34:08 crc kubenswrapper[4700]: I0227 18:34:08.997877 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4" path="/var/lib/kubelet/pods/d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4/volumes" Feb 27 18:34:13 crc kubenswrapper[4700]: I0227 18:34:13.907054 4700 scope.go:117] "RemoveContainer" containerID="bef2cb63a3b1ee53e15e8d40529d1fc26948d802f5478c8b2e25db61e155c8e1" Feb 27 18:34:13 crc kubenswrapper[4700]: I0227 18:34:13.979493 4700 scope.go:117] "RemoveContainer" containerID="eddf47b752176a5dca7ef89b2e9da9f47d2a28693a50fab9e03d35eee675a117" Feb 27 18:34:14 crc kubenswrapper[4700]: I0227 18:34:14.062749 4700 scope.go:117] "RemoveContainer" containerID="99203e60bd2f4332e38617f1a8b60b443dcded5a1ec8b88d77f76aeeb4c0d145" Feb 27 18:34:17 crc kubenswrapper[4700]: I0227 18:34:17.379384 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wqxph/must-gather-6pk65"] Feb 27 18:34:17 crc kubenswrapper[4700]: E0227 18:34:17.380190 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dc264f3-4cbc-4c56-b9f8-4090451d6f8d" containerName="oc" Feb 27 18:34:17 crc kubenswrapper[4700]: I0227 18:34:17.380202 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dc264f3-4cbc-4c56-b9f8-4090451d6f8d" containerName="oc" Feb 27 18:34:17 crc kubenswrapper[4700]: E0227 18:34:17.380227 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4" containerName="registry-server" Feb 27 18:34:17 crc kubenswrapper[4700]: I0227 18:34:17.380234 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4" containerName="registry-server" Feb 27 18:34:17 crc kubenswrapper[4700]: E0227 18:34:17.380254 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4" containerName="extract-utilities" Feb 27 18:34:17 crc kubenswrapper[4700]: I0227 18:34:17.380260 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4" containerName="extract-utilities" Feb 27 18:34:17 crc kubenswrapper[4700]: E0227 18:34:17.380268 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4" containerName="extract-content" Feb 27 18:34:17 crc kubenswrapper[4700]: I0227 18:34:17.380273 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4" containerName="extract-content" Feb 27 18:34:17 crc kubenswrapper[4700]: I0227 18:34:17.380450 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5c5ed20-36ef-46f2-adc5-1c8f3ea314a4" containerName="registry-server" Feb 27 18:34:17 crc kubenswrapper[4700]: I0227 18:34:17.380484 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dc264f3-4cbc-4c56-b9f8-4090451d6f8d" containerName="oc" Feb 27 18:34:17 crc kubenswrapper[4700]: I0227 18:34:17.381482 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wqxph/must-gather-6pk65" Feb 27 18:34:17 crc kubenswrapper[4700]: I0227 18:34:17.383916 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-wqxph"/"openshift-service-ca.crt" Feb 27 18:34:17 crc kubenswrapper[4700]: I0227 18:34:17.384182 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-wqxph"/"default-dockercfg-jr8fm" Feb 27 18:34:17 crc kubenswrapper[4700]: I0227 18:34:17.384317 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-wqxph"/"kube-root-ca.crt" Feb 27 18:34:17 crc kubenswrapper[4700]: I0227 18:34:17.390690 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-wqxph/must-gather-6pk65"] Feb 27 18:34:17 crc kubenswrapper[4700]: I0227 18:34:17.444948 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/489734cd-8c02-4e0b-b300-62a2ce5b6c94-must-gather-output\") pod \"must-gather-6pk65\" (UID: \"489734cd-8c02-4e0b-b300-62a2ce5b6c94\") " pod="openshift-must-gather-wqxph/must-gather-6pk65" Feb 27 18:34:17 crc kubenswrapper[4700]: I0227 18:34:17.445231 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8w67\" (UniqueName: \"kubernetes.io/projected/489734cd-8c02-4e0b-b300-62a2ce5b6c94-kube-api-access-c8w67\") pod \"must-gather-6pk65\" (UID: \"489734cd-8c02-4e0b-b300-62a2ce5b6c94\") " pod="openshift-must-gather-wqxph/must-gather-6pk65" Feb 27 18:34:17 crc kubenswrapper[4700]: I0227 18:34:17.548249 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/489734cd-8c02-4e0b-b300-62a2ce5b6c94-must-gather-output\") pod \"must-gather-6pk65\" (UID: \"489734cd-8c02-4e0b-b300-62a2ce5b6c94\") " pod="openshift-must-gather-wqxph/must-gather-6pk65" Feb 27 18:34:17 crc kubenswrapper[4700]: I0227 18:34:17.548318 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8w67\" (UniqueName: \"kubernetes.io/projected/489734cd-8c02-4e0b-b300-62a2ce5b6c94-kube-api-access-c8w67\") pod \"must-gather-6pk65\" (UID: \"489734cd-8c02-4e0b-b300-62a2ce5b6c94\") " pod="openshift-must-gather-wqxph/must-gather-6pk65" Feb 27 18:34:17 crc kubenswrapper[4700]: I0227 18:34:17.548824 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/489734cd-8c02-4e0b-b300-62a2ce5b6c94-must-gather-output\") pod \"must-gather-6pk65\" (UID: \"489734cd-8c02-4e0b-b300-62a2ce5b6c94\") " pod="openshift-must-gather-wqxph/must-gather-6pk65" Feb 27 18:34:17 crc kubenswrapper[4700]: I0227 18:34:17.569169 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8w67\" (UniqueName: \"kubernetes.io/projected/489734cd-8c02-4e0b-b300-62a2ce5b6c94-kube-api-access-c8w67\") pod \"must-gather-6pk65\" (UID: \"489734cd-8c02-4e0b-b300-62a2ce5b6c94\") " pod="openshift-must-gather-wqxph/must-gather-6pk65" Feb 27 18:34:17 crc kubenswrapper[4700]: I0227 18:34:17.700800 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wqxph/must-gather-6pk65" Feb 27 18:34:17 crc kubenswrapper[4700]: E0227 18:34:17.985330 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:34:18 crc kubenswrapper[4700]: I0227 18:34:18.235079 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-wqxph/must-gather-6pk65"] Feb 27 18:34:18 crc kubenswrapper[4700]: I0227 18:34:18.591900 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wqxph/must-gather-6pk65" event={"ID":"489734cd-8c02-4e0b-b300-62a2ce5b6c94","Type":"ContainerStarted","Data":"d4e6a2713a577415642633b7c2a5b5c1df7cfb06e16d8ff594aad9bb5c272722"} Feb 27 18:34:27 crc kubenswrapper[4700]: I0227 18:34:27.713701 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wqxph/must-gather-6pk65" event={"ID":"489734cd-8c02-4e0b-b300-62a2ce5b6c94","Type":"ContainerStarted","Data":"062b668618606ea03fd04a44e970d2a53effed5ec68456e58361756b89dd4e55"} Feb 27 18:34:27 crc kubenswrapper[4700]: I0227 18:34:27.714980 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wqxph/must-gather-6pk65" event={"ID":"489734cd-8c02-4e0b-b300-62a2ce5b6c94","Type":"ContainerStarted","Data":"70b8519010dd2402205e46f8823f4e36e179cf07706d5a2a05faa1a6cbaa6cc7"} Feb 27 18:34:27 crc kubenswrapper[4700]: I0227 18:34:27.758698 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-wqxph/must-gather-6pk65" podStartSLOduration=2.556587865 podStartE2EDuration="10.758662821s" podCreationTimestamp="2026-02-27 18:34:17 +0000 UTC" firstStartedPulling="2026-02-27 18:34:18.246349574 +0000 UTC m=+5618.231662321" lastFinishedPulling="2026-02-27 18:34:26.44842453 +0000 UTC m=+5626.433737277" observedRunningTime="2026-02-27 18:34:27.736812582 +0000 UTC m=+5627.722125329" watchObservedRunningTime="2026-02-27 18:34:27.758662821 +0000 UTC m=+5627.743975608" Feb 27 18:34:31 crc kubenswrapper[4700]: I0227 18:34:31.714494 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wqxph/crc-debug-z6pvj"] Feb 27 18:34:31 crc kubenswrapper[4700]: I0227 18:34:31.716197 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wqxph/crc-debug-z6pvj" Feb 27 18:34:31 crc kubenswrapper[4700]: I0227 18:34:31.827616 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkdl8\" (UniqueName: \"kubernetes.io/projected/d2e295e4-5458-4b9c-b33f-c4437a708090-kube-api-access-wkdl8\") pod \"crc-debug-z6pvj\" (UID: \"d2e295e4-5458-4b9c-b33f-c4437a708090\") " pod="openshift-must-gather-wqxph/crc-debug-z6pvj" Feb 27 18:34:31 crc kubenswrapper[4700]: I0227 18:34:31.827898 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d2e295e4-5458-4b9c-b33f-c4437a708090-host\") pod \"crc-debug-z6pvj\" (UID: \"d2e295e4-5458-4b9c-b33f-c4437a708090\") " pod="openshift-must-gather-wqxph/crc-debug-z6pvj" Feb 27 18:34:31 crc kubenswrapper[4700]: I0227 18:34:31.929634 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkdl8\" (UniqueName: \"kubernetes.io/projected/d2e295e4-5458-4b9c-b33f-c4437a708090-kube-api-access-wkdl8\") pod \"crc-debug-z6pvj\" (UID: \"d2e295e4-5458-4b9c-b33f-c4437a708090\") " pod="openshift-must-gather-wqxph/crc-debug-z6pvj" Feb 27 18:34:31 crc kubenswrapper[4700]: I0227 18:34:31.929751 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d2e295e4-5458-4b9c-b33f-c4437a708090-host\") pod \"crc-debug-z6pvj\" (UID: \"d2e295e4-5458-4b9c-b33f-c4437a708090\") " pod="openshift-must-gather-wqxph/crc-debug-z6pvj" Feb 27 18:34:31 crc kubenswrapper[4700]: I0227 18:34:31.929880 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d2e295e4-5458-4b9c-b33f-c4437a708090-host\") pod \"crc-debug-z6pvj\" (UID: \"d2e295e4-5458-4b9c-b33f-c4437a708090\") " pod="openshift-must-gather-wqxph/crc-debug-z6pvj" Feb 27 18:34:31 crc kubenswrapper[4700]: I0227 18:34:31.964281 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkdl8\" (UniqueName: \"kubernetes.io/projected/d2e295e4-5458-4b9c-b33f-c4437a708090-kube-api-access-wkdl8\") pod \"crc-debug-z6pvj\" (UID: \"d2e295e4-5458-4b9c-b33f-c4437a708090\") " pod="openshift-must-gather-wqxph/crc-debug-z6pvj" Feb 27 18:34:32 crc kubenswrapper[4700]: I0227 18:34:32.037474 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wqxph/crc-debug-z6pvj" Feb 27 18:34:32 crc kubenswrapper[4700]: W0227 18:34:32.066658 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2e295e4_5458_4b9c_b33f_c4437a708090.slice/crio-4320f314603531de2c50968e4a69d372d076bec4f08e94af1fc745919504db57 WatchSource:0}: Error finding container 4320f314603531de2c50968e4a69d372d076bec4f08e94af1fc745919504db57: Status 404 returned error can't find the container with id 4320f314603531de2c50968e4a69d372d076bec4f08e94af1fc745919504db57 Feb 27 18:34:32 crc kubenswrapper[4700]: I0227 18:34:32.794702 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wqxph/crc-debug-z6pvj" event={"ID":"d2e295e4-5458-4b9c-b33f-c4437a708090","Type":"ContainerStarted","Data":"4320f314603531de2c50968e4a69d372d076bec4f08e94af1fc745919504db57"} Feb 27 18:34:32 crc kubenswrapper[4700]: E0227 18:34:32.983294 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:34:36 crc kubenswrapper[4700]: I0227 18:34:36.410789 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 18:34:36 crc kubenswrapper[4700]: I0227 18:34:36.411616 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 18:34:36 crc kubenswrapper[4700]: I0227 18:34:36.411695 4700 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" Feb 27 18:34:36 crc kubenswrapper[4700]: I0227 18:34:36.412961 4700 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fc0ec093303e51bd2ed80860af9a81018e732a7449b44faf6a9d28fb0570f2b0"} pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 18:34:36 crc kubenswrapper[4700]: I0227 18:34:36.413040 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" containerID="cri-o://fc0ec093303e51bd2ed80860af9a81018e732a7449b44faf6a9d28fb0570f2b0" gracePeriod=600 Feb 27 18:34:36 crc kubenswrapper[4700]: I0227 18:34:36.837071 4700 generic.go:334] "Generic (PLEG): container finished" podID="84b45511-b94c-479f-98db-fd2c4eceec46" containerID="fc0ec093303e51bd2ed80860af9a81018e732a7449b44faf6a9d28fb0570f2b0" exitCode=0 Feb 27 18:34:36 crc kubenswrapper[4700]: I0227 18:34:36.837149 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerDied","Data":"fc0ec093303e51bd2ed80860af9a81018e732a7449b44faf6a9d28fb0570f2b0"} Feb 27 18:34:36 crc kubenswrapper[4700]: I0227 18:34:36.837609 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerStarted","Data":"c7ad608c50ddee2aac4992df7a6ccaaeda7c737e2a0d2fad5282b4482eac9e50"} Feb 27 18:34:36 crc kubenswrapper[4700]: I0227 18:34:36.837649 4700 scope.go:117] "RemoveContainer" containerID="b3fdfb06db2c06428e3b3e96ef9804be97292a50a8f04f891caec5e729731adf" Feb 27 18:34:43 crc kubenswrapper[4700]: E0227 18:34:43.984312 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:34:44 crc kubenswrapper[4700]: I0227 18:34:44.932792 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wqxph/crc-debug-z6pvj" event={"ID":"d2e295e4-5458-4b9c-b33f-c4437a708090","Type":"ContainerStarted","Data":"4e352c90a326d1b43366a561782d2e0128653a7eb0793a99ddc14f5a8491618a"} Feb 27 18:34:44 crc kubenswrapper[4700]: I0227 18:34:44.956020 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-wqxph/crc-debug-z6pvj" podStartSLOduration=2.254350016 podStartE2EDuration="13.955999786s" podCreationTimestamp="2026-02-27 18:34:31 +0000 UTC" firstStartedPulling="2026-02-27 18:34:32.068871503 +0000 UTC m=+5632.054184250" lastFinishedPulling="2026-02-27 18:34:43.770521273 +0000 UTC m=+5643.755834020" observedRunningTime="2026-02-27 18:34:44.947079153 +0000 UTC m=+5644.932391910" watchObservedRunningTime="2026-02-27 18:34:44.955999786 +0000 UTC m=+5644.941312543" Feb 27 18:34:56 crc kubenswrapper[4700]: E0227 18:34:56.912551 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:35:11 crc kubenswrapper[4700]: E0227 18:35:11.041044 4700 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 27 18:35:11 crc kubenswrapper[4700]: E0227 18:35:11.041555 4700 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 27 18:35:11 crc kubenswrapper[4700]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 27 18:35:11 crc kubenswrapper[4700]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qkjfz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29536934-rw77f_openshift-infra(caa4e57d-b02d-4d65-944d-451b8294f884): ErrImagePull: copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error) Feb 27 18:35:11 crc kubenswrapper[4700]: > logger="UnhandledError" Feb 27 18:35:11 crc kubenswrapper[4700]: E0227 18:35:11.042685 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: reading signatures: reading signature from https://registry.redhat.io/containers/sigstore/openshift4/ose-cli@sha256=69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9/signature-7: status 500 (Internal Server Error)\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:35:21 crc kubenswrapper[4700]: E0227 18:35:21.983188 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:35:31 crc kubenswrapper[4700]: I0227 18:35:31.424050 4700 generic.go:334] "Generic (PLEG): container finished" podID="d2e295e4-5458-4b9c-b33f-c4437a708090" containerID="4e352c90a326d1b43366a561782d2e0128653a7eb0793a99ddc14f5a8491618a" exitCode=0 Feb 27 18:35:31 crc kubenswrapper[4700]: I0227 18:35:31.424123 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wqxph/crc-debug-z6pvj" event={"ID":"d2e295e4-5458-4b9c-b33f-c4437a708090","Type":"ContainerDied","Data":"4e352c90a326d1b43366a561782d2e0128653a7eb0793a99ddc14f5a8491618a"} Feb 27 18:35:32 crc kubenswrapper[4700]: I0227 18:35:32.569345 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wqxph/crc-debug-z6pvj" Feb 27 18:35:32 crc kubenswrapper[4700]: I0227 18:35:32.630213 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wqxph/crc-debug-z6pvj"] Feb 27 18:35:32 crc kubenswrapper[4700]: I0227 18:35:32.640008 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wqxph/crc-debug-z6pvj"] Feb 27 18:35:32 crc kubenswrapper[4700]: I0227 18:35:32.722598 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d2e295e4-5458-4b9c-b33f-c4437a708090-host\") pod \"d2e295e4-5458-4b9c-b33f-c4437a708090\" (UID: \"d2e295e4-5458-4b9c-b33f-c4437a708090\") " Feb 27 18:35:32 crc kubenswrapper[4700]: I0227 18:35:32.722806 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d2e295e4-5458-4b9c-b33f-c4437a708090-host" (OuterVolumeSpecName: "host") pod "d2e295e4-5458-4b9c-b33f-c4437a708090" (UID: "d2e295e4-5458-4b9c-b33f-c4437a708090"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 18:35:32 crc kubenswrapper[4700]: I0227 18:35:32.723080 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wkdl8\" (UniqueName: \"kubernetes.io/projected/d2e295e4-5458-4b9c-b33f-c4437a708090-kube-api-access-wkdl8\") pod \"d2e295e4-5458-4b9c-b33f-c4437a708090\" (UID: \"d2e295e4-5458-4b9c-b33f-c4437a708090\") " Feb 27 18:35:32 crc kubenswrapper[4700]: I0227 18:35:32.724067 4700 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d2e295e4-5458-4b9c-b33f-c4437a708090-host\") on node \"crc\" DevicePath \"\"" Feb 27 18:35:33 crc kubenswrapper[4700]: I0227 18:35:33.450418 4700 scope.go:117] "RemoveContainer" containerID="4e352c90a326d1b43366a561782d2e0128653a7eb0793a99ddc14f5a8491618a" Feb 27 18:35:33 crc kubenswrapper[4700]: I0227 18:35:33.450449 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wqxph/crc-debug-z6pvj" Feb 27 18:35:33 crc kubenswrapper[4700]: I0227 18:35:33.550996 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2e295e4-5458-4b9c-b33f-c4437a708090-kube-api-access-wkdl8" (OuterVolumeSpecName: "kube-api-access-wkdl8") pod "d2e295e4-5458-4b9c-b33f-c4437a708090" (UID: "d2e295e4-5458-4b9c-b33f-c4437a708090"). InnerVolumeSpecName "kube-api-access-wkdl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:35:33 crc kubenswrapper[4700]: I0227 18:35:33.652526 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wkdl8\" (UniqueName: \"kubernetes.io/projected/d2e295e4-5458-4b9c-b33f-c4437a708090-kube-api-access-wkdl8\") on node \"crc\" DevicePath \"\"" Feb 27 18:35:34 crc kubenswrapper[4700]: E0227 18:35:33.983559 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:35:34 crc kubenswrapper[4700]: I0227 18:35:34.896865 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wqxph/crc-debug-6b2pn"] Feb 27 18:35:34 crc kubenswrapper[4700]: E0227 18:35:34.897322 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2e295e4-5458-4b9c-b33f-c4437a708090" containerName="container-00" Feb 27 18:35:34 crc kubenswrapper[4700]: I0227 18:35:34.897335 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2e295e4-5458-4b9c-b33f-c4437a708090" containerName="container-00" Feb 27 18:35:34 crc kubenswrapper[4700]: I0227 18:35:34.898417 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2e295e4-5458-4b9c-b33f-c4437a708090" containerName="container-00" Feb 27 18:35:34 crc kubenswrapper[4700]: I0227 18:35:34.899162 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wqxph/crc-debug-6b2pn" Feb 27 18:35:34 crc kubenswrapper[4700]: I0227 18:35:34.982908 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5nkq\" (UniqueName: \"kubernetes.io/projected/185698a1-44e6-4cc6-a648-b756b8f085ef-kube-api-access-b5nkq\") pod \"crc-debug-6b2pn\" (UID: \"185698a1-44e6-4cc6-a648-b756b8f085ef\") " pod="openshift-must-gather-wqxph/crc-debug-6b2pn" Feb 27 18:35:34 crc kubenswrapper[4700]: I0227 18:35:34.983246 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/185698a1-44e6-4cc6-a648-b756b8f085ef-host\") pod \"crc-debug-6b2pn\" (UID: \"185698a1-44e6-4cc6-a648-b756b8f085ef\") " pod="openshift-must-gather-wqxph/crc-debug-6b2pn" Feb 27 18:35:34 crc kubenswrapper[4700]: I0227 18:35:34.995039 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2e295e4-5458-4b9c-b33f-c4437a708090" path="/var/lib/kubelet/pods/d2e295e4-5458-4b9c-b33f-c4437a708090/volumes" Feb 27 18:35:35 crc kubenswrapper[4700]: I0227 18:35:35.085188 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5nkq\" (UniqueName: \"kubernetes.io/projected/185698a1-44e6-4cc6-a648-b756b8f085ef-kube-api-access-b5nkq\") pod \"crc-debug-6b2pn\" (UID: \"185698a1-44e6-4cc6-a648-b756b8f085ef\") " pod="openshift-must-gather-wqxph/crc-debug-6b2pn" Feb 27 18:35:35 crc kubenswrapper[4700]: I0227 18:35:35.085428 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/185698a1-44e6-4cc6-a648-b756b8f085ef-host\") pod \"crc-debug-6b2pn\" (UID: \"185698a1-44e6-4cc6-a648-b756b8f085ef\") " pod="openshift-must-gather-wqxph/crc-debug-6b2pn" Feb 27 18:35:35 crc kubenswrapper[4700]: I0227 18:35:35.085968 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/185698a1-44e6-4cc6-a648-b756b8f085ef-host\") pod \"crc-debug-6b2pn\" (UID: \"185698a1-44e6-4cc6-a648-b756b8f085ef\") " pod="openshift-must-gather-wqxph/crc-debug-6b2pn" Feb 27 18:35:35 crc kubenswrapper[4700]: I0227 18:35:35.119242 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5nkq\" (UniqueName: \"kubernetes.io/projected/185698a1-44e6-4cc6-a648-b756b8f085ef-kube-api-access-b5nkq\") pod \"crc-debug-6b2pn\" (UID: \"185698a1-44e6-4cc6-a648-b756b8f085ef\") " pod="openshift-must-gather-wqxph/crc-debug-6b2pn" Feb 27 18:35:35 crc kubenswrapper[4700]: I0227 18:35:35.225535 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wqxph/crc-debug-6b2pn" Feb 27 18:35:35 crc kubenswrapper[4700]: I0227 18:35:35.471659 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wqxph/crc-debug-6b2pn" event={"ID":"185698a1-44e6-4cc6-a648-b756b8f085ef","Type":"ContainerStarted","Data":"5beb081530839d4c35ed181458f4b692d9ceb5aaec0fd07b4f7f7fa59e327201"} Feb 27 18:35:36 crc kubenswrapper[4700]: I0227 18:35:36.482276 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wqxph/crc-debug-6b2pn" event={"ID":"185698a1-44e6-4cc6-a648-b756b8f085ef","Type":"ContainerStarted","Data":"21f02e929bbc8f2c11dc0adba14f1cebe7881738dd247872b8935fd9af3c5e56"} Feb 27 18:35:37 crc kubenswrapper[4700]: I0227 18:35:37.513357 4700 generic.go:334] "Generic (PLEG): container finished" podID="185698a1-44e6-4cc6-a648-b756b8f085ef" containerID="21f02e929bbc8f2c11dc0adba14f1cebe7881738dd247872b8935fd9af3c5e56" exitCode=0 Feb 27 18:35:37 crc kubenswrapper[4700]: I0227 18:35:37.513758 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wqxph/crc-debug-6b2pn" event={"ID":"185698a1-44e6-4cc6-a648-b756b8f085ef","Type":"ContainerDied","Data":"21f02e929bbc8f2c11dc0adba14f1cebe7881738dd247872b8935fd9af3c5e56"} Feb 27 18:35:38 crc kubenswrapper[4700]: I0227 18:35:38.617980 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wqxph/crc-debug-6b2pn" Feb 27 18:35:38 crc kubenswrapper[4700]: I0227 18:35:38.766486 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5nkq\" (UniqueName: \"kubernetes.io/projected/185698a1-44e6-4cc6-a648-b756b8f085ef-kube-api-access-b5nkq\") pod \"185698a1-44e6-4cc6-a648-b756b8f085ef\" (UID: \"185698a1-44e6-4cc6-a648-b756b8f085ef\") " Feb 27 18:35:38 crc kubenswrapper[4700]: I0227 18:35:38.766878 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/185698a1-44e6-4cc6-a648-b756b8f085ef-host\") pod \"185698a1-44e6-4cc6-a648-b756b8f085ef\" (UID: \"185698a1-44e6-4cc6-a648-b756b8f085ef\") " Feb 27 18:35:38 crc kubenswrapper[4700]: I0227 18:35:38.767534 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/185698a1-44e6-4cc6-a648-b756b8f085ef-host" (OuterVolumeSpecName: "host") pod "185698a1-44e6-4cc6-a648-b756b8f085ef" (UID: "185698a1-44e6-4cc6-a648-b756b8f085ef"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 18:35:38 crc kubenswrapper[4700]: I0227 18:35:38.772739 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/185698a1-44e6-4cc6-a648-b756b8f085ef-kube-api-access-b5nkq" (OuterVolumeSpecName: "kube-api-access-b5nkq") pod "185698a1-44e6-4cc6-a648-b756b8f085ef" (UID: "185698a1-44e6-4cc6-a648-b756b8f085ef"). InnerVolumeSpecName "kube-api-access-b5nkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:35:38 crc kubenswrapper[4700]: I0227 18:35:38.843482 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wqxph/crc-debug-6b2pn"] Feb 27 18:35:38 crc kubenswrapper[4700]: I0227 18:35:38.859129 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wqxph/crc-debug-6b2pn"] Feb 27 18:35:38 crc kubenswrapper[4700]: I0227 18:35:38.869751 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5nkq\" (UniqueName: \"kubernetes.io/projected/185698a1-44e6-4cc6-a648-b756b8f085ef-kube-api-access-b5nkq\") on node \"crc\" DevicePath \"\"" Feb 27 18:35:38 crc kubenswrapper[4700]: I0227 18:35:38.869789 4700 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/185698a1-44e6-4cc6-a648-b756b8f085ef-host\") on node \"crc\" DevicePath \"\"" Feb 27 18:35:38 crc kubenswrapper[4700]: I0227 18:35:38.994774 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="185698a1-44e6-4cc6-a648-b756b8f085ef" path="/var/lib/kubelet/pods/185698a1-44e6-4cc6-a648-b756b8f085ef/volumes" Feb 27 18:35:39 crc kubenswrapper[4700]: I0227 18:35:39.535937 4700 scope.go:117] "RemoveContainer" containerID="21f02e929bbc8f2c11dc0adba14f1cebe7881738dd247872b8935fd9af3c5e56" Feb 27 18:35:39 crc kubenswrapper[4700]: I0227 18:35:39.536186 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wqxph/crc-debug-6b2pn" Feb 27 18:35:40 crc kubenswrapper[4700]: I0227 18:35:40.023141 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wqxph/crc-debug-6rfk4"] Feb 27 18:35:40 crc kubenswrapper[4700]: E0227 18:35:40.024219 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="185698a1-44e6-4cc6-a648-b756b8f085ef" containerName="container-00" Feb 27 18:35:40 crc kubenswrapper[4700]: I0227 18:35:40.024239 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="185698a1-44e6-4cc6-a648-b756b8f085ef" containerName="container-00" Feb 27 18:35:40 crc kubenswrapper[4700]: I0227 18:35:40.024613 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="185698a1-44e6-4cc6-a648-b756b8f085ef" containerName="container-00" Feb 27 18:35:40 crc kubenswrapper[4700]: I0227 18:35:40.025604 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wqxph/crc-debug-6rfk4" Feb 27 18:35:40 crc kubenswrapper[4700]: I0227 18:35:40.096815 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brvjp\" (UniqueName: \"kubernetes.io/projected/ee9fee67-9166-4ec5-8a0b-71ee94b804ba-kube-api-access-brvjp\") pod \"crc-debug-6rfk4\" (UID: \"ee9fee67-9166-4ec5-8a0b-71ee94b804ba\") " pod="openshift-must-gather-wqxph/crc-debug-6rfk4" Feb 27 18:35:40 crc kubenswrapper[4700]: I0227 18:35:40.097037 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ee9fee67-9166-4ec5-8a0b-71ee94b804ba-host\") pod \"crc-debug-6rfk4\" (UID: \"ee9fee67-9166-4ec5-8a0b-71ee94b804ba\") " pod="openshift-must-gather-wqxph/crc-debug-6rfk4" Feb 27 18:35:40 crc kubenswrapper[4700]: I0227 18:35:40.199201 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brvjp\" (UniqueName: \"kubernetes.io/projected/ee9fee67-9166-4ec5-8a0b-71ee94b804ba-kube-api-access-brvjp\") pod \"crc-debug-6rfk4\" (UID: \"ee9fee67-9166-4ec5-8a0b-71ee94b804ba\") " pod="openshift-must-gather-wqxph/crc-debug-6rfk4" Feb 27 18:35:40 crc kubenswrapper[4700]: I0227 18:35:40.199341 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ee9fee67-9166-4ec5-8a0b-71ee94b804ba-host\") pod \"crc-debug-6rfk4\" (UID: \"ee9fee67-9166-4ec5-8a0b-71ee94b804ba\") " pod="openshift-must-gather-wqxph/crc-debug-6rfk4" Feb 27 18:35:40 crc kubenswrapper[4700]: I0227 18:35:40.199573 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ee9fee67-9166-4ec5-8a0b-71ee94b804ba-host\") pod \"crc-debug-6rfk4\" (UID: \"ee9fee67-9166-4ec5-8a0b-71ee94b804ba\") " pod="openshift-must-gather-wqxph/crc-debug-6rfk4" Feb 27 18:35:40 crc kubenswrapper[4700]: I0227 18:35:40.241303 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brvjp\" (UniqueName: \"kubernetes.io/projected/ee9fee67-9166-4ec5-8a0b-71ee94b804ba-kube-api-access-brvjp\") pod \"crc-debug-6rfk4\" (UID: \"ee9fee67-9166-4ec5-8a0b-71ee94b804ba\") " pod="openshift-must-gather-wqxph/crc-debug-6rfk4" Feb 27 18:35:40 crc kubenswrapper[4700]: I0227 18:35:40.351448 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wqxph/crc-debug-6rfk4" Feb 27 18:35:40 crc kubenswrapper[4700]: W0227 18:35:40.398873 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee9fee67_9166_4ec5_8a0b_71ee94b804ba.slice/crio-89b5b95e5f0d0e5d31482ffe3a516b63cc9e9986a1c50ccbfdf61db50a4d2f9e WatchSource:0}: Error finding container 89b5b95e5f0d0e5d31482ffe3a516b63cc9e9986a1c50ccbfdf61db50a4d2f9e: Status 404 returned error can't find the container with id 89b5b95e5f0d0e5d31482ffe3a516b63cc9e9986a1c50ccbfdf61db50a4d2f9e Feb 27 18:35:40 crc kubenswrapper[4700]: I0227 18:35:40.548394 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wqxph/crc-debug-6rfk4" event={"ID":"ee9fee67-9166-4ec5-8a0b-71ee94b804ba","Type":"ContainerStarted","Data":"89b5b95e5f0d0e5d31482ffe3a516b63cc9e9986a1c50ccbfdf61db50a4d2f9e"} Feb 27 18:35:41 crc kubenswrapper[4700]: I0227 18:35:41.563567 4700 generic.go:334] "Generic (PLEG): container finished" podID="ee9fee67-9166-4ec5-8a0b-71ee94b804ba" containerID="04a812c6125aee558c02b212cfcf1357f76dd59af499d73c44b80a0e7190a10c" exitCode=0 Feb 27 18:35:41 crc kubenswrapper[4700]: I0227 18:35:41.563643 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wqxph/crc-debug-6rfk4" event={"ID":"ee9fee67-9166-4ec5-8a0b-71ee94b804ba","Type":"ContainerDied","Data":"04a812c6125aee558c02b212cfcf1357f76dd59af499d73c44b80a0e7190a10c"} Feb 27 18:35:41 crc kubenswrapper[4700]: I0227 18:35:41.624989 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wqxph/crc-debug-6rfk4"] Feb 27 18:35:41 crc kubenswrapper[4700]: I0227 18:35:41.636012 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wqxph/crc-debug-6rfk4"] Feb 27 18:35:42 crc kubenswrapper[4700]: I0227 18:35:42.700911 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wqxph/crc-debug-6rfk4" Feb 27 18:35:42 crc kubenswrapper[4700]: I0227 18:35:42.767746 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-brvjp\" (UniqueName: \"kubernetes.io/projected/ee9fee67-9166-4ec5-8a0b-71ee94b804ba-kube-api-access-brvjp\") pod \"ee9fee67-9166-4ec5-8a0b-71ee94b804ba\" (UID: \"ee9fee67-9166-4ec5-8a0b-71ee94b804ba\") " Feb 27 18:35:42 crc kubenswrapper[4700]: I0227 18:35:42.767941 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ee9fee67-9166-4ec5-8a0b-71ee94b804ba-host\") pod \"ee9fee67-9166-4ec5-8a0b-71ee94b804ba\" (UID: \"ee9fee67-9166-4ec5-8a0b-71ee94b804ba\") " Feb 27 18:35:42 crc kubenswrapper[4700]: I0227 18:35:42.768924 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee9fee67-9166-4ec5-8a0b-71ee94b804ba-host" (OuterVolumeSpecName: "host") pod "ee9fee67-9166-4ec5-8a0b-71ee94b804ba" (UID: "ee9fee67-9166-4ec5-8a0b-71ee94b804ba"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 18:35:42 crc kubenswrapper[4700]: I0227 18:35:42.774804 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee9fee67-9166-4ec5-8a0b-71ee94b804ba-kube-api-access-brvjp" (OuterVolumeSpecName: "kube-api-access-brvjp") pod "ee9fee67-9166-4ec5-8a0b-71ee94b804ba" (UID: "ee9fee67-9166-4ec5-8a0b-71ee94b804ba"). InnerVolumeSpecName "kube-api-access-brvjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:35:42 crc kubenswrapper[4700]: I0227 18:35:42.871741 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-brvjp\" (UniqueName: \"kubernetes.io/projected/ee9fee67-9166-4ec5-8a0b-71ee94b804ba-kube-api-access-brvjp\") on node \"crc\" DevicePath \"\"" Feb 27 18:35:42 crc kubenswrapper[4700]: I0227 18:35:42.871793 4700 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ee9fee67-9166-4ec5-8a0b-71ee94b804ba-host\") on node \"crc\" DevicePath \"\"" Feb 27 18:35:42 crc kubenswrapper[4700]: I0227 18:35:42.998672 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee9fee67-9166-4ec5-8a0b-71ee94b804ba" path="/var/lib/kubelet/pods/ee9fee67-9166-4ec5-8a0b-71ee94b804ba/volumes" Feb 27 18:35:43 crc kubenswrapper[4700]: I0227 18:35:43.583076 4700 scope.go:117] "RemoveContainer" containerID="04a812c6125aee558c02b212cfcf1357f76dd59af499d73c44b80a0e7190a10c" Feb 27 18:35:43 crc kubenswrapper[4700]: I0227 18:35:43.583107 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wqxph/crc-debug-6rfk4" Feb 27 18:35:47 crc kubenswrapper[4700]: E0227 18:35:47.983819 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:36:00 crc kubenswrapper[4700]: I0227 18:36:00.218645 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536956-sz7xx"] Feb 27 18:36:00 crc kubenswrapper[4700]: E0227 18:36:00.219954 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee9fee67-9166-4ec5-8a0b-71ee94b804ba" containerName="container-00" Feb 27 18:36:00 crc kubenswrapper[4700]: I0227 18:36:00.219977 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee9fee67-9166-4ec5-8a0b-71ee94b804ba" containerName="container-00" Feb 27 18:36:00 crc kubenswrapper[4700]: I0227 18:36:00.220387 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee9fee67-9166-4ec5-8a0b-71ee94b804ba" containerName="container-00" Feb 27 18:36:00 crc kubenswrapper[4700]: I0227 18:36:00.222627 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536956-sz7xx" Feb 27 18:36:00 crc kubenswrapper[4700]: I0227 18:36:00.256545 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536956-sz7xx"] Feb 27 18:36:00 crc kubenswrapper[4700]: I0227 18:36:00.294737 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x59md\" (UniqueName: \"kubernetes.io/projected/074ce5d9-7bd8-4205-9c43-24c9cf22ea32-kube-api-access-x59md\") pod \"auto-csr-approver-29536956-sz7xx\" (UID: \"074ce5d9-7bd8-4205-9c43-24c9cf22ea32\") " pod="openshift-infra/auto-csr-approver-29536956-sz7xx" Feb 27 18:36:00 crc kubenswrapper[4700]: I0227 18:36:00.397957 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x59md\" (UniqueName: \"kubernetes.io/projected/074ce5d9-7bd8-4205-9c43-24c9cf22ea32-kube-api-access-x59md\") pod \"auto-csr-approver-29536956-sz7xx\" (UID: \"074ce5d9-7bd8-4205-9c43-24c9cf22ea32\") " pod="openshift-infra/auto-csr-approver-29536956-sz7xx" Feb 27 18:36:00 crc kubenswrapper[4700]: I0227 18:36:00.418179 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x59md\" (UniqueName: \"kubernetes.io/projected/074ce5d9-7bd8-4205-9c43-24c9cf22ea32-kube-api-access-x59md\") pod \"auto-csr-approver-29536956-sz7xx\" (UID: \"074ce5d9-7bd8-4205-9c43-24c9cf22ea32\") " pod="openshift-infra/auto-csr-approver-29536956-sz7xx" Feb 27 18:36:00 crc kubenswrapper[4700]: I0227 18:36:00.552095 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536956-sz7xx" Feb 27 18:36:01 crc kubenswrapper[4700]: I0227 18:36:01.062738 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536956-sz7xx"] Feb 27 18:36:01 crc kubenswrapper[4700]: I0227 18:36:01.789405 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536956-sz7xx" event={"ID":"074ce5d9-7bd8-4205-9c43-24c9cf22ea32","Type":"ContainerStarted","Data":"2edf8e841ec7d7d28b6786847ca76725af08bb3918b2c2c5d5c2455d10c887e2"} Feb 27 18:36:02 crc kubenswrapper[4700]: I0227 18:36:02.798624 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536956-sz7xx" event={"ID":"074ce5d9-7bd8-4205-9c43-24c9cf22ea32","Type":"ContainerStarted","Data":"d487919697c779b8b8845178b947df4809622ee74a2201b36d2cf94c70270ba3"} Feb 27 18:36:02 crc kubenswrapper[4700]: I0227 18:36:02.814278 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536956-sz7xx" podStartSLOduration=1.667624316 podStartE2EDuration="2.814261166s" podCreationTimestamp="2026-02-27 18:36:00 +0000 UTC" firstStartedPulling="2026-02-27 18:36:01.071602188 +0000 UTC m=+5721.056914935" lastFinishedPulling="2026-02-27 18:36:02.218239028 +0000 UTC m=+5722.203551785" observedRunningTime="2026-02-27 18:36:02.812349986 +0000 UTC m=+5722.797662733" watchObservedRunningTime="2026-02-27 18:36:02.814261166 +0000 UTC m=+5722.799573913" Feb 27 18:36:02 crc kubenswrapper[4700]: E0227 18:36:02.982625 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:36:03 crc kubenswrapper[4700]: I0227 18:36:03.807963 4700 generic.go:334] "Generic (PLEG): container finished" podID="074ce5d9-7bd8-4205-9c43-24c9cf22ea32" containerID="d487919697c779b8b8845178b947df4809622ee74a2201b36d2cf94c70270ba3" exitCode=0 Feb 27 18:36:03 crc kubenswrapper[4700]: I0227 18:36:03.808001 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536956-sz7xx" event={"ID":"074ce5d9-7bd8-4205-9c43-24c9cf22ea32","Type":"ContainerDied","Data":"d487919697c779b8b8845178b947df4809622ee74a2201b36d2cf94c70270ba3"} Feb 27 18:36:05 crc kubenswrapper[4700]: I0227 18:36:05.256979 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536956-sz7xx" Feb 27 18:36:05 crc kubenswrapper[4700]: I0227 18:36:05.415052 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x59md\" (UniqueName: \"kubernetes.io/projected/074ce5d9-7bd8-4205-9c43-24c9cf22ea32-kube-api-access-x59md\") pod \"074ce5d9-7bd8-4205-9c43-24c9cf22ea32\" (UID: \"074ce5d9-7bd8-4205-9c43-24c9cf22ea32\") " Feb 27 18:36:05 crc kubenswrapper[4700]: I0227 18:36:05.429622 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/074ce5d9-7bd8-4205-9c43-24c9cf22ea32-kube-api-access-x59md" (OuterVolumeSpecName: "kube-api-access-x59md") pod "074ce5d9-7bd8-4205-9c43-24c9cf22ea32" (UID: "074ce5d9-7bd8-4205-9c43-24c9cf22ea32"). InnerVolumeSpecName "kube-api-access-x59md". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:36:05 crc kubenswrapper[4700]: I0227 18:36:05.517273 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x59md\" (UniqueName: \"kubernetes.io/projected/074ce5d9-7bd8-4205-9c43-24c9cf22ea32-kube-api-access-x59md\") on node \"crc\" DevicePath \"\"" Feb 27 18:36:05 crc kubenswrapper[4700]: I0227 18:36:05.825202 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536956-sz7xx" event={"ID":"074ce5d9-7bd8-4205-9c43-24c9cf22ea32","Type":"ContainerDied","Data":"2edf8e841ec7d7d28b6786847ca76725af08bb3918b2c2c5d5c2455d10c887e2"} Feb 27 18:36:05 crc kubenswrapper[4700]: I0227 18:36:05.825236 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536956-sz7xx" Feb 27 18:36:05 crc kubenswrapper[4700]: I0227 18:36:05.825237 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2edf8e841ec7d7d28b6786847ca76725af08bb3918b2c2c5d5c2455d10c887e2" Feb 27 18:36:05 crc kubenswrapper[4700]: I0227 18:36:05.923495 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536950-458j5"] Feb 27 18:36:05 crc kubenswrapper[4700]: I0227 18:36:05.943044 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536950-458j5"] Feb 27 18:36:06 crc kubenswrapper[4700]: I0227 18:36:06.992543 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74489fab-6aca-4a3f-9303-04a2b82f1fde" path="/var/lib/kubelet/pods/74489fab-6aca-4a3f-9303-04a2b82f1fde/volumes" Feb 27 18:36:09 crc kubenswrapper[4700]: I0227 18:36:09.749264 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="f2325e29-9f4c-438f-9400-e643e2494808" containerName="galera" probeResult="failure" output="command timed out" Feb 27 18:36:09 crc kubenswrapper[4700]: I0227 18:36:09.749438 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="f2325e29-9f4c-438f-9400-e643e2494808" containerName="galera" probeResult="failure" output="command timed out" Feb 27 18:36:14 crc kubenswrapper[4700]: E0227 18:36:14.984226 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:36:22 crc kubenswrapper[4700]: I0227 18:36:22.875073 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6cbb9d98-xkx97_a9b30341-3783-4498-95ac-0f303a599cce/barbican-api-log/0.log" Feb 27 18:36:22 crc kubenswrapper[4700]: I0227 18:36:22.884024 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6cbb9d98-xkx97_a9b30341-3783-4498-95ac-0f303a599cce/barbican-api/0.log" Feb 27 18:36:23 crc kubenswrapper[4700]: I0227 18:36:23.061751 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7f87979786-hh2zw_3b6c33fa-4855-4b22-a634-cbc739ec0580/barbican-keystone-listener/0.log" Feb 27 18:36:23 crc kubenswrapper[4700]: I0227 18:36:23.122763 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7f87979786-hh2zw_3b6c33fa-4855-4b22-a634-cbc739ec0580/barbican-keystone-listener-log/0.log" Feb 27 18:36:23 crc kubenswrapper[4700]: I0227 18:36:23.139509 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-756b7f868f-z7qr7_1311d698-8af2-4960-9790-f470a67f8457/barbican-worker/0.log" Feb 27 18:36:23 crc kubenswrapper[4700]: I0227 18:36:23.248487 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-756b7f868f-z7qr7_1311d698-8af2-4960-9790-f470a67f8457/barbican-worker-log/0.log" Feb 27 18:36:23 crc kubenswrapper[4700]: I0227 18:36:23.371548 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-5f27j_583579de-19a5-4728-97f9-37398ce4a452/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 18:36:23 crc kubenswrapper[4700]: I0227 18:36:23.526309 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a50deb6-9983-40f1-bf2c-f8ce7584e809/ceilometer-central-agent/0.log" Feb 27 18:36:23 crc kubenswrapper[4700]: I0227 18:36:23.565724 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a50deb6-9983-40f1-bf2c-f8ce7584e809/proxy-httpd/0.log" Feb 27 18:36:23 crc kubenswrapper[4700]: I0227 18:36:23.659363 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a50deb6-9983-40f1-bf2c-f8ce7584e809/ceilometer-notification-agent/0.log" Feb 27 18:36:23 crc kubenswrapper[4700]: I0227 18:36:23.833049 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a50deb6-9983-40f1-bf2c-f8ce7584e809/sg-core/0.log" Feb 27 18:36:23 crc kubenswrapper[4700]: I0227 18:36:23.956284 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_d9080fd5-c068-49f1-8167-e3843d34d471/cinder-api-log/0.log" Feb 27 18:36:24 crc kubenswrapper[4700]: I0227 18:36:24.261161 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_d9080fd5-c068-49f1-8167-e3843d34d471/cinder-api/0.log" Feb 27 18:36:24 crc kubenswrapper[4700]: I0227 18:36:24.720713 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_d9569ae4-5bf3-4720-bc48-0a1b735876ea/cinder-scheduler/0.log" Feb 27 18:36:24 crc kubenswrapper[4700]: I0227 18:36:24.737698 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_4891dc6b-c6d7-4468-ab14-e22ccc8462d2/probe/0.log" Feb 27 18:36:24 crc kubenswrapper[4700]: I0227 18:36:24.870432 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_4891dc6b-c6d7-4468-ab14-e22ccc8462d2/cinder-backup/0.log" Feb 27 18:36:25 crc kubenswrapper[4700]: I0227 18:36:25.021678 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_d9569ae4-5bf3-4720-bc48-0a1b735876ea/probe/0.log" Feb 27 18:36:25 crc kubenswrapper[4700]: I0227 18:36:25.129120 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-nfs-0_4c3e4a38-4e77-4f9d-aa8d-355b41080b84/probe/0.log" Feb 27 18:36:25 crc kubenswrapper[4700]: I0227 18:36:25.188472 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-nfs-0_4c3e4a38-4e77-4f9d-aa8d-355b41080b84/cinder-volume/0.log" Feb 27 18:36:25 crc kubenswrapper[4700]: I0227 18:36:25.369336 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-nfs-2-0_0e4cd883-f86f-43d5-b2f2-eb96bd9ed323/cinder-volume/0.log" Feb 27 18:36:25 crc kubenswrapper[4700]: I0227 18:36:25.428041 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-nfs-2-0_0e4cd883-f86f-43d5-b2f2-eb96bd9ed323/probe/0.log" Feb 27 18:36:25 crc kubenswrapper[4700]: I0227 18:36:25.527206 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-xndbg_493471ba-3317-48f3-accc-a2ddb143984c/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 18:36:25 crc kubenswrapper[4700]: I0227 18:36:25.657163 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-wd8dw_d3872381-519a-4292-bf34-fc008f8828c1/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 18:36:25 crc kubenswrapper[4700]: I0227 18:36:25.694892 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cfbbd847f-bd2kg_eaac16a0-f258-43bb-9c2f-ca23dec08397/init/0.log" Feb 27 18:36:25 crc kubenswrapper[4700]: I0227 18:36:25.941894 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-t8clg_e72836a3-fb67-48cf-971a-e90210a3d43d/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 18:36:25 crc kubenswrapper[4700]: I0227 18:36:25.963192 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cfbbd847f-bd2kg_eaac16a0-f258-43bb-9c2f-ca23dec08397/init/0.log" Feb 27 18:36:26 crc kubenswrapper[4700]: I0227 18:36:26.093892 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cfbbd847f-bd2kg_eaac16a0-f258-43bb-9c2f-ca23dec08397/dnsmasq-dns/0.log" Feb 27 18:36:26 crc kubenswrapper[4700]: I0227 18:36:26.165659 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_4bba45c6-8787-499a-a2a4-9f90e3ef1d7d/glance-log/0.log" Feb 27 18:36:26 crc kubenswrapper[4700]: I0227 18:36:26.185593 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_4bba45c6-8787-499a-a2a4-9f90e3ef1d7d/glance-httpd/0.log" Feb 27 18:36:26 crc kubenswrapper[4700]: I0227 18:36:26.264379 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_99391efa-6166-404b-9113-6e28f50d7136/glance-log/0.log" Feb 27 18:36:26 crc kubenswrapper[4700]: I0227 18:36:26.284500 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_99391efa-6166-404b-9113-6e28f50d7136/glance-httpd/0.log" Feb 27 18:36:26 crc kubenswrapper[4700]: I0227 18:36:26.433849 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6cc494ff8-ggq9w_ed7353a3-748a-40cb-8865-c4badeb8e402/horizon/0.log" Feb 27 18:36:26 crc kubenswrapper[4700]: I0227 18:36:26.532324 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-lpmfz_8a6cb27d-80dd-4ef8-b8a1-6c8a46329b45/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 18:36:26 crc kubenswrapper[4700]: I0227 18:36:26.692810 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-r7dmx_bfd90117-4b1d-45ed-9c0b-51a98fd914a4/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 18:36:27 crc kubenswrapper[4700]: I0227 18:36:27.107852 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6cc494ff8-ggq9w_ed7353a3-748a-40cb-8865-c4badeb8e402/horizon-log/0.log" Feb 27 18:36:27 crc kubenswrapper[4700]: I0227 18:36:27.123329 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29536921-kzv47_167e1521-363c-4a1a-9ff0-67981b7ebf37/keystone-cron/0.log" Feb 27 18:36:27 crc kubenswrapper[4700]: I0227 18:36:27.131966 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_f100325f-1efc-423e-9e23-e964c5547135/kube-state-metrics/0.log" Feb 27 18:36:27 crc kubenswrapper[4700]: I0227 18:36:27.349445 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-7dcdff5d6d-9f49t_47d9c3da-3a41-4ed3-93d9-1218f2d7a441/keystone-api/0.log" Feb 27 18:36:27 crc kubenswrapper[4700]: I0227 18:36:27.367085 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-p8j7p_abd22381-3367-4aef-a05d-dd895c6de545/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 18:36:27 crc kubenswrapper[4700]: I0227 18:36:27.876359 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-dc4cd4577-rvnzh_8d4d08db-27c4-4d2c-8902-968271280975/neutron-httpd/0.log" Feb 27 18:36:27 crc kubenswrapper[4700]: I0227 18:36:27.891642 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-dc4cd4577-rvnzh_8d4d08db-27c4-4d2c-8902-968271280975/neutron-api/0.log" Feb 27 18:36:27 crc kubenswrapper[4700]: E0227 18:36:27.983572 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:36:28 crc kubenswrapper[4700]: I0227 18:36:28.030144 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-ntcjg_a7c20871-f705-46a0-9856-e747aed1874e/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 18:36:28 crc kubenswrapper[4700]: I0227 18:36:28.162189 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_notifications-rabbitmq-server-0_fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da/setup-container/0.log" Feb 27 18:36:28 crc kubenswrapper[4700]: I0227 18:36:28.337269 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_notifications-rabbitmq-server-0_fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da/setup-container/0.log" Feb 27 18:36:28 crc kubenswrapper[4700]: I0227 18:36:28.404988 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_notifications-rabbitmq-server-0_fe7f30e4-63fa-4c8e-bb9a-d33c8c5264da/rabbitmq/0.log" Feb 27 18:36:28 crc kubenswrapper[4700]: I0227 18:36:28.959636 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_338e2de1-bc7a-4224-a7c1-e90c99eba72e/nova-cell0-conductor-conductor/0.log" Feb 27 18:36:29 crc kubenswrapper[4700]: I0227 18:36:29.279839 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_7602b1c6-8920-4dfd-a379-a008447edb85/nova-cell1-conductor-conductor/0.log" Feb 27 18:36:29 crc kubenswrapper[4700]: I0227 18:36:29.554846 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_e395b037-9e65-41ac-b615-bee1dbf5f0f9/nova-cell1-novncproxy-novncproxy/0.log" Feb 27 18:36:29 crc kubenswrapper[4700]: I0227 18:36:29.764212 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-7kqqq_fff27101-a84c-478b-9160-a7a5e6a189ea/nova-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 18:36:29 crc kubenswrapper[4700]: I0227 18:36:29.797699 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_87dac950-8564-49c9-a8bc-d51b0dd9773b/nova-api-log/0.log" Feb 27 18:36:30 crc kubenswrapper[4700]: I0227 18:36:30.033696 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_08f8b14c-db2a-4ac2-818a-4978ca9abf8b/nova-metadata-log/0.log" Feb 27 18:36:30 crc kubenswrapper[4700]: I0227 18:36:30.120321 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_87dac950-8564-49c9-a8bc-d51b0dd9773b/nova-api-api/0.log" Feb 27 18:36:30 crc kubenswrapper[4700]: I0227 18:36:30.388418 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223/mysql-bootstrap/0.log" Feb 27 18:36:30 crc kubenswrapper[4700]: I0227 18:36:30.513724 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_c0c63170-09b4-45f8-9b8b-d52ec9af2181/nova-scheduler-scheduler/0.log" Feb 27 18:36:30 crc kubenswrapper[4700]: I0227 18:36:30.536609 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223/mysql-bootstrap/0.log" Feb 27 18:36:30 crc kubenswrapper[4700]: I0227 18:36:30.677653 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_4a5c2d87-48b4-41da-ba7f-d9c3ddd7f223/galera/0.log" Feb 27 18:36:30 crc kubenswrapper[4700]: I0227 18:36:30.806874 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f2325e29-9f4c-438f-9400-e643e2494808/mysql-bootstrap/0.log" Feb 27 18:36:31 crc kubenswrapper[4700]: I0227 18:36:31.412722 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f2325e29-9f4c-438f-9400-e643e2494808/mysql-bootstrap/0.log" Feb 27 18:36:31 crc kubenswrapper[4700]: I0227 18:36:31.449490 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f2325e29-9f4c-438f-9400-e643e2494808/galera/0.log" Feb 27 18:36:31 crc kubenswrapper[4700]: I0227 18:36:31.593391 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_b2cd011d-8143-4800-ae5a-35c9d348892c/openstackclient/0.log" Feb 27 18:36:31 crc kubenswrapper[4700]: I0227 18:36:31.695619 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-7r49n_faaab80d-004f-48ab-b216-e687a0769a7b/openstack-network-exporter/0.log" Feb 27 18:36:31 crc kubenswrapper[4700]: I0227 18:36:31.891043 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-qqqm8_c7fc37b2-4775-468e-b118-e399646a95e9/ovsdb-server-init/0.log" Feb 27 18:36:32 crc kubenswrapper[4700]: I0227 18:36:32.341916 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-qqqm8_c7fc37b2-4775-468e-b118-e399646a95e9/ovsdb-server-init/0.log" Feb 27 18:36:32 crc kubenswrapper[4700]: I0227 18:36:32.342595 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_08f8b14c-db2a-4ac2-818a-4978ca9abf8b/nova-metadata-metadata/0.log" Feb 27 18:36:32 crc kubenswrapper[4700]: I0227 18:36:32.380997 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-qqqm8_c7fc37b2-4775-468e-b118-e399646a95e9/ovsdb-server/0.log" Feb 27 18:36:32 crc kubenswrapper[4700]: I0227 18:36:32.767632 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-qqqm8_c7fc37b2-4775-468e-b118-e399646a95e9/ovs-vswitchd/0.log" Feb 27 18:36:33 crc kubenswrapper[4700]: I0227 18:36:33.076887 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-sfldb_a3c2abcc-cc58-4b41-ae6a-c1f67a10c32a/ovn-controller/0.log" Feb 27 18:36:33 crc kubenswrapper[4700]: I0227 18:36:33.181152 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-kf8kd_e50903c5-4b11-4c01-b14f-7f0e51be5319/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 18:36:33 crc kubenswrapper[4700]: I0227 18:36:33.357074 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_616dbda2-825d-4a95-a3fc-96c44dbd49aa/openstack-network-exporter/0.log" Feb 27 18:36:33 crc kubenswrapper[4700]: I0227 18:36:33.422751 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_616dbda2-825d-4a95-a3fc-96c44dbd49aa/ovn-northd/0.log" Feb 27 18:36:33 crc kubenswrapper[4700]: I0227 18:36:33.531061 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a4b9e543-db0e-4b6a-bd32-9935f603556e/openstack-network-exporter/0.log" Feb 27 18:36:33 crc kubenswrapper[4700]: I0227 18:36:33.584314 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a4b9e543-db0e-4b6a-bd32-9935f603556e/ovsdbserver-nb/0.log" Feb 27 18:36:33 crc kubenswrapper[4700]: I0227 18:36:33.655992 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e/openstack-network-exporter/0.log" Feb 27 18:36:33 crc kubenswrapper[4700]: I0227 18:36:33.776290 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_f9e933fb-76ab-4cf0-8168-86f0e3f0ea7e/ovsdbserver-sb/0.log" Feb 27 18:36:34 crc kubenswrapper[4700]: I0227 18:36:34.034684 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-cc78b9998-2c2dk_779fb051-4fb5-4283-be4b-c46ff1602076/placement-api/0.log" Feb 27 18:36:34 crc kubenswrapper[4700]: I0227 18:36:34.144092 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_306434a9-5de0-42d0-8c10-184238aa5f97/init-config-reloader/0.log" Feb 27 18:36:34 crc kubenswrapper[4700]: I0227 18:36:34.166887 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-cc78b9998-2c2dk_779fb051-4fb5-4283-be4b-c46ff1602076/placement-log/0.log" Feb 27 18:36:34 crc kubenswrapper[4700]: I0227 18:36:34.345426 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_306434a9-5de0-42d0-8c10-184238aa5f97/init-config-reloader/0.log" Feb 27 18:36:34 crc kubenswrapper[4700]: I0227 18:36:34.398829 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_306434a9-5de0-42d0-8c10-184238aa5f97/prometheus/0.log" Feb 27 18:36:34 crc kubenswrapper[4700]: I0227 18:36:34.437662 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_306434a9-5de0-42d0-8c10-184238aa5f97/config-reloader/0.log" Feb 27 18:36:34 crc kubenswrapper[4700]: I0227 18:36:34.442634 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_306434a9-5de0-42d0-8c10-184238aa5f97/thanos-sidecar/0.log" Feb 27 18:36:34 crc kubenswrapper[4700]: I0227 18:36:34.585329 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fpxkz"] Feb 27 18:36:34 crc kubenswrapper[4700]: E0227 18:36:34.587633 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="074ce5d9-7bd8-4205-9c43-24c9cf22ea32" containerName="oc" Feb 27 18:36:34 crc kubenswrapper[4700]: I0227 18:36:34.587648 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="074ce5d9-7bd8-4205-9c43-24c9cf22ea32" containerName="oc" Feb 27 18:36:34 crc kubenswrapper[4700]: I0227 18:36:34.587851 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="074ce5d9-7bd8-4205-9c43-24c9cf22ea32" containerName="oc" Feb 27 18:36:34 crc kubenswrapper[4700]: I0227 18:36:34.595775 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fpxkz" Feb 27 18:36:34 crc kubenswrapper[4700]: I0227 18:36:34.603196 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fpxkz"] Feb 27 18:36:34 crc kubenswrapper[4700]: I0227 18:36:34.628143 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_43a2852f-4d8f-468c-988a-60dcac328039/setup-container/0.log" Feb 27 18:36:34 crc kubenswrapper[4700]: I0227 18:36:34.657882 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed2293ce-4fd2-4674-81e9-0e63c000ab21-utilities\") pod \"redhat-marketplace-fpxkz\" (UID: \"ed2293ce-4fd2-4674-81e9-0e63c000ab21\") " pod="openshift-marketplace/redhat-marketplace-fpxkz" Feb 27 18:36:34 crc kubenswrapper[4700]: I0227 18:36:34.658214 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed2293ce-4fd2-4674-81e9-0e63c000ab21-catalog-content\") pod \"redhat-marketplace-fpxkz\" (UID: \"ed2293ce-4fd2-4674-81e9-0e63c000ab21\") " pod="openshift-marketplace/redhat-marketplace-fpxkz" Feb 27 18:36:34 crc kubenswrapper[4700]: I0227 18:36:34.658341 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bqn7\" (UniqueName: \"kubernetes.io/projected/ed2293ce-4fd2-4674-81e9-0e63c000ab21-kube-api-access-9bqn7\") pod \"redhat-marketplace-fpxkz\" (UID: \"ed2293ce-4fd2-4674-81e9-0e63c000ab21\") " pod="openshift-marketplace/redhat-marketplace-fpxkz" Feb 27 18:36:34 crc kubenswrapper[4700]: I0227 18:36:34.759959 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed2293ce-4fd2-4674-81e9-0e63c000ab21-utilities\") pod \"redhat-marketplace-fpxkz\" (UID: \"ed2293ce-4fd2-4674-81e9-0e63c000ab21\") " pod="openshift-marketplace/redhat-marketplace-fpxkz" Feb 27 18:36:34 crc kubenswrapper[4700]: I0227 18:36:34.760276 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed2293ce-4fd2-4674-81e9-0e63c000ab21-catalog-content\") pod \"redhat-marketplace-fpxkz\" (UID: \"ed2293ce-4fd2-4674-81e9-0e63c000ab21\") " pod="openshift-marketplace/redhat-marketplace-fpxkz" Feb 27 18:36:34 crc kubenswrapper[4700]: I0227 18:36:34.760451 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bqn7\" (UniqueName: \"kubernetes.io/projected/ed2293ce-4fd2-4674-81e9-0e63c000ab21-kube-api-access-9bqn7\") pod \"redhat-marketplace-fpxkz\" (UID: \"ed2293ce-4fd2-4674-81e9-0e63c000ab21\") " pod="openshift-marketplace/redhat-marketplace-fpxkz" Feb 27 18:36:34 crc kubenswrapper[4700]: I0227 18:36:34.760755 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed2293ce-4fd2-4674-81e9-0e63c000ab21-utilities\") pod \"redhat-marketplace-fpxkz\" (UID: \"ed2293ce-4fd2-4674-81e9-0e63c000ab21\") " pod="openshift-marketplace/redhat-marketplace-fpxkz" Feb 27 18:36:34 crc kubenswrapper[4700]: I0227 18:36:34.760827 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed2293ce-4fd2-4674-81e9-0e63c000ab21-catalog-content\") pod \"redhat-marketplace-fpxkz\" (UID: \"ed2293ce-4fd2-4674-81e9-0e63c000ab21\") " pod="openshift-marketplace/redhat-marketplace-fpxkz" Feb 27 18:36:34 crc kubenswrapper[4700]: I0227 18:36:34.782552 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bqn7\" (UniqueName: \"kubernetes.io/projected/ed2293ce-4fd2-4674-81e9-0e63c000ab21-kube-api-access-9bqn7\") pod \"redhat-marketplace-fpxkz\" (UID: \"ed2293ce-4fd2-4674-81e9-0e63c000ab21\") " pod="openshift-marketplace/redhat-marketplace-fpxkz" Feb 27 18:36:34 crc kubenswrapper[4700]: I0227 18:36:34.864649 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_43a2852f-4d8f-468c-988a-60dcac328039/setup-container/0.log" Feb 27 18:36:34 crc kubenswrapper[4700]: I0227 18:36:34.930664 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fpxkz" Feb 27 18:36:34 crc kubenswrapper[4700]: I0227 18:36:34.941853 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_15c786bb-7353-4d06-b410-a44458a5f954/setup-container/0.log" Feb 27 18:36:35 crc kubenswrapper[4700]: I0227 18:36:35.085990 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_43a2852f-4d8f-468c-988a-60dcac328039/rabbitmq/0.log" Feb 27 18:36:35 crc kubenswrapper[4700]: I0227 18:36:35.412415 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_15c786bb-7353-4d06-b410-a44458a5f954/setup-container/0.log" Feb 27 18:36:35 crc kubenswrapper[4700]: I0227 18:36:35.460981 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fpxkz"] Feb 27 18:36:35 crc kubenswrapper[4700]: I0227 18:36:35.463766 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_15c786bb-7353-4d06-b410-a44458a5f954/rabbitmq/0.log" Feb 27 18:36:35 crc kubenswrapper[4700]: I0227 18:36:35.506841 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-274bw_f3d2123c-c2d1-44d0-90c2-2fa65b9d062c/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 18:36:35 crc kubenswrapper[4700]: I0227 18:36:35.741590 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-rqjvd_85c1092c-f937-492c-9791-9df6150f439d/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 18:36:35 crc kubenswrapper[4700]: I0227 18:36:35.753591 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-8b64m_ee61dca5-6400-424d-b76c-53ef8e3b0434/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 18:36:35 crc kubenswrapper[4700]: I0227 18:36:35.985516 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-dmfwb_7ee8fa48-c17e-43a3-aef8-d5ac737e0b36/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 18:36:36 crc kubenswrapper[4700]: I0227 18:36:36.097295 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-sznqn_0115f2bd-d709-44d2-8a30-0c8debd51afa/ssh-known-hosts-edpm-deployment/0.log" Feb 27 18:36:36 crc kubenswrapper[4700]: I0227 18:36:36.170897 4700 generic.go:334] "Generic (PLEG): container finished" podID="ed2293ce-4fd2-4674-81e9-0e63c000ab21" containerID="3d443f6731f8e8daa0b9140d077b809faca81293f18e6dadac10b6f42bdf12a2" exitCode=0 Feb 27 18:36:36 crc kubenswrapper[4700]: I0227 18:36:36.170946 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fpxkz" event={"ID":"ed2293ce-4fd2-4674-81e9-0e63c000ab21","Type":"ContainerDied","Data":"3d443f6731f8e8daa0b9140d077b809faca81293f18e6dadac10b6f42bdf12a2"} Feb 27 18:36:36 crc kubenswrapper[4700]: I0227 18:36:36.170976 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fpxkz" event={"ID":"ed2293ce-4fd2-4674-81e9-0e63c000ab21","Type":"ContainerStarted","Data":"19db586553ec6e133c709eedbd0e15493c5fcc5700bc4c81bcc2798fcefc1fc1"} Feb 27 18:36:36 crc kubenswrapper[4700]: I0227 18:36:36.329543 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-65457d8799-dclxm_e43c5d5e-63ef-45b9-af4e-627a035b376e/proxy-server/0.log" Feb 27 18:36:36 crc kubenswrapper[4700]: I0227 18:36:36.409912 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-65457d8799-dclxm_e43c5d5e-63ef-45b9-af4e-627a035b376e/proxy-httpd/0.log" Feb 27 18:36:36 crc kubenswrapper[4700]: I0227 18:36:36.410016 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 18:36:36 crc kubenswrapper[4700]: I0227 18:36:36.410074 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 18:36:36 crc kubenswrapper[4700]: I0227 18:36:36.750705 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_9cf3890b-d331-4312-a6ce-a49945e61147/memcached/0.log" Feb 27 18:36:36 crc kubenswrapper[4700]: I0227 18:36:36.772090 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-m7qsc_89befb8e-a997-46a9-9db2-24f0dc5bf10d/swift-ring-rebalance/0.log" Feb 27 18:36:36 crc kubenswrapper[4700]: I0227 18:36:36.865526 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9df57c73-8d37-4f7d-b62f-209210dbba42/account-reaper/0.log" Feb 27 18:36:36 crc kubenswrapper[4700]: I0227 18:36:36.906107 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9df57c73-8d37-4f7d-b62f-209210dbba42/account-auditor/0.log" Feb 27 18:36:37 crc kubenswrapper[4700]: I0227 18:36:37.005279 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9df57c73-8d37-4f7d-b62f-209210dbba42/account-replicator/0.log" Feb 27 18:36:37 crc kubenswrapper[4700]: I0227 18:36:37.030116 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9df57c73-8d37-4f7d-b62f-209210dbba42/account-server/0.log" Feb 27 18:36:37 crc kubenswrapper[4700]: I0227 18:36:37.128665 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9df57c73-8d37-4f7d-b62f-209210dbba42/container-server/0.log" Feb 27 18:36:37 crc kubenswrapper[4700]: I0227 18:36:37.142659 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9df57c73-8d37-4f7d-b62f-209210dbba42/container-auditor/0.log" Feb 27 18:36:37 crc kubenswrapper[4700]: I0227 18:36:37.143407 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9df57c73-8d37-4f7d-b62f-209210dbba42/container-replicator/0.log" Feb 27 18:36:37 crc kubenswrapper[4700]: I0227 18:36:37.188432 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fpxkz" event={"ID":"ed2293ce-4fd2-4674-81e9-0e63c000ab21","Type":"ContainerStarted","Data":"37553ebda967c5e62330986828108bd28672efba3c070ede576339fd10dc2961"} Feb 27 18:36:37 crc kubenswrapper[4700]: I0227 18:36:37.202515 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9df57c73-8d37-4f7d-b62f-209210dbba42/container-updater/0.log" Feb 27 18:36:37 crc kubenswrapper[4700]: I0227 18:36:37.300495 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9df57c73-8d37-4f7d-b62f-209210dbba42/object-auditor/0.log" Feb 27 18:36:37 crc kubenswrapper[4700]: I0227 18:36:37.328903 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9df57c73-8d37-4f7d-b62f-209210dbba42/object-expirer/0.log" Feb 27 18:36:37 crc kubenswrapper[4700]: I0227 18:36:37.360506 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9df57c73-8d37-4f7d-b62f-209210dbba42/object-server/0.log" Feb 27 18:36:37 crc kubenswrapper[4700]: I0227 18:36:37.402146 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9df57c73-8d37-4f7d-b62f-209210dbba42/object-replicator/0.log" Feb 27 18:36:37 crc kubenswrapper[4700]: I0227 18:36:37.412266 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9df57c73-8d37-4f7d-b62f-209210dbba42/object-updater/0.log" Feb 27 18:36:37 crc kubenswrapper[4700]: I0227 18:36:37.519545 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9df57c73-8d37-4f7d-b62f-209210dbba42/rsync/0.log" Feb 27 18:36:37 crc kubenswrapper[4700]: I0227 18:36:37.580543 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9df57c73-8d37-4f7d-b62f-209210dbba42/swift-recon-cron/0.log" Feb 27 18:36:37 crc kubenswrapper[4700]: I0227 18:36:37.630893 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-z5pgn_d9b66d11-37c1-4826-96c3-967dd98f053c/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 18:36:37 crc kubenswrapper[4700]: I0227 18:36:37.781081 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_805fcd3a-360c-445d-85d8-889bf15d310b/tempest-tests-tempest-tests-runner/0.log" Feb 27 18:36:37 crc kubenswrapper[4700]: I0227 18:36:37.845069 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_82d7f7cf-9ad7-4f87-b31c-a3c31b454289/test-operator-logs-container/0.log" Feb 27 18:36:37 crc kubenswrapper[4700]: I0227 18:36:37.915928 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-dgj8b_2855040c-fd3a-4dae-b6c0-626cc01ee321/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 18:36:38 crc kubenswrapper[4700]: I0227 18:36:38.199247 4700 generic.go:334] "Generic (PLEG): container finished" podID="ed2293ce-4fd2-4674-81e9-0e63c000ab21" containerID="37553ebda967c5e62330986828108bd28672efba3c070ede576339fd10dc2961" exitCode=0 Feb 27 18:36:38 crc kubenswrapper[4700]: I0227 18:36:38.199320 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fpxkz" event={"ID":"ed2293ce-4fd2-4674-81e9-0e63c000ab21","Type":"ContainerDied","Data":"37553ebda967c5e62330986828108bd28672efba3c070ede576339fd10dc2961"} Feb 27 18:36:38 crc kubenswrapper[4700]: I0227 18:36:38.633248 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-applier-0_e3bb7b7a-9fbd-47ed-99f5-ac0dd93ecceb/watcher-applier/0.log" Feb 27 18:36:39 crc kubenswrapper[4700]: I0227 18:36:39.216702 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fpxkz" event={"ID":"ed2293ce-4fd2-4674-81e9-0e63c000ab21","Type":"ContainerStarted","Data":"75aa121919fa0eff302676ebb165728a7548583eaf75209401717b1c6cea8d5d"} Feb 27 18:36:39 crc kubenswrapper[4700]: I0227 18:36:39.239191 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fpxkz" podStartSLOduration=2.756688992 podStartE2EDuration="5.23917124s" podCreationTimestamp="2026-02-27 18:36:34 +0000 UTC" firstStartedPulling="2026-02-27 18:36:36.17313989 +0000 UTC m=+5756.158452637" lastFinishedPulling="2026-02-27 18:36:38.655622138 +0000 UTC m=+5758.640934885" observedRunningTime="2026-02-27 18:36:39.23109304 +0000 UTC m=+5759.216405807" watchObservedRunningTime="2026-02-27 18:36:39.23917124 +0000 UTC m=+5759.224483987" Feb 27 18:36:39 crc kubenswrapper[4700]: I0227 18:36:39.309221 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-api-0_36944f75-7dc8-4a1a-b20f-ab423b8282fa/watcher-api-log/0.log" Feb 27 18:36:40 crc kubenswrapper[4700]: E0227 18:36:40.993370 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:36:41 crc kubenswrapper[4700]: I0227 18:36:41.242874 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-decision-engine-0_e7f47383-ff4b-4d9b-912a-9fe2f5ec26da/watcher-decision-engine/0.log" Feb 27 18:36:41 crc kubenswrapper[4700]: I0227 18:36:41.875627 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-api-0_36944f75-7dc8-4a1a-b20f-ab423b8282fa/watcher-api/0.log" Feb 27 18:36:44 crc kubenswrapper[4700]: I0227 18:36:44.931100 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fpxkz" Feb 27 18:36:44 crc kubenswrapper[4700]: I0227 18:36:44.931514 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fpxkz" Feb 27 18:36:44 crc kubenswrapper[4700]: I0227 18:36:44.991828 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fpxkz" Feb 27 18:36:45 crc kubenswrapper[4700]: I0227 18:36:45.581295 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fpxkz" Feb 27 18:36:45 crc kubenswrapper[4700]: I0227 18:36:45.630795 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fpxkz"] Feb 27 18:36:47 crc kubenswrapper[4700]: I0227 18:36:47.506139 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fpxkz" podUID="ed2293ce-4fd2-4674-81e9-0e63c000ab21" containerName="registry-server" containerID="cri-o://75aa121919fa0eff302676ebb165728a7548583eaf75209401717b1c6cea8d5d" gracePeriod=2 Feb 27 18:36:47 crc kubenswrapper[4700]: I0227 18:36:47.969945 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fpxkz" Feb 27 18:36:48 crc kubenswrapper[4700]: I0227 18:36:48.148183 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed2293ce-4fd2-4674-81e9-0e63c000ab21-utilities\") pod \"ed2293ce-4fd2-4674-81e9-0e63c000ab21\" (UID: \"ed2293ce-4fd2-4674-81e9-0e63c000ab21\") " Feb 27 18:36:48 crc kubenswrapper[4700]: I0227 18:36:48.148467 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9bqn7\" (UniqueName: \"kubernetes.io/projected/ed2293ce-4fd2-4674-81e9-0e63c000ab21-kube-api-access-9bqn7\") pod \"ed2293ce-4fd2-4674-81e9-0e63c000ab21\" (UID: \"ed2293ce-4fd2-4674-81e9-0e63c000ab21\") " Feb 27 18:36:48 crc kubenswrapper[4700]: I0227 18:36:48.148508 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed2293ce-4fd2-4674-81e9-0e63c000ab21-catalog-content\") pod \"ed2293ce-4fd2-4674-81e9-0e63c000ab21\" (UID: \"ed2293ce-4fd2-4674-81e9-0e63c000ab21\") " Feb 27 18:36:48 crc kubenswrapper[4700]: I0227 18:36:48.149236 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed2293ce-4fd2-4674-81e9-0e63c000ab21-utilities" (OuterVolumeSpecName: "utilities") pod "ed2293ce-4fd2-4674-81e9-0e63c000ab21" (UID: "ed2293ce-4fd2-4674-81e9-0e63c000ab21"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 18:36:48 crc kubenswrapper[4700]: I0227 18:36:48.155477 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed2293ce-4fd2-4674-81e9-0e63c000ab21-kube-api-access-9bqn7" (OuterVolumeSpecName: "kube-api-access-9bqn7") pod "ed2293ce-4fd2-4674-81e9-0e63c000ab21" (UID: "ed2293ce-4fd2-4674-81e9-0e63c000ab21"). InnerVolumeSpecName "kube-api-access-9bqn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:36:48 crc kubenswrapper[4700]: I0227 18:36:48.183996 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed2293ce-4fd2-4674-81e9-0e63c000ab21-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ed2293ce-4fd2-4674-81e9-0e63c000ab21" (UID: "ed2293ce-4fd2-4674-81e9-0e63c000ab21"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 18:36:48 crc kubenswrapper[4700]: I0227 18:36:48.250181 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9bqn7\" (UniqueName: \"kubernetes.io/projected/ed2293ce-4fd2-4674-81e9-0e63c000ab21-kube-api-access-9bqn7\") on node \"crc\" DevicePath \"\"" Feb 27 18:36:48 crc kubenswrapper[4700]: I0227 18:36:48.250222 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed2293ce-4fd2-4674-81e9-0e63c000ab21-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 18:36:48 crc kubenswrapper[4700]: I0227 18:36:48.250233 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed2293ce-4fd2-4674-81e9-0e63c000ab21-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 18:36:48 crc kubenswrapper[4700]: I0227 18:36:48.515965 4700 generic.go:334] "Generic (PLEG): container finished" podID="ed2293ce-4fd2-4674-81e9-0e63c000ab21" containerID="75aa121919fa0eff302676ebb165728a7548583eaf75209401717b1c6cea8d5d" exitCode=0 Feb 27 18:36:48 crc kubenswrapper[4700]: I0227 18:36:48.516008 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fpxkz" event={"ID":"ed2293ce-4fd2-4674-81e9-0e63c000ab21","Type":"ContainerDied","Data":"75aa121919fa0eff302676ebb165728a7548583eaf75209401717b1c6cea8d5d"} Feb 27 18:36:48 crc kubenswrapper[4700]: I0227 18:36:48.516041 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fpxkz" event={"ID":"ed2293ce-4fd2-4674-81e9-0e63c000ab21","Type":"ContainerDied","Data":"19db586553ec6e133c709eedbd0e15493c5fcc5700bc4c81bcc2798fcefc1fc1"} Feb 27 18:36:48 crc kubenswrapper[4700]: I0227 18:36:48.516046 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fpxkz" Feb 27 18:36:48 crc kubenswrapper[4700]: I0227 18:36:48.516059 4700 scope.go:117] "RemoveContainer" containerID="75aa121919fa0eff302676ebb165728a7548583eaf75209401717b1c6cea8d5d" Feb 27 18:36:48 crc kubenswrapper[4700]: I0227 18:36:48.546243 4700 scope.go:117] "RemoveContainer" containerID="37553ebda967c5e62330986828108bd28672efba3c070ede576339fd10dc2961" Feb 27 18:36:48 crc kubenswrapper[4700]: I0227 18:36:48.566942 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fpxkz"] Feb 27 18:36:48 crc kubenswrapper[4700]: I0227 18:36:48.583505 4700 scope.go:117] "RemoveContainer" containerID="3d443f6731f8e8daa0b9140d077b809faca81293f18e6dadac10b6f42bdf12a2" Feb 27 18:36:48 crc kubenswrapper[4700]: I0227 18:36:48.587895 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fpxkz"] Feb 27 18:36:48 crc kubenswrapper[4700]: I0227 18:36:48.618982 4700 scope.go:117] "RemoveContainer" containerID="75aa121919fa0eff302676ebb165728a7548583eaf75209401717b1c6cea8d5d" Feb 27 18:36:48 crc kubenswrapper[4700]: E0227 18:36:48.619714 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75aa121919fa0eff302676ebb165728a7548583eaf75209401717b1c6cea8d5d\": container with ID starting with 75aa121919fa0eff302676ebb165728a7548583eaf75209401717b1c6cea8d5d not found: ID does not exist" containerID="75aa121919fa0eff302676ebb165728a7548583eaf75209401717b1c6cea8d5d" Feb 27 18:36:48 crc kubenswrapper[4700]: I0227 18:36:48.619759 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75aa121919fa0eff302676ebb165728a7548583eaf75209401717b1c6cea8d5d"} err="failed to get container status \"75aa121919fa0eff302676ebb165728a7548583eaf75209401717b1c6cea8d5d\": rpc error: code = NotFound desc = could not find container \"75aa121919fa0eff302676ebb165728a7548583eaf75209401717b1c6cea8d5d\": container with ID starting with 75aa121919fa0eff302676ebb165728a7548583eaf75209401717b1c6cea8d5d not found: ID does not exist" Feb 27 18:36:48 crc kubenswrapper[4700]: I0227 18:36:48.619786 4700 scope.go:117] "RemoveContainer" containerID="37553ebda967c5e62330986828108bd28672efba3c070ede576339fd10dc2961" Feb 27 18:36:48 crc kubenswrapper[4700]: E0227 18:36:48.619998 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37553ebda967c5e62330986828108bd28672efba3c070ede576339fd10dc2961\": container with ID starting with 37553ebda967c5e62330986828108bd28672efba3c070ede576339fd10dc2961 not found: ID does not exist" containerID="37553ebda967c5e62330986828108bd28672efba3c070ede576339fd10dc2961" Feb 27 18:36:48 crc kubenswrapper[4700]: I0227 18:36:48.620017 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37553ebda967c5e62330986828108bd28672efba3c070ede576339fd10dc2961"} err="failed to get container status \"37553ebda967c5e62330986828108bd28672efba3c070ede576339fd10dc2961\": rpc error: code = NotFound desc = could not find container \"37553ebda967c5e62330986828108bd28672efba3c070ede576339fd10dc2961\": container with ID starting with 37553ebda967c5e62330986828108bd28672efba3c070ede576339fd10dc2961 not found: ID does not exist" Feb 27 18:36:48 crc kubenswrapper[4700]: I0227 18:36:48.620029 4700 scope.go:117] "RemoveContainer" containerID="3d443f6731f8e8daa0b9140d077b809faca81293f18e6dadac10b6f42bdf12a2" Feb 27 18:36:48 crc kubenswrapper[4700]: E0227 18:36:48.620178 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d443f6731f8e8daa0b9140d077b809faca81293f18e6dadac10b6f42bdf12a2\": container with ID starting with 3d443f6731f8e8daa0b9140d077b809faca81293f18e6dadac10b6f42bdf12a2 not found: ID does not exist" containerID="3d443f6731f8e8daa0b9140d077b809faca81293f18e6dadac10b6f42bdf12a2" Feb 27 18:36:48 crc kubenswrapper[4700]: I0227 18:36:48.620198 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d443f6731f8e8daa0b9140d077b809faca81293f18e6dadac10b6f42bdf12a2"} err="failed to get container status \"3d443f6731f8e8daa0b9140d077b809faca81293f18e6dadac10b6f42bdf12a2\": rpc error: code = NotFound desc = could not find container \"3d443f6731f8e8daa0b9140d077b809faca81293f18e6dadac10b6f42bdf12a2\": container with ID starting with 3d443f6731f8e8daa0b9140d077b809faca81293f18e6dadac10b6f42bdf12a2 not found: ID does not exist" Feb 27 18:36:48 crc kubenswrapper[4700]: I0227 18:36:48.991955 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed2293ce-4fd2-4674-81e9-0e63c000ab21" path="/var/lib/kubelet/pods/ed2293ce-4fd2-4674-81e9-0e63c000ab21/volumes" Feb 27 18:36:55 crc kubenswrapper[4700]: E0227 18:36:55.983543 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:37:06 crc kubenswrapper[4700]: I0227 18:37:06.410647 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 18:37:06 crc kubenswrapper[4700]: I0227 18:37:06.411781 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 18:37:10 crc kubenswrapper[4700]: E0227 18:37:10.990172 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:37:12 crc kubenswrapper[4700]: I0227 18:37:12.509515 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a70a7d17df4824bd764910ad293b109e72da56f2f4bf07c2a6401f9135j2sd8_670f4e8a-0015-44af-8397-87e1ef84c3e0/util/0.log" Feb 27 18:37:13 crc kubenswrapper[4700]: I0227 18:37:13.585246 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a70a7d17df4824bd764910ad293b109e72da56f2f4bf07c2a6401f9135j2sd8_670f4e8a-0015-44af-8397-87e1ef84c3e0/util/0.log" Feb 27 18:37:13 crc kubenswrapper[4700]: I0227 18:37:13.660807 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a70a7d17df4824bd764910ad293b109e72da56f2f4bf07c2a6401f9135j2sd8_670f4e8a-0015-44af-8397-87e1ef84c3e0/pull/0.log" Feb 27 18:37:13 crc kubenswrapper[4700]: I0227 18:37:13.737207 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a70a7d17df4824bd764910ad293b109e72da56f2f4bf07c2a6401f9135j2sd8_670f4e8a-0015-44af-8397-87e1ef84c3e0/pull/0.log" Feb 27 18:37:13 crc kubenswrapper[4700]: I0227 18:37:13.911842 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a70a7d17df4824bd764910ad293b109e72da56f2f4bf07c2a6401f9135j2sd8_670f4e8a-0015-44af-8397-87e1ef84c3e0/util/0.log" Feb 27 18:37:13 crc kubenswrapper[4700]: I0227 18:37:13.919789 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a70a7d17df4824bd764910ad293b109e72da56f2f4bf07c2a6401f9135j2sd8_670f4e8a-0015-44af-8397-87e1ef84c3e0/pull/0.log" Feb 27 18:37:13 crc kubenswrapper[4700]: I0227 18:37:13.922185 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a70a7d17df4824bd764910ad293b109e72da56f2f4bf07c2a6401f9135j2sd8_670f4e8a-0015-44af-8397-87e1ef84c3e0/extract/0.log" Feb 27 18:37:14 crc kubenswrapper[4700]: I0227 18:37:14.353642 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-5d87c9d997-9mxq4_9b575398-8936-48f1-a412-ec26c21303cf/manager/0.log" Feb 27 18:37:14 crc kubenswrapper[4700]: I0227 18:37:14.355052 4700 scope.go:117] "RemoveContainer" containerID="4a54bd2797b5714d8a57bcd263accf252512d3e44013a6579b8785a2cc66e4d6" Feb 27 18:37:14 crc kubenswrapper[4700]: I0227 18:37:14.714957 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-64db6967f8-l2542_181557a9-fcb7-4554-82ae-d3446b460867/manager/0.log" Feb 27 18:37:15 crc kubenswrapper[4700]: I0227 18:37:15.002181 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-cf99c678f-n55xj_ab00dea0-59fa-49ea-92f9-a230da6ea536/manager/0.log" Feb 27 18:37:15 crc kubenswrapper[4700]: I0227 18:37:15.538816 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-78bc7f9bd9-dm8wj_c3ac9c3b-2142-409d-8c49-1bf2837252d7/manager/0.log" Feb 27 18:37:15 crc kubenswrapper[4700]: I0227 18:37:15.800320 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-545456dc4-stc8v_9c1db815-93de-4bcc-9698-22bf6588ac88/manager/0.log" Feb 27 18:37:16 crc kubenswrapper[4700]: I0227 18:37:16.160422 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-55ffd4876b-q66w8_7cd71afd-89a7-42b9-9913-6827ea7a22d3/manager/0.log" Feb 27 18:37:16 crc kubenswrapper[4700]: I0227 18:37:16.329497 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-67d996989d-f625b_e6d0216e-25c4-4957-8e6b-473b49dd7543/manager/0.log" Feb 27 18:37:16 crc kubenswrapper[4700]: I0227 18:37:16.556155 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-f7fcc58b9-84zw4_56bc1032-efb2-4480-9091-11e529096a3c/manager/0.log" Feb 27 18:37:16 crc kubenswrapper[4700]: I0227 18:37:16.631724 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-556b8b874-4bmbt_c1fdff54-9c43-4fe4-925b-f87590c75e72/manager/0.log" Feb 27 18:37:16 crc kubenswrapper[4700]: I0227 18:37:16.763072 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-55d77d7b5c-dc72n_dcdee90e-b946-4cf5-807c-6babcd83071f/manager/0.log" Feb 27 18:37:16 crc kubenswrapper[4700]: I0227 18:37:16.805743 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-54688575f-8vhl9_e00942b5-6f82-48d4-8413-2f568daf5d7f/manager/0.log" Feb 27 18:37:16 crc kubenswrapper[4700]: I0227 18:37:16.924531 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-74b6b5dc96-72l74_5bbcc1bb-6070-4196-8212-cf8e04eaa923/manager/0.log" Feb 27 18:37:17 crc kubenswrapper[4700]: I0227 18:37:17.008702 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-5d86c7ddb7-vt248_bd70bcca-beb1-464b-a1a3-b154ff45c1f5/manager/0.log" Feb 27 18:37:17 crc kubenswrapper[4700]: I0227 18:37:17.160935 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg_f6a3d37c-2b18-40bc-a990-56dc67093abf/manager/0.log" Feb 27 18:37:17 crc kubenswrapper[4700]: I0227 18:37:17.412953 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-64d4474cb4-mb4nb_d65f20c2-a9a2-43a1-9adf-9925736c38c5/operator/0.log" Feb 27 18:37:17 crc kubenswrapper[4700]: I0227 18:37:17.590264 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-tdw6d_16e196e0-ad2c-4e3e-8056-d15d71661a96/registry-server/0.log" Feb 27 18:37:17 crc kubenswrapper[4700]: I0227 18:37:17.836014 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-75684d597f-6cfj6_00a21bc7-24a6-4a6d-ba48-18db1d9233c6/manager/0.log" Feb 27 18:37:17 crc kubenswrapper[4700]: I0227 18:37:17.914015 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-648564c9fc-dmd7w_5d94c9a7-4f4e-4370-a3da-9d80309f327a/manager/0.log" Feb 27 18:37:18 crc kubenswrapper[4700]: I0227 18:37:18.148039 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-qf2g9_1010adc0-3454-486d-a676-80d329745d4c/operator/0.log" Feb 27 18:37:18 crc kubenswrapper[4700]: I0227 18:37:18.366371 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9b9ff9f4d-cb7w5_15764e3b-3a40-4742-91a4-3d339cb86a0b/manager/0.log" Feb 27 18:37:18 crc kubenswrapper[4700]: I0227 18:37:18.662681 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5fdb694969-44djz_629a3ab5-c8ce-49cb-bd57-355d643253c5/manager/0.log" Feb 27 18:37:18 crc kubenswrapper[4700]: I0227 18:37:18.669018 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-55b5ff4dbb-xgf2x_269a9c0b-7242-44dc-b860-386757217052/manager/0.log" Feb 27 18:37:19 crc kubenswrapper[4700]: I0227 18:37:19.076018 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-8f7484475-8pdn6_33fd8eaf-c78f-4949-be53-86d8981ca296/manager/0.log" Feb 27 18:37:19 crc kubenswrapper[4700]: I0227 18:37:19.433218 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7ff59cf98b-bntz4_fe5a4feb-297b-4c17-bc6c-cab1bf1dee41/manager/0.log" Feb 27 18:37:23 crc kubenswrapper[4700]: E0227 18:37:23.984821 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:37:25 crc kubenswrapper[4700]: I0227 18:37:25.899883 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6db6876945-nmv94_37c36ebd-e968-4024-a8c5-77c6507b95cf/manager/0.log" Feb 27 18:37:36 crc kubenswrapper[4700]: I0227 18:37:36.410439 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 18:37:36 crc kubenswrapper[4700]: I0227 18:37:36.411136 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 18:37:36 crc kubenswrapper[4700]: I0227 18:37:36.411202 4700 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" Feb 27 18:37:36 crc kubenswrapper[4700]: I0227 18:37:36.412045 4700 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c7ad608c50ddee2aac4992df7a6ccaaeda7c737e2a0d2fad5282b4482eac9e50"} pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 18:37:36 crc kubenswrapper[4700]: I0227 18:37:36.412128 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" containerID="cri-o://c7ad608c50ddee2aac4992df7a6ccaaeda7c737e2a0d2fad5282b4482eac9e50" gracePeriod=600 Feb 27 18:37:36 crc kubenswrapper[4700]: E0227 18:37:36.550129 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:37:37 crc kubenswrapper[4700]: I0227 18:37:37.010451 4700 generic.go:334] "Generic (PLEG): container finished" podID="84b45511-b94c-479f-98db-fd2c4eceec46" containerID="c7ad608c50ddee2aac4992df7a6ccaaeda7c737e2a0d2fad5282b4482eac9e50" exitCode=0 Feb 27 18:37:37 crc kubenswrapper[4700]: I0227 18:37:37.010521 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerDied","Data":"c7ad608c50ddee2aac4992df7a6ccaaeda7c737e2a0d2fad5282b4482eac9e50"} Feb 27 18:37:37 crc kubenswrapper[4700]: I0227 18:37:37.010946 4700 scope.go:117] "RemoveContainer" containerID="fc0ec093303e51bd2ed80860af9a81018e732a7449b44faf6a9d28fb0570f2b0" Feb 27 18:37:37 crc kubenswrapper[4700]: I0227 18:37:37.012196 4700 scope.go:117] "RemoveContainer" containerID="c7ad608c50ddee2aac4992df7a6ccaaeda7c737e2a0d2fad5282b4482eac9e50" Feb 27 18:37:37 crc kubenswrapper[4700]: E0227 18:37:37.012772 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:37:38 crc kubenswrapper[4700]: E0227 18:37:38.983603 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:37:44 crc kubenswrapper[4700]: I0227 18:37:44.427603 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-sbqxk_f3526e22-b6e6-4a3a-95ec-67a831e5678a/control-plane-machine-set-operator/0.log" Feb 27 18:37:44 crc kubenswrapper[4700]: I0227 18:37:44.599924 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-b9772_4ea98230-93e8-4ef6-86d6-e25b8448c632/kube-rbac-proxy/0.log" Feb 27 18:37:44 crc kubenswrapper[4700]: I0227 18:37:44.608257 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-b9772_4ea98230-93e8-4ef6-86d6-e25b8448c632/machine-api-operator/0.log" Feb 27 18:37:49 crc kubenswrapper[4700]: I0227 18:37:49.983093 4700 scope.go:117] "RemoveContainer" containerID="c7ad608c50ddee2aac4992df7a6ccaaeda7c737e2a0d2fad5282b4482eac9e50" Feb 27 18:37:49 crc kubenswrapper[4700]: E0227 18:37:49.984931 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:37:51 crc kubenswrapper[4700]: E0227 18:37:51.983002 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:37:59 crc kubenswrapper[4700]: I0227 18:37:59.749238 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="f2325e29-9f4c-438f-9400-e643e2494808" containerName="galera" probeResult="failure" output="command timed out" Feb 27 18:38:00 crc kubenswrapper[4700]: I0227 18:38:00.144595 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536958-tdbth"] Feb 27 18:38:00 crc kubenswrapper[4700]: E0227 18:38:00.144990 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed2293ce-4fd2-4674-81e9-0e63c000ab21" containerName="extract-utilities" Feb 27 18:38:00 crc kubenswrapper[4700]: I0227 18:38:00.145007 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed2293ce-4fd2-4674-81e9-0e63c000ab21" containerName="extract-utilities" Feb 27 18:38:00 crc kubenswrapper[4700]: E0227 18:38:00.145019 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed2293ce-4fd2-4674-81e9-0e63c000ab21" containerName="extract-content" Feb 27 18:38:00 crc kubenswrapper[4700]: I0227 18:38:00.145026 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed2293ce-4fd2-4674-81e9-0e63c000ab21" containerName="extract-content" Feb 27 18:38:00 crc kubenswrapper[4700]: E0227 18:38:00.145051 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed2293ce-4fd2-4674-81e9-0e63c000ab21" containerName="registry-server" Feb 27 18:38:00 crc kubenswrapper[4700]: I0227 18:38:00.145058 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed2293ce-4fd2-4674-81e9-0e63c000ab21" containerName="registry-server" Feb 27 18:38:00 crc kubenswrapper[4700]: I0227 18:38:00.145256 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed2293ce-4fd2-4674-81e9-0e63c000ab21" containerName="registry-server" Feb 27 18:38:00 crc kubenswrapper[4700]: I0227 18:38:00.145902 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536958-tdbth" Feb 27 18:38:00 crc kubenswrapper[4700]: I0227 18:38:00.169659 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536958-tdbth"] Feb 27 18:38:00 crc kubenswrapper[4700]: I0227 18:38:00.291789 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kj5x4\" (UniqueName: \"kubernetes.io/projected/b103b0e5-f0f5-4659-87d5-40a4d4edf5b7-kube-api-access-kj5x4\") pod \"auto-csr-approver-29536958-tdbth\" (UID: \"b103b0e5-f0f5-4659-87d5-40a4d4edf5b7\") " pod="openshift-infra/auto-csr-approver-29536958-tdbth" Feb 27 18:38:00 crc kubenswrapper[4700]: I0227 18:38:00.393344 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kj5x4\" (UniqueName: \"kubernetes.io/projected/b103b0e5-f0f5-4659-87d5-40a4d4edf5b7-kube-api-access-kj5x4\") pod \"auto-csr-approver-29536958-tdbth\" (UID: \"b103b0e5-f0f5-4659-87d5-40a4d4edf5b7\") " pod="openshift-infra/auto-csr-approver-29536958-tdbth" Feb 27 18:38:00 crc kubenswrapper[4700]: I0227 18:38:00.417240 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kj5x4\" (UniqueName: \"kubernetes.io/projected/b103b0e5-f0f5-4659-87d5-40a4d4edf5b7-kube-api-access-kj5x4\") pod \"auto-csr-approver-29536958-tdbth\" (UID: \"b103b0e5-f0f5-4659-87d5-40a4d4edf5b7\") " pod="openshift-infra/auto-csr-approver-29536958-tdbth" Feb 27 18:38:00 crc kubenswrapper[4700]: I0227 18:38:00.464520 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536958-tdbth" Feb 27 18:38:00 crc kubenswrapper[4700]: I0227 18:38:00.941287 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536958-tdbth"] Feb 27 18:38:01 crc kubenswrapper[4700]: I0227 18:38:01.257179 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536958-tdbth" event={"ID":"b103b0e5-f0f5-4659-87d5-40a4d4edf5b7","Type":"ContainerStarted","Data":"3ea823db4a03650ea5aa3074b1c78aa09656ab3ae15e5ca1e0e7ff95fecaad5d"} Feb 27 18:38:01 crc kubenswrapper[4700]: I0227 18:38:01.980683 4700 scope.go:117] "RemoveContainer" containerID="c7ad608c50ddee2aac4992df7a6ccaaeda7c737e2a0d2fad5282b4482eac9e50" Feb 27 18:38:01 crc kubenswrapper[4700]: E0227 18:38:01.981223 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:38:02 crc kubenswrapper[4700]: I0227 18:38:02.006536 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-6xcnk_fc05be47-71de-4f97-b1f2-b58bacfc752e/cert-manager-controller/0.log" Feb 27 18:38:02 crc kubenswrapper[4700]: I0227 18:38:02.221829 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-fdwz4_c00612dd-19ac-4a4b-870f-d1e70fa1604f/cert-manager-cainjector/0.log" Feb 27 18:38:02 crc kubenswrapper[4700]: I0227 18:38:02.274257 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-szl9g_e1d9b74a-c2c9-4654-b440-b4eb4f2deef1/cert-manager-webhook/0.log" Feb 27 18:38:03 crc kubenswrapper[4700]: I0227 18:38:03.275215 4700 generic.go:334] "Generic (PLEG): container finished" podID="b103b0e5-f0f5-4659-87d5-40a4d4edf5b7" containerID="0b22d4489434c448b1d3fdc0c20aebee56e79aca5fd52010bde72614dce6bcf6" exitCode=0 Feb 27 18:38:03 crc kubenswrapper[4700]: I0227 18:38:03.275303 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536958-tdbth" event={"ID":"b103b0e5-f0f5-4659-87d5-40a4d4edf5b7","Type":"ContainerDied","Data":"0b22d4489434c448b1d3fdc0c20aebee56e79aca5fd52010bde72614dce6bcf6"} Feb 27 18:38:04 crc kubenswrapper[4700]: I0227 18:38:04.717794 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536958-tdbth" Feb 27 18:38:04 crc kubenswrapper[4700]: I0227 18:38:04.780933 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kj5x4\" (UniqueName: \"kubernetes.io/projected/b103b0e5-f0f5-4659-87d5-40a4d4edf5b7-kube-api-access-kj5x4\") pod \"b103b0e5-f0f5-4659-87d5-40a4d4edf5b7\" (UID: \"b103b0e5-f0f5-4659-87d5-40a4d4edf5b7\") " Feb 27 18:38:04 crc kubenswrapper[4700]: I0227 18:38:04.787152 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b103b0e5-f0f5-4659-87d5-40a4d4edf5b7-kube-api-access-kj5x4" (OuterVolumeSpecName: "kube-api-access-kj5x4") pod "b103b0e5-f0f5-4659-87d5-40a4d4edf5b7" (UID: "b103b0e5-f0f5-4659-87d5-40a4d4edf5b7"). InnerVolumeSpecName "kube-api-access-kj5x4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:38:04 crc kubenswrapper[4700]: I0227 18:38:04.883060 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kj5x4\" (UniqueName: \"kubernetes.io/projected/b103b0e5-f0f5-4659-87d5-40a4d4edf5b7-kube-api-access-kj5x4\") on node \"crc\" DevicePath \"\"" Feb 27 18:38:04 crc kubenswrapper[4700]: E0227 18:38:04.987477 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:38:05 crc kubenswrapper[4700]: I0227 18:38:05.306635 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536958-tdbth" event={"ID":"b103b0e5-f0f5-4659-87d5-40a4d4edf5b7","Type":"ContainerDied","Data":"3ea823db4a03650ea5aa3074b1c78aa09656ab3ae15e5ca1e0e7ff95fecaad5d"} Feb 27 18:38:05 crc kubenswrapper[4700]: I0227 18:38:05.307494 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ea823db4a03650ea5aa3074b1c78aa09656ab3ae15e5ca1e0e7ff95fecaad5d" Feb 27 18:38:05 crc kubenswrapper[4700]: I0227 18:38:05.307436 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536958-tdbth" Feb 27 18:38:05 crc kubenswrapper[4700]: I0227 18:38:05.792944 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536952-p4tmr"] Feb 27 18:38:05 crc kubenswrapper[4700]: I0227 18:38:05.800018 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536952-p4tmr"] Feb 27 18:38:07 crc kubenswrapper[4700]: I0227 18:38:07.000563 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38cf4801-ea2b-4538-8417-ae67e1668122" path="/var/lib/kubelet/pods/38cf4801-ea2b-4538-8417-ae67e1668122/volumes" Feb 27 18:38:14 crc kubenswrapper[4700]: I0227 18:38:14.986176 4700 scope.go:117] "RemoveContainer" containerID="c7ad608c50ddee2aac4992df7a6ccaaeda7c737e2a0d2fad5282b4482eac9e50" Feb 27 18:38:14 crc kubenswrapper[4700]: E0227 18:38:14.987556 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:38:17 crc kubenswrapper[4700]: E0227 18:38:17.984076 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:38:18 crc kubenswrapper[4700]: I0227 18:38:18.778128 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5dcbbd79cf-m9hcq_fbb064be-36df-4a47-8c0e-9ffdd43d1123/nmstate-console-plugin/0.log" Feb 27 18:38:18 crc kubenswrapper[4700]: I0227 18:38:18.942599 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-85tkk_0686b736-3cb6-4fb9-8710-8359e2066637/nmstate-handler/0.log" Feb 27 18:38:19 crc kubenswrapper[4700]: I0227 18:38:19.080520 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-69594cc75-znbzd_47c2c9bd-d6c8-4b33-974c-ab0e57ea05fc/kube-rbac-proxy/0.log" Feb 27 18:38:19 crc kubenswrapper[4700]: I0227 18:38:19.124004 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-69594cc75-znbzd_47c2c9bd-d6c8-4b33-974c-ab0e57ea05fc/nmstate-metrics/0.log" Feb 27 18:38:19 crc kubenswrapper[4700]: I0227 18:38:19.227239 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-75c5dccd6c-6rrfb_3dc864b3-a8d5-4587-9356-ecffe805f637/nmstate-operator/0.log" Feb 27 18:38:19 crc kubenswrapper[4700]: I0227 18:38:19.301388 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-786f45cff4-gg78f_97a282f2-3d9c-4911-bb64-4593b586bb4e/nmstate-webhook/0.log" Feb 27 18:38:26 crc kubenswrapper[4700]: I0227 18:38:26.982550 4700 scope.go:117] "RemoveContainer" containerID="c7ad608c50ddee2aac4992df7a6ccaaeda7c737e2a0d2fad5282b4482eac9e50" Feb 27 18:38:26 crc kubenswrapper[4700]: E0227 18:38:26.983985 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:38:32 crc kubenswrapper[4700]: E0227 18:38:32.985108 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:38:37 crc kubenswrapper[4700]: I0227 18:38:37.669330 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-mw89d_530391a5-04d2-480a-8a3d-99d343188028/prometheus-operator/0.log" Feb 27 18:38:37 crc kubenswrapper[4700]: I0227 18:38:37.832707 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6894b8c5db-7gblc_fd6f868c-c0c2-4d1f-ac0a-8f81da9db16f/prometheus-operator-admission-webhook/0.log" Feb 27 18:38:37 crc kubenswrapper[4700]: I0227 18:38:37.906285 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6894b8c5db-82ptt_97b66bd1-71dd-493c-a8cf-86b371409c8c/prometheus-operator-admission-webhook/0.log" Feb 27 18:38:38 crc kubenswrapper[4700]: I0227 18:38:38.047646 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-hc2rv_eaaadf56-7fe6-4df6-a59b-7cdcab1dc47b/operator/0.log" Feb 27 18:38:38 crc kubenswrapper[4700]: I0227 18:38:38.086156 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-z6mk4_c4b67ce7-9585-4c61-a8d7-bf37151fb6ea/perses-operator/0.log" Feb 27 18:38:39 crc kubenswrapper[4700]: I0227 18:38:39.981252 4700 scope.go:117] "RemoveContainer" containerID="c7ad608c50ddee2aac4992df7a6ccaaeda7c737e2a0d2fad5282b4482eac9e50" Feb 27 18:38:39 crc kubenswrapper[4700]: E0227 18:38:39.981846 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:38:43 crc kubenswrapper[4700]: E0227 18:38:43.986493 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:38:54 crc kubenswrapper[4700]: I0227 18:38:54.980815 4700 scope.go:117] "RemoveContainer" containerID="c7ad608c50ddee2aac4992df7a6ccaaeda7c737e2a0d2fad5282b4482eac9e50" Feb 27 18:38:54 crc kubenswrapper[4700]: E0227 18:38:54.981492 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:38:55 crc kubenswrapper[4700]: I0227 18:38:55.842389 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-86ddb6bd46-fvdl8_16cfcb53-eb3f-49d0-9cbc-f83b73527d8b/kube-rbac-proxy/0.log" Feb 27 18:38:55 crc kubenswrapper[4700]: E0227 18:38:55.989822 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:38:56 crc kubenswrapper[4700]: I0227 18:38:56.038930 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-86ddb6bd46-fvdl8_16cfcb53-eb3f-49d0-9cbc-f83b73527d8b/controller/0.log" Feb 27 18:38:56 crc kubenswrapper[4700]: I0227 18:38:56.098115 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j65dj_74b35197-b618-42e5-afba-88be4729f5b9/cp-frr-files/0.log" Feb 27 18:38:56 crc kubenswrapper[4700]: I0227 18:38:56.280604 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j65dj_74b35197-b618-42e5-afba-88be4729f5b9/cp-reloader/0.log" Feb 27 18:38:56 crc kubenswrapper[4700]: I0227 18:38:56.286627 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j65dj_74b35197-b618-42e5-afba-88be4729f5b9/cp-frr-files/0.log" Feb 27 18:38:56 crc kubenswrapper[4700]: I0227 18:38:56.302787 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j65dj_74b35197-b618-42e5-afba-88be4729f5b9/cp-metrics/0.log" Feb 27 18:38:56 crc kubenswrapper[4700]: I0227 18:38:56.341281 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j65dj_74b35197-b618-42e5-afba-88be4729f5b9/cp-reloader/0.log" Feb 27 18:38:56 crc kubenswrapper[4700]: I0227 18:38:56.501594 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j65dj_74b35197-b618-42e5-afba-88be4729f5b9/cp-metrics/0.log" Feb 27 18:38:56 crc kubenswrapper[4700]: I0227 18:38:56.509240 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j65dj_74b35197-b618-42e5-afba-88be4729f5b9/cp-metrics/0.log" Feb 27 18:38:56 crc kubenswrapper[4700]: I0227 18:38:56.521986 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j65dj_74b35197-b618-42e5-afba-88be4729f5b9/cp-reloader/0.log" Feb 27 18:38:56 crc kubenswrapper[4700]: I0227 18:38:56.528439 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j65dj_74b35197-b618-42e5-afba-88be4729f5b9/cp-frr-files/0.log" Feb 27 18:38:56 crc kubenswrapper[4700]: I0227 18:38:56.732187 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j65dj_74b35197-b618-42e5-afba-88be4729f5b9/cp-frr-files/0.log" Feb 27 18:38:56 crc kubenswrapper[4700]: I0227 18:38:56.734865 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j65dj_74b35197-b618-42e5-afba-88be4729f5b9/cp-reloader/0.log" Feb 27 18:38:56 crc kubenswrapper[4700]: I0227 18:38:56.794866 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j65dj_74b35197-b618-42e5-afba-88be4729f5b9/cp-metrics/0.log" Feb 27 18:38:56 crc kubenswrapper[4700]: I0227 18:38:56.809107 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j65dj_74b35197-b618-42e5-afba-88be4729f5b9/controller/0.log" Feb 27 18:38:56 crc kubenswrapper[4700]: I0227 18:38:56.945835 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j65dj_74b35197-b618-42e5-afba-88be4729f5b9/frr-metrics/0.log" Feb 27 18:38:56 crc kubenswrapper[4700]: I0227 18:38:56.997431 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j65dj_74b35197-b618-42e5-afba-88be4729f5b9/kube-rbac-proxy/0.log" Feb 27 18:38:57 crc kubenswrapper[4700]: I0227 18:38:57.028492 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j65dj_74b35197-b618-42e5-afba-88be4729f5b9/kube-rbac-proxy-frr/0.log" Feb 27 18:38:57 crc kubenswrapper[4700]: I0227 18:38:57.160495 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j65dj_74b35197-b618-42e5-afba-88be4729f5b9/reloader/0.log" Feb 27 18:38:57 crc kubenswrapper[4700]: I0227 18:38:57.280050 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7f989f654f-kf8ck_0aa251ee-47d2-4917-b18b-00ed276a6d24/frr-k8s-webhook-server/0.log" Feb 27 18:38:57 crc kubenswrapper[4700]: I0227 18:38:57.432208 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-cc5d6c7cf-x6qzc_1d823d65-036b-4040-af8a-9b6698071d32/manager/0.log" Feb 27 18:38:57 crc kubenswrapper[4700]: I0227 18:38:57.525186 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-f4987db88-t66kg_520ca54b-383f-4ca4-819a-7922a704404e/webhook-server/0.log" Feb 27 18:38:57 crc kubenswrapper[4700]: I0227 18:38:57.775832 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-7ktps_6108b3df-5dda-48cd-8cb4-55b36e5843aa/kube-rbac-proxy/0.log" Feb 27 18:38:58 crc kubenswrapper[4700]: I0227 18:38:58.222866 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-7ktps_6108b3df-5dda-48cd-8cb4-55b36e5843aa/speaker/0.log" Feb 27 18:38:58 crc kubenswrapper[4700]: I0227 18:38:58.893173 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j65dj_74b35197-b618-42e5-afba-88be4729f5b9/frr/0.log" Feb 27 18:39:05 crc kubenswrapper[4700]: I0227 18:39:05.981134 4700 scope.go:117] "RemoveContainer" containerID="c7ad608c50ddee2aac4992df7a6ccaaeda7c737e2a0d2fad5282b4482eac9e50" Feb 27 18:39:05 crc kubenswrapper[4700]: E0227 18:39:05.982137 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:39:10 crc kubenswrapper[4700]: E0227 18:39:10.997398 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:39:14 crc kubenswrapper[4700]: I0227 18:39:14.480836 4700 scope.go:117] "RemoveContainer" containerID="b4e1fe01b7eeeb7aff8e5c7ec44de422fdc079aa27f51ebade305bc61c0305db" Feb 27 18:39:15 crc kubenswrapper[4700]: I0227 18:39:15.416327 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp_e8d807a9-fc26-4417-94db-051e3e890be4/util/0.log" Feb 27 18:39:16 crc kubenswrapper[4700]: I0227 18:39:16.045729 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp_e8d807a9-fc26-4417-94db-051e3e890be4/util/0.log" Feb 27 18:39:16 crc kubenswrapper[4700]: I0227 18:39:16.077893 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp_e8d807a9-fc26-4417-94db-051e3e890be4/pull/0.log" Feb 27 18:39:16 crc kubenswrapper[4700]: I0227 18:39:16.090554 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp_e8d807a9-fc26-4417-94db-051e3e890be4/pull/0.log" Feb 27 18:39:16 crc kubenswrapper[4700]: I0227 18:39:16.287486 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp_e8d807a9-fc26-4417-94db-051e3e890be4/pull/0.log" Feb 27 18:39:16 crc kubenswrapper[4700]: I0227 18:39:16.309512 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp_e8d807a9-fc26-4417-94db-051e3e890be4/util/0.log" Feb 27 18:39:16 crc kubenswrapper[4700]: I0227 18:39:16.322804 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82ccmcp_e8d807a9-fc26-4417-94db-051e3e890be4/extract/0.log" Feb 27 18:39:16 crc kubenswrapper[4700]: I0227 18:39:16.479763 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ct5kj_0cc8109b-159b-4c74-90af-110d79f1b6f4/util/0.log" Feb 27 18:39:16 crc kubenswrapper[4700]: I0227 18:39:16.604803 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ct5kj_0cc8109b-159b-4c74-90af-110d79f1b6f4/util/0.log" Feb 27 18:39:16 crc kubenswrapper[4700]: I0227 18:39:16.663255 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ct5kj_0cc8109b-159b-4c74-90af-110d79f1b6f4/pull/0.log" Feb 27 18:39:16 crc kubenswrapper[4700]: I0227 18:39:16.688856 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ct5kj_0cc8109b-159b-4c74-90af-110d79f1b6f4/pull/0.log" Feb 27 18:39:16 crc kubenswrapper[4700]: I0227 18:39:16.855897 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ct5kj_0cc8109b-159b-4c74-90af-110d79f1b6f4/util/0.log" Feb 27 18:39:16 crc kubenswrapper[4700]: I0227 18:39:16.858195 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ct5kj_0cc8109b-159b-4c74-90af-110d79f1b6f4/extract/0.log" Feb 27 18:39:16 crc kubenswrapper[4700]: I0227 18:39:16.867005 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ct5kj_0cc8109b-159b-4c74-90af-110d79f1b6f4/pull/0.log" Feb 27 18:39:17 crc kubenswrapper[4700]: I0227 18:39:17.018259 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qmhw5_b1a1f4b1-e13f-4b11-bdeb-5d815af7b98b/extract-utilities/0.log" Feb 27 18:39:17 crc kubenswrapper[4700]: I0227 18:39:17.235738 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qmhw5_b1a1f4b1-e13f-4b11-bdeb-5d815af7b98b/extract-content/0.log" Feb 27 18:39:17 crc kubenswrapper[4700]: I0227 18:39:17.253877 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qmhw5_b1a1f4b1-e13f-4b11-bdeb-5d815af7b98b/extract-content/0.log" Feb 27 18:39:17 crc kubenswrapper[4700]: I0227 18:39:17.257346 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qmhw5_b1a1f4b1-e13f-4b11-bdeb-5d815af7b98b/extract-utilities/0.log" Feb 27 18:39:17 crc kubenswrapper[4700]: I0227 18:39:17.371840 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qmhw5_b1a1f4b1-e13f-4b11-bdeb-5d815af7b98b/extract-utilities/0.log" Feb 27 18:39:17 crc kubenswrapper[4700]: I0227 18:39:17.399183 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qmhw5_b1a1f4b1-e13f-4b11-bdeb-5d815af7b98b/extract-content/0.log" Feb 27 18:39:17 crc kubenswrapper[4700]: I0227 18:39:17.556214 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-98xlg_622df037-57ba-4a94-8e21-d78cfb08a79c/extract-utilities/0.log" Feb 27 18:39:17 crc kubenswrapper[4700]: I0227 18:39:17.799718 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-98xlg_622df037-57ba-4a94-8e21-d78cfb08a79c/extract-utilities/0.log" Feb 27 18:39:17 crc kubenswrapper[4700]: I0227 18:39:17.800291 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-98xlg_622df037-57ba-4a94-8e21-d78cfb08a79c/extract-content/0.log" Feb 27 18:39:17 crc kubenswrapper[4700]: I0227 18:39:17.807860 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-98xlg_622df037-57ba-4a94-8e21-d78cfb08a79c/extract-content/0.log" Feb 27 18:39:17 crc kubenswrapper[4700]: I0227 18:39:17.947974 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qmhw5_b1a1f4b1-e13f-4b11-bdeb-5d815af7b98b/registry-server/0.log" Feb 27 18:39:18 crc kubenswrapper[4700]: I0227 18:39:18.008205 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-98xlg_622df037-57ba-4a94-8e21-d78cfb08a79c/extract-utilities/0.log" Feb 27 18:39:18 crc kubenswrapper[4700]: I0227 18:39:18.027553 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-98xlg_622df037-57ba-4a94-8e21-d78cfb08a79c/extract-content/0.log" Feb 27 18:39:18 crc kubenswrapper[4700]: I0227 18:39:18.259717 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d_ccc245e7-023f-4d49-99d0-9ebc992a9ccc/util/0.log" Feb 27 18:39:18 crc kubenswrapper[4700]: I0227 18:39:18.401921 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d_ccc245e7-023f-4d49-99d0-9ebc992a9ccc/util/0.log" Feb 27 18:39:18 crc kubenswrapper[4700]: I0227 18:39:18.478516 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d_ccc245e7-023f-4d49-99d0-9ebc992a9ccc/pull/0.log" Feb 27 18:39:18 crc kubenswrapper[4700]: I0227 18:39:18.485373 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d_ccc245e7-023f-4d49-99d0-9ebc992a9ccc/pull/0.log" Feb 27 18:39:18 crc kubenswrapper[4700]: I0227 18:39:18.666608 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-98xlg_622df037-57ba-4a94-8e21-d78cfb08a79c/registry-server/0.log" Feb 27 18:39:18 crc kubenswrapper[4700]: I0227 18:39:18.705317 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d_ccc245e7-023f-4d49-99d0-9ebc992a9ccc/util/0.log" Feb 27 18:39:18 crc kubenswrapper[4700]: I0227 18:39:18.706596 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d_ccc245e7-023f-4d49-99d0-9ebc992a9ccc/pull/0.log" Feb 27 18:39:18 crc kubenswrapper[4700]: I0227 18:39:18.717876 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4xf54d_ccc245e7-023f-4d49-99d0-9ebc992a9ccc/extract/0.log" Feb 27 18:39:18 crc kubenswrapper[4700]: I0227 18:39:18.914988 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-wlsk4_aca7409e-ab2c-4d9b-a775-e82f2cd958d5/marketplace-operator/0.log" Feb 27 18:39:18 crc kubenswrapper[4700]: I0227 18:39:18.920304 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-sl4bz_53351da8-d63b-4af8-912e-9a7aa5da939a/extract-utilities/0.log" Feb 27 18:39:19 crc kubenswrapper[4700]: I0227 18:39:19.126227 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-sl4bz_53351da8-d63b-4af8-912e-9a7aa5da939a/extract-utilities/0.log" Feb 27 18:39:19 crc kubenswrapper[4700]: I0227 18:39:19.134548 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-sl4bz_53351da8-d63b-4af8-912e-9a7aa5da939a/extract-content/0.log" Feb 27 18:39:19 crc kubenswrapper[4700]: I0227 18:39:19.195928 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-sl4bz_53351da8-d63b-4af8-912e-9a7aa5da939a/extract-content/0.log" Feb 27 18:39:19 crc kubenswrapper[4700]: I0227 18:39:19.303689 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-sl4bz_53351da8-d63b-4af8-912e-9a7aa5da939a/extract-content/0.log" Feb 27 18:39:19 crc kubenswrapper[4700]: I0227 18:39:19.359446 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-sl4bz_53351da8-d63b-4af8-912e-9a7aa5da939a/extract-utilities/0.log" Feb 27 18:39:19 crc kubenswrapper[4700]: I0227 18:39:19.455058 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-sl4bz_53351da8-d63b-4af8-912e-9a7aa5da939a/registry-server/0.log" Feb 27 18:39:19 crc kubenswrapper[4700]: I0227 18:39:19.517746 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q8xmq_f9dae327-14da-40ce-8e98-b29453b0352f/extract-utilities/0.log" Feb 27 18:39:19 crc kubenswrapper[4700]: I0227 18:39:19.617985 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q8xmq_f9dae327-14da-40ce-8e98-b29453b0352f/extract-utilities/0.log" Feb 27 18:39:19 crc kubenswrapper[4700]: I0227 18:39:19.640658 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q8xmq_f9dae327-14da-40ce-8e98-b29453b0352f/extract-content/0.log" Feb 27 18:39:19 crc kubenswrapper[4700]: I0227 18:39:19.659609 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q8xmq_f9dae327-14da-40ce-8e98-b29453b0352f/extract-content/0.log" Feb 27 18:39:19 crc kubenswrapper[4700]: I0227 18:39:19.826884 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q8xmq_f9dae327-14da-40ce-8e98-b29453b0352f/extract-utilities/0.log" Feb 27 18:39:19 crc kubenswrapper[4700]: I0227 18:39:19.833330 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q8xmq_f9dae327-14da-40ce-8e98-b29453b0352f/extract-content/0.log" Feb 27 18:39:20 crc kubenswrapper[4700]: I0227 18:39:20.484839 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q8xmq_f9dae327-14da-40ce-8e98-b29453b0352f/registry-server/0.log" Feb 27 18:39:20 crc kubenswrapper[4700]: I0227 18:39:20.987651 4700 scope.go:117] "RemoveContainer" containerID="c7ad608c50ddee2aac4992df7a6ccaaeda7c737e2a0d2fad5282b4482eac9e50" Feb 27 18:39:20 crc kubenswrapper[4700]: E0227 18:39:20.988073 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:39:25 crc kubenswrapper[4700]: E0227 18:39:25.985683 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:39:33 crc kubenswrapper[4700]: I0227 18:39:33.981879 4700 scope.go:117] "RemoveContainer" containerID="c7ad608c50ddee2aac4992df7a6ccaaeda7c737e2a0d2fad5282b4482eac9e50" Feb 27 18:39:33 crc kubenswrapper[4700]: E0227 18:39:33.983082 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:39:36 crc kubenswrapper[4700]: I0227 18:39:36.331912 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6894b8c5db-7gblc_fd6f868c-c0c2-4d1f-ac0a-8f81da9db16f/prometheus-operator-admission-webhook/0.log" Feb 27 18:39:36 crc kubenswrapper[4700]: I0227 18:39:36.339254 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-mw89d_530391a5-04d2-480a-8a3d-99d343188028/prometheus-operator/0.log" Feb 27 18:39:36 crc kubenswrapper[4700]: I0227 18:39:36.389162 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6894b8c5db-82ptt_97b66bd1-71dd-493c-a8cf-86b371409c8c/prometheus-operator-admission-webhook/0.log" Feb 27 18:39:36 crc kubenswrapper[4700]: I0227 18:39:36.513472 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-hc2rv_eaaadf56-7fe6-4df6-a59b-7cdcab1dc47b/operator/0.log" Feb 27 18:39:36 crc kubenswrapper[4700]: I0227 18:39:36.537493 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-z6mk4_c4b67ce7-9585-4c61-a8d7-bf37151fb6ea/perses-operator/0.log" Feb 27 18:39:37 crc kubenswrapper[4700]: E0227 18:39:37.983828 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:39:40 crc kubenswrapper[4700]: E0227 18:39:40.429149 4700 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.222:44846->38.102.83.222:41369: write tcp 38.102.83.222:44846->38.102.83.222:41369: write: broken pipe Feb 27 18:39:48 crc kubenswrapper[4700]: I0227 18:39:48.981928 4700 scope.go:117] "RemoveContainer" containerID="c7ad608c50ddee2aac4992df7a6ccaaeda7c737e2a0d2fad5282b4482eac9e50" Feb 27 18:39:48 crc kubenswrapper[4700]: E0227 18:39:48.983054 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:39:48 crc kubenswrapper[4700]: E0227 18:39:48.984601 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:39:52 crc kubenswrapper[4700]: E0227 18:39:52.400864 4700 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.222:41558->38.102.83.222:41369: write tcp 38.102.83.222:41558->38.102.83.222:41369: write: broken pipe Feb 27 18:40:00 crc kubenswrapper[4700]: I0227 18:40:00.161276 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536960-z9fjb"] Feb 27 18:40:00 crc kubenswrapper[4700]: E0227 18:40:00.162718 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b103b0e5-f0f5-4659-87d5-40a4d4edf5b7" containerName="oc" Feb 27 18:40:00 crc kubenswrapper[4700]: I0227 18:40:00.162735 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="b103b0e5-f0f5-4659-87d5-40a4d4edf5b7" containerName="oc" Feb 27 18:40:00 crc kubenswrapper[4700]: I0227 18:40:00.162969 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="b103b0e5-f0f5-4659-87d5-40a4d4edf5b7" containerName="oc" Feb 27 18:40:00 crc kubenswrapper[4700]: I0227 18:40:00.163908 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536960-z9fjb" Feb 27 18:40:00 crc kubenswrapper[4700]: I0227 18:40:00.173952 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536960-z9fjb"] Feb 27 18:40:00 crc kubenswrapper[4700]: I0227 18:40:00.260849 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rm98w\" (UniqueName: \"kubernetes.io/projected/2960fff4-655f-432d-a3d0-8e2090558d8b-kube-api-access-rm98w\") pod \"auto-csr-approver-29536960-z9fjb\" (UID: \"2960fff4-655f-432d-a3d0-8e2090558d8b\") " pod="openshift-infra/auto-csr-approver-29536960-z9fjb" Feb 27 18:40:00 crc kubenswrapper[4700]: I0227 18:40:00.362954 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rm98w\" (UniqueName: \"kubernetes.io/projected/2960fff4-655f-432d-a3d0-8e2090558d8b-kube-api-access-rm98w\") pod \"auto-csr-approver-29536960-z9fjb\" (UID: \"2960fff4-655f-432d-a3d0-8e2090558d8b\") " pod="openshift-infra/auto-csr-approver-29536960-z9fjb" Feb 27 18:40:00 crc kubenswrapper[4700]: I0227 18:40:00.394022 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rm98w\" (UniqueName: \"kubernetes.io/projected/2960fff4-655f-432d-a3d0-8e2090558d8b-kube-api-access-rm98w\") pod \"auto-csr-approver-29536960-z9fjb\" (UID: \"2960fff4-655f-432d-a3d0-8e2090558d8b\") " pod="openshift-infra/auto-csr-approver-29536960-z9fjb" Feb 27 18:40:00 crc kubenswrapper[4700]: I0227 18:40:00.488754 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536960-z9fjb" Feb 27 18:40:01 crc kubenswrapper[4700]: I0227 18:40:00.994229 4700 scope.go:117] "RemoveContainer" containerID="c7ad608c50ddee2aac4992df7a6ccaaeda7c737e2a0d2fad5282b4482eac9e50" Feb 27 18:40:01 crc kubenswrapper[4700]: E0227 18:40:00.995757 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:40:01 crc kubenswrapper[4700]: I0227 18:40:01.005533 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536960-z9fjb"] Feb 27 18:40:01 crc kubenswrapper[4700]: I0227 18:40:01.028025 4700 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 18:40:01 crc kubenswrapper[4700]: I0227 18:40:01.831583 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536960-z9fjb" event={"ID":"2960fff4-655f-432d-a3d0-8e2090558d8b","Type":"ContainerStarted","Data":"3400c9bc95c40adfb4209af7a0ca3f7197cfcec8732e760308a545126c591058"} Feb 27 18:40:02 crc kubenswrapper[4700]: I0227 18:40:02.847007 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536960-z9fjb" event={"ID":"2960fff4-655f-432d-a3d0-8e2090558d8b","Type":"ContainerStarted","Data":"1516063b9488f41ba916b1c4e0f7a55394b7ac28e08688f60a4f2308a9c540bf"} Feb 27 18:40:02 crc kubenswrapper[4700]: I0227 18:40:02.884526 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536960-z9fjb" podStartSLOduration=1.7364326879999998 podStartE2EDuration="2.884495555s" podCreationTimestamp="2026-02-27 18:40:00 +0000 UTC" firstStartedPulling="2026-02-27 18:40:01.027815208 +0000 UTC m=+5961.013127955" lastFinishedPulling="2026-02-27 18:40:02.175878085 +0000 UTC m=+5962.161190822" observedRunningTime="2026-02-27 18:40:02.876540548 +0000 UTC m=+5962.861853325" watchObservedRunningTime="2026-02-27 18:40:02.884495555 +0000 UTC m=+5962.869808342" Feb 27 18:40:03 crc kubenswrapper[4700]: I0227 18:40:03.859832 4700 generic.go:334] "Generic (PLEG): container finished" podID="2960fff4-655f-432d-a3d0-8e2090558d8b" containerID="1516063b9488f41ba916b1c4e0f7a55394b7ac28e08688f60a4f2308a9c540bf" exitCode=0 Feb 27 18:40:03 crc kubenswrapper[4700]: I0227 18:40:03.859937 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536960-z9fjb" event={"ID":"2960fff4-655f-432d-a3d0-8e2090558d8b","Type":"ContainerDied","Data":"1516063b9488f41ba916b1c4e0f7a55394b7ac28e08688f60a4f2308a9c540bf"} Feb 27 18:40:03 crc kubenswrapper[4700]: E0227 18:40:03.982977 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29536934-rw77f" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" Feb 27 18:40:05 crc kubenswrapper[4700]: I0227 18:40:05.560976 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536960-z9fjb" Feb 27 18:40:05 crc kubenswrapper[4700]: I0227 18:40:05.620666 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rm98w\" (UniqueName: \"kubernetes.io/projected/2960fff4-655f-432d-a3d0-8e2090558d8b-kube-api-access-rm98w\") pod \"2960fff4-655f-432d-a3d0-8e2090558d8b\" (UID: \"2960fff4-655f-432d-a3d0-8e2090558d8b\") " Feb 27 18:40:05 crc kubenswrapper[4700]: I0227 18:40:05.627559 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2960fff4-655f-432d-a3d0-8e2090558d8b-kube-api-access-rm98w" (OuterVolumeSpecName: "kube-api-access-rm98w") pod "2960fff4-655f-432d-a3d0-8e2090558d8b" (UID: "2960fff4-655f-432d-a3d0-8e2090558d8b"). InnerVolumeSpecName "kube-api-access-rm98w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:40:05 crc kubenswrapper[4700]: I0227 18:40:05.722906 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rm98w\" (UniqueName: \"kubernetes.io/projected/2960fff4-655f-432d-a3d0-8e2090558d8b-kube-api-access-rm98w\") on node \"crc\" DevicePath \"\"" Feb 27 18:40:05 crc kubenswrapper[4700]: I0227 18:40:05.884961 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536960-z9fjb" event={"ID":"2960fff4-655f-432d-a3d0-8e2090558d8b","Type":"ContainerDied","Data":"3400c9bc95c40adfb4209af7a0ca3f7197cfcec8732e760308a545126c591058"} Feb 27 18:40:05 crc kubenswrapper[4700]: I0227 18:40:05.885267 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3400c9bc95c40adfb4209af7a0ca3f7197cfcec8732e760308a545126c591058" Feb 27 18:40:05 crc kubenswrapper[4700]: I0227 18:40:05.885072 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536960-z9fjb" Feb 27 18:40:06 crc kubenswrapper[4700]: I0227 18:40:06.671259 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536954-6cc9p"] Feb 27 18:40:06 crc kubenswrapper[4700]: I0227 18:40:06.681294 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536954-6cc9p"] Feb 27 18:40:06 crc kubenswrapper[4700]: I0227 18:40:06.999764 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3dc264f3-4cbc-4c56-b9f8-4090451d6f8d" path="/var/lib/kubelet/pods/3dc264f3-4cbc-4c56-b9f8-4090451d6f8d/volumes" Feb 27 18:40:13 crc kubenswrapper[4700]: I0227 18:40:13.982329 4700 scope.go:117] "RemoveContainer" containerID="c7ad608c50ddee2aac4992df7a6ccaaeda7c737e2a0d2fad5282b4482eac9e50" Feb 27 18:40:13 crc kubenswrapper[4700]: E0227 18:40:13.983673 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:40:14 crc kubenswrapper[4700]: I0227 18:40:14.553753 4700 scope.go:117] "RemoveContainer" containerID="a20d6b3024c7096e4d6dbd8f38beca19c244a4c825e28672f18614b6a3a89ef5" Feb 27 18:40:18 crc kubenswrapper[4700]: I0227 18:40:18.055339 4700 generic.go:334] "Generic (PLEG): container finished" podID="caa4e57d-b02d-4d65-944d-451b8294f884" containerID="79244f56f855c6bc88df8ccf332c926d06aefac46e3c36cd557d9452928c97b7" exitCode=0 Feb 27 18:40:18 crc kubenswrapper[4700]: I0227 18:40:18.055440 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536934-rw77f" event={"ID":"caa4e57d-b02d-4d65-944d-451b8294f884","Type":"ContainerDied","Data":"79244f56f855c6bc88df8ccf332c926d06aefac46e3c36cd557d9452928c97b7"} Feb 27 18:40:19 crc kubenswrapper[4700]: I0227 18:40:19.517180 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536934-rw77f" Feb 27 18:40:19 crc kubenswrapper[4700]: I0227 18:40:19.665193 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qkjfz\" (UniqueName: \"kubernetes.io/projected/caa4e57d-b02d-4d65-944d-451b8294f884-kube-api-access-qkjfz\") pod \"caa4e57d-b02d-4d65-944d-451b8294f884\" (UID: \"caa4e57d-b02d-4d65-944d-451b8294f884\") " Feb 27 18:40:19 crc kubenswrapper[4700]: I0227 18:40:19.674624 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/caa4e57d-b02d-4d65-944d-451b8294f884-kube-api-access-qkjfz" (OuterVolumeSpecName: "kube-api-access-qkjfz") pod "caa4e57d-b02d-4d65-944d-451b8294f884" (UID: "caa4e57d-b02d-4d65-944d-451b8294f884"). InnerVolumeSpecName "kube-api-access-qkjfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:40:19 crc kubenswrapper[4700]: I0227 18:40:19.769010 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qkjfz\" (UniqueName: \"kubernetes.io/projected/caa4e57d-b02d-4d65-944d-451b8294f884-kube-api-access-qkjfz\") on node \"crc\" DevicePath \"\"" Feb 27 18:40:20 crc kubenswrapper[4700]: I0227 18:40:20.080915 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536934-rw77f" event={"ID":"caa4e57d-b02d-4d65-944d-451b8294f884","Type":"ContainerDied","Data":"8462629da14118f9f26d64718724c34cf1a6253798bba5deabaac00477a43029"} Feb 27 18:40:20 crc kubenswrapper[4700]: I0227 18:40:20.080952 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8462629da14118f9f26d64718724c34cf1a6253798bba5deabaac00477a43029" Feb 27 18:40:20 crc kubenswrapper[4700]: I0227 18:40:20.080985 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536934-rw77f" Feb 27 18:40:20 crc kubenswrapper[4700]: I0227 18:40:20.605168 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536934-rw77f"] Feb 27 18:40:20 crc kubenswrapper[4700]: I0227 18:40:20.617783 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536934-rw77f"] Feb 27 18:40:21 crc kubenswrapper[4700]: I0227 18:40:21.037694 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" path="/var/lib/kubelet/pods/caa4e57d-b02d-4d65-944d-451b8294f884/volumes" Feb 27 18:40:25 crc kubenswrapper[4700]: I0227 18:40:25.983034 4700 scope.go:117] "RemoveContainer" containerID="c7ad608c50ddee2aac4992df7a6ccaaeda7c737e2a0d2fad5282b4482eac9e50" Feb 27 18:40:25 crc kubenswrapper[4700]: E0227 18:40:25.983800 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:40:41 crc kubenswrapper[4700]: I0227 18:40:40.997892 4700 scope.go:117] "RemoveContainer" containerID="c7ad608c50ddee2aac4992df7a6ccaaeda7c737e2a0d2fad5282b4482eac9e50" Feb 27 18:40:41 crc kubenswrapper[4700]: E0227 18:40:40.998968 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:40:54 crc kubenswrapper[4700]: I0227 18:40:54.981404 4700 scope.go:117] "RemoveContainer" containerID="c7ad608c50ddee2aac4992df7a6ccaaeda7c737e2a0d2fad5282b4482eac9e50" Feb 27 18:40:54 crc kubenswrapper[4700]: E0227 18:40:54.982576 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:41:04 crc kubenswrapper[4700]: I0227 18:41:04.199311 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5pqm8"] Feb 27 18:41:04 crc kubenswrapper[4700]: E0227 18:41:04.200451 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" containerName="oc" Feb 27 18:41:04 crc kubenswrapper[4700]: I0227 18:41:04.200476 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" containerName="oc" Feb 27 18:41:04 crc kubenswrapper[4700]: E0227 18:41:04.200519 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2960fff4-655f-432d-a3d0-8e2090558d8b" containerName="oc" Feb 27 18:41:04 crc kubenswrapper[4700]: I0227 18:41:04.200527 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="2960fff4-655f-432d-a3d0-8e2090558d8b" containerName="oc" Feb 27 18:41:04 crc kubenswrapper[4700]: I0227 18:41:04.200721 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="caa4e57d-b02d-4d65-944d-451b8294f884" containerName="oc" Feb 27 18:41:04 crc kubenswrapper[4700]: I0227 18:41:04.200735 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="2960fff4-655f-432d-a3d0-8e2090558d8b" containerName="oc" Feb 27 18:41:04 crc kubenswrapper[4700]: I0227 18:41:04.202418 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5pqm8" Feb 27 18:41:04 crc kubenswrapper[4700]: I0227 18:41:04.229724 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5pqm8"] Feb 27 18:41:04 crc kubenswrapper[4700]: I0227 18:41:04.334227 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckcx7\" (UniqueName: \"kubernetes.io/projected/f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670-kube-api-access-ckcx7\") pod \"redhat-operators-5pqm8\" (UID: \"f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670\") " pod="openshift-marketplace/redhat-operators-5pqm8" Feb 27 18:41:04 crc kubenswrapper[4700]: I0227 18:41:04.334398 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670-catalog-content\") pod \"redhat-operators-5pqm8\" (UID: \"f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670\") " pod="openshift-marketplace/redhat-operators-5pqm8" Feb 27 18:41:04 crc kubenswrapper[4700]: I0227 18:41:04.334531 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670-utilities\") pod \"redhat-operators-5pqm8\" (UID: \"f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670\") " pod="openshift-marketplace/redhat-operators-5pqm8" Feb 27 18:41:04 crc kubenswrapper[4700]: I0227 18:41:04.437179 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670-catalog-content\") pod \"redhat-operators-5pqm8\" (UID: \"f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670\") " pod="openshift-marketplace/redhat-operators-5pqm8" Feb 27 18:41:04 crc kubenswrapper[4700]: I0227 18:41:04.437313 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670-utilities\") pod \"redhat-operators-5pqm8\" (UID: \"f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670\") " pod="openshift-marketplace/redhat-operators-5pqm8" Feb 27 18:41:04 crc kubenswrapper[4700]: I0227 18:41:04.437379 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckcx7\" (UniqueName: \"kubernetes.io/projected/f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670-kube-api-access-ckcx7\") pod \"redhat-operators-5pqm8\" (UID: \"f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670\") " pod="openshift-marketplace/redhat-operators-5pqm8" Feb 27 18:41:04 crc kubenswrapper[4700]: I0227 18:41:04.437837 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670-catalog-content\") pod \"redhat-operators-5pqm8\" (UID: \"f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670\") " pod="openshift-marketplace/redhat-operators-5pqm8" Feb 27 18:41:04 crc kubenswrapper[4700]: I0227 18:41:04.438201 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670-utilities\") pod \"redhat-operators-5pqm8\" (UID: \"f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670\") " pod="openshift-marketplace/redhat-operators-5pqm8" Feb 27 18:41:04 crc kubenswrapper[4700]: I0227 18:41:04.463336 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckcx7\" (UniqueName: \"kubernetes.io/projected/f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670-kube-api-access-ckcx7\") pod \"redhat-operators-5pqm8\" (UID: \"f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670\") " pod="openshift-marketplace/redhat-operators-5pqm8" Feb 27 18:41:04 crc kubenswrapper[4700]: I0227 18:41:04.536474 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5pqm8" Feb 27 18:41:05 crc kubenswrapper[4700]: I0227 18:41:05.048265 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5pqm8"] Feb 27 18:41:05 crc kubenswrapper[4700]: I0227 18:41:05.706018 4700 generic.go:334] "Generic (PLEG): container finished" podID="f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670" containerID="f5a62cf8d3a0b543c40f58460495274c3f16d8a3a47dc3c887ea09d40d6ea36c" exitCode=0 Feb 27 18:41:05 crc kubenswrapper[4700]: I0227 18:41:05.706110 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5pqm8" event={"ID":"f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670","Type":"ContainerDied","Data":"f5a62cf8d3a0b543c40f58460495274c3f16d8a3a47dc3c887ea09d40d6ea36c"} Feb 27 18:41:05 crc kubenswrapper[4700]: I0227 18:41:05.706260 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5pqm8" event={"ID":"f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670","Type":"ContainerStarted","Data":"a8912e2733f2ca58161c8fa264b6417554ad67333069c79b48d1b48fd4860201"} Feb 27 18:41:06 crc kubenswrapper[4700]: I0227 18:41:06.718282 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5pqm8" event={"ID":"f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670","Type":"ContainerStarted","Data":"9fc9801802539ab927f1ddb5a3637cf411a7dcefb8563790daf7e5b0847790fd"} Feb 27 18:41:06 crc kubenswrapper[4700]: I0227 18:41:06.982378 4700 scope.go:117] "RemoveContainer" containerID="c7ad608c50ddee2aac4992df7a6ccaaeda7c737e2a0d2fad5282b4482eac9e50" Feb 27 18:41:06 crc kubenswrapper[4700]: E0227 18:41:06.982938 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:41:12 crc kubenswrapper[4700]: I0227 18:41:12.789054 4700 generic.go:334] "Generic (PLEG): container finished" podID="f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670" containerID="9fc9801802539ab927f1ddb5a3637cf411a7dcefb8563790daf7e5b0847790fd" exitCode=0 Feb 27 18:41:12 crc kubenswrapper[4700]: I0227 18:41:12.789355 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5pqm8" event={"ID":"f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670","Type":"ContainerDied","Data":"9fc9801802539ab927f1ddb5a3637cf411a7dcefb8563790daf7e5b0847790fd"} Feb 27 18:41:13 crc kubenswrapper[4700]: I0227 18:41:13.806094 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5pqm8" event={"ID":"f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670","Type":"ContainerStarted","Data":"80dc0101c5ab739011c108f75c02c94192cd970c976c4eb4d52864e08ad11270"} Feb 27 18:41:13 crc kubenswrapper[4700]: I0227 18:41:13.843635 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5pqm8" podStartSLOduration=2.339067747 podStartE2EDuration="9.843611302s" podCreationTimestamp="2026-02-27 18:41:04 +0000 UTC" firstStartedPulling="2026-02-27 18:41:05.707802942 +0000 UTC m=+6025.693115689" lastFinishedPulling="2026-02-27 18:41:13.212346487 +0000 UTC m=+6033.197659244" observedRunningTime="2026-02-27 18:41:13.835757337 +0000 UTC m=+6033.821070124" watchObservedRunningTime="2026-02-27 18:41:13.843611302 +0000 UTC m=+6033.828924089" Feb 27 18:41:14 crc kubenswrapper[4700]: I0227 18:41:14.537709 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5pqm8" Feb 27 18:41:14 crc kubenswrapper[4700]: I0227 18:41:14.538412 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5pqm8" Feb 27 18:41:15 crc kubenswrapper[4700]: I0227 18:41:15.620766 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5pqm8" podUID="f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670" containerName="registry-server" probeResult="failure" output=< Feb 27 18:41:15 crc kubenswrapper[4700]: timeout: failed to connect service ":50051" within 1s Feb 27 18:41:15 crc kubenswrapper[4700]: > Feb 27 18:41:20 crc kubenswrapper[4700]: I0227 18:41:20.411949 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-znqck"] Feb 27 18:41:20 crc kubenswrapper[4700]: I0227 18:41:20.420206 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-znqck" Feb 27 18:41:20 crc kubenswrapper[4700]: I0227 18:41:20.432052 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-znqck"] Feb 27 18:41:20 crc kubenswrapper[4700]: I0227 18:41:20.598331 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52e1f8c9-d3ad-43bc-9e10-fbf842232ab7-utilities\") pod \"community-operators-znqck\" (UID: \"52e1f8c9-d3ad-43bc-9e10-fbf842232ab7\") " pod="openshift-marketplace/community-operators-znqck" Feb 27 18:41:20 crc kubenswrapper[4700]: I0227 18:41:20.598687 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52e1f8c9-d3ad-43bc-9e10-fbf842232ab7-catalog-content\") pod \"community-operators-znqck\" (UID: \"52e1f8c9-d3ad-43bc-9e10-fbf842232ab7\") " pod="openshift-marketplace/community-operators-znqck" Feb 27 18:41:20 crc kubenswrapper[4700]: I0227 18:41:20.598938 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v82tm\" (UniqueName: \"kubernetes.io/projected/52e1f8c9-d3ad-43bc-9e10-fbf842232ab7-kube-api-access-v82tm\") pod \"community-operators-znqck\" (UID: \"52e1f8c9-d3ad-43bc-9e10-fbf842232ab7\") " pod="openshift-marketplace/community-operators-znqck" Feb 27 18:41:20 crc kubenswrapper[4700]: I0227 18:41:20.701350 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52e1f8c9-d3ad-43bc-9e10-fbf842232ab7-utilities\") pod \"community-operators-znqck\" (UID: \"52e1f8c9-d3ad-43bc-9e10-fbf842232ab7\") " pod="openshift-marketplace/community-operators-znqck" Feb 27 18:41:20 crc kubenswrapper[4700]: I0227 18:41:20.701490 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52e1f8c9-d3ad-43bc-9e10-fbf842232ab7-catalog-content\") pod \"community-operators-znqck\" (UID: \"52e1f8c9-d3ad-43bc-9e10-fbf842232ab7\") " pod="openshift-marketplace/community-operators-znqck" Feb 27 18:41:20 crc kubenswrapper[4700]: I0227 18:41:20.701545 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v82tm\" (UniqueName: \"kubernetes.io/projected/52e1f8c9-d3ad-43bc-9e10-fbf842232ab7-kube-api-access-v82tm\") pod \"community-operators-znqck\" (UID: \"52e1f8c9-d3ad-43bc-9e10-fbf842232ab7\") " pod="openshift-marketplace/community-operators-znqck" Feb 27 18:41:20 crc kubenswrapper[4700]: I0227 18:41:20.701898 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52e1f8c9-d3ad-43bc-9e10-fbf842232ab7-utilities\") pod \"community-operators-znqck\" (UID: \"52e1f8c9-d3ad-43bc-9e10-fbf842232ab7\") " pod="openshift-marketplace/community-operators-znqck" Feb 27 18:41:20 crc kubenswrapper[4700]: I0227 18:41:20.702267 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52e1f8c9-d3ad-43bc-9e10-fbf842232ab7-catalog-content\") pod \"community-operators-znqck\" (UID: \"52e1f8c9-d3ad-43bc-9e10-fbf842232ab7\") " pod="openshift-marketplace/community-operators-znqck" Feb 27 18:41:20 crc kubenswrapper[4700]: I0227 18:41:20.718611 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v82tm\" (UniqueName: \"kubernetes.io/projected/52e1f8c9-d3ad-43bc-9e10-fbf842232ab7-kube-api-access-v82tm\") pod \"community-operators-znqck\" (UID: \"52e1f8c9-d3ad-43bc-9e10-fbf842232ab7\") " pod="openshift-marketplace/community-operators-znqck" Feb 27 18:41:20 crc kubenswrapper[4700]: I0227 18:41:20.796330 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-znqck" Feb 27 18:41:21 crc kubenswrapper[4700]: I0227 18:41:21.351691 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-znqck"] Feb 27 18:41:21 crc kubenswrapper[4700]: W0227 18:41:21.361747 4700 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod52e1f8c9_d3ad_43bc_9e10_fbf842232ab7.slice/crio-d018cc3f0b771b523b315f462baa16df3b0e823a816862622b6da6de640d89a7 WatchSource:0}: Error finding container d018cc3f0b771b523b315f462baa16df3b0e823a816862622b6da6de640d89a7: Status 404 returned error can't find the container with id d018cc3f0b771b523b315f462baa16df3b0e823a816862622b6da6de640d89a7 Feb 27 18:41:21 crc kubenswrapper[4700]: I0227 18:41:21.896366 4700 generic.go:334] "Generic (PLEG): container finished" podID="52e1f8c9-d3ad-43bc-9e10-fbf842232ab7" containerID="9f628253e56fad7a67c1983f093f20872cfe88e5386d2319bd48f063baf7906b" exitCode=0 Feb 27 18:41:21 crc kubenswrapper[4700]: I0227 18:41:21.896524 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-znqck" event={"ID":"52e1f8c9-d3ad-43bc-9e10-fbf842232ab7","Type":"ContainerDied","Data":"9f628253e56fad7a67c1983f093f20872cfe88e5386d2319bd48f063baf7906b"} Feb 27 18:41:21 crc kubenswrapper[4700]: I0227 18:41:21.896839 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-znqck" event={"ID":"52e1f8c9-d3ad-43bc-9e10-fbf842232ab7","Type":"ContainerStarted","Data":"d018cc3f0b771b523b315f462baa16df3b0e823a816862622b6da6de640d89a7"} Feb 27 18:41:21 crc kubenswrapper[4700]: I0227 18:41:21.981897 4700 scope.go:117] "RemoveContainer" containerID="c7ad608c50ddee2aac4992df7a6ccaaeda7c737e2a0d2fad5282b4482eac9e50" Feb 27 18:41:21 crc kubenswrapper[4700]: E0227 18:41:21.982204 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:41:22 crc kubenswrapper[4700]: I0227 18:41:22.911681 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-znqck" event={"ID":"52e1f8c9-d3ad-43bc-9e10-fbf842232ab7","Type":"ContainerStarted","Data":"9b9823ecfe1665c71565f4b1dbcadb4efb85881a733a4a36a8c2051c3efadc86"} Feb 27 18:41:24 crc kubenswrapper[4700]: I0227 18:41:24.936237 4700 generic.go:334] "Generic (PLEG): container finished" podID="52e1f8c9-d3ad-43bc-9e10-fbf842232ab7" containerID="9b9823ecfe1665c71565f4b1dbcadb4efb85881a733a4a36a8c2051c3efadc86" exitCode=0 Feb 27 18:41:24 crc kubenswrapper[4700]: I0227 18:41:24.936333 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-znqck" event={"ID":"52e1f8c9-d3ad-43bc-9e10-fbf842232ab7","Type":"ContainerDied","Data":"9b9823ecfe1665c71565f4b1dbcadb4efb85881a733a4a36a8c2051c3efadc86"} Feb 27 18:41:25 crc kubenswrapper[4700]: I0227 18:41:25.594996 4700 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5pqm8" podUID="f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670" containerName="registry-server" probeResult="failure" output=< Feb 27 18:41:25 crc kubenswrapper[4700]: timeout: failed to connect service ":50051" within 1s Feb 27 18:41:25 crc kubenswrapper[4700]: > Feb 27 18:41:25 crc kubenswrapper[4700]: I0227 18:41:25.949899 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-znqck" event={"ID":"52e1f8c9-d3ad-43bc-9e10-fbf842232ab7","Type":"ContainerStarted","Data":"b2b9051b2901974872ecda32a2f1a37b222b36d5fd108cc7d1ce622200a3bd3b"} Feb 27 18:41:25 crc kubenswrapper[4700]: I0227 18:41:25.976449 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-znqck" podStartSLOduration=2.4447792440000002 podStartE2EDuration="5.976424194s" podCreationTimestamp="2026-02-27 18:41:20 +0000 UTC" firstStartedPulling="2026-02-27 18:41:21.898852552 +0000 UTC m=+6041.884165309" lastFinishedPulling="2026-02-27 18:41:25.430497512 +0000 UTC m=+6045.415810259" observedRunningTime="2026-02-27 18:41:25.970252693 +0000 UTC m=+6045.955565460" watchObservedRunningTime="2026-02-27 18:41:25.976424194 +0000 UTC m=+6045.961736941" Feb 27 18:41:30 crc kubenswrapper[4700]: I0227 18:41:30.798050 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-znqck" Feb 27 18:41:30 crc kubenswrapper[4700]: I0227 18:41:30.798560 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-znqck" Feb 27 18:41:31 crc kubenswrapper[4700]: I0227 18:41:31.428964 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-znqck" Feb 27 18:41:31 crc kubenswrapper[4700]: I0227 18:41:31.498835 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-znqck" Feb 27 18:41:31 crc kubenswrapper[4700]: I0227 18:41:31.675201 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-znqck"] Feb 27 18:41:32 crc kubenswrapper[4700]: I0227 18:41:32.982100 4700 scope.go:117] "RemoveContainer" containerID="c7ad608c50ddee2aac4992df7a6ccaaeda7c737e2a0d2fad5282b4482eac9e50" Feb 27 18:41:32 crc kubenswrapper[4700]: E0227 18:41:32.984251 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:41:33 crc kubenswrapper[4700]: I0227 18:41:33.022684 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-znqck" podUID="52e1f8c9-d3ad-43bc-9e10-fbf842232ab7" containerName="registry-server" containerID="cri-o://b2b9051b2901974872ecda32a2f1a37b222b36d5fd108cc7d1ce622200a3bd3b" gracePeriod=2 Feb 27 18:41:33 crc kubenswrapper[4700]: I0227 18:41:33.559176 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-znqck" Feb 27 18:41:33 crc kubenswrapper[4700]: I0227 18:41:33.726299 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v82tm\" (UniqueName: \"kubernetes.io/projected/52e1f8c9-d3ad-43bc-9e10-fbf842232ab7-kube-api-access-v82tm\") pod \"52e1f8c9-d3ad-43bc-9e10-fbf842232ab7\" (UID: \"52e1f8c9-d3ad-43bc-9e10-fbf842232ab7\") " Feb 27 18:41:33 crc kubenswrapper[4700]: I0227 18:41:33.726422 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52e1f8c9-d3ad-43bc-9e10-fbf842232ab7-utilities\") pod \"52e1f8c9-d3ad-43bc-9e10-fbf842232ab7\" (UID: \"52e1f8c9-d3ad-43bc-9e10-fbf842232ab7\") " Feb 27 18:41:33 crc kubenswrapper[4700]: I0227 18:41:33.726641 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52e1f8c9-d3ad-43bc-9e10-fbf842232ab7-catalog-content\") pod \"52e1f8c9-d3ad-43bc-9e10-fbf842232ab7\" (UID: \"52e1f8c9-d3ad-43bc-9e10-fbf842232ab7\") " Feb 27 18:41:33 crc kubenswrapper[4700]: I0227 18:41:33.727189 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52e1f8c9-d3ad-43bc-9e10-fbf842232ab7-utilities" (OuterVolumeSpecName: "utilities") pod "52e1f8c9-d3ad-43bc-9e10-fbf842232ab7" (UID: "52e1f8c9-d3ad-43bc-9e10-fbf842232ab7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 18:41:33 crc kubenswrapper[4700]: I0227 18:41:33.727740 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52e1f8c9-d3ad-43bc-9e10-fbf842232ab7-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 18:41:33 crc kubenswrapper[4700]: I0227 18:41:33.746022 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52e1f8c9-d3ad-43bc-9e10-fbf842232ab7-kube-api-access-v82tm" (OuterVolumeSpecName: "kube-api-access-v82tm") pod "52e1f8c9-d3ad-43bc-9e10-fbf842232ab7" (UID: "52e1f8c9-d3ad-43bc-9e10-fbf842232ab7"). InnerVolumeSpecName "kube-api-access-v82tm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:41:33 crc kubenswrapper[4700]: I0227 18:41:33.807445 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52e1f8c9-d3ad-43bc-9e10-fbf842232ab7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "52e1f8c9-d3ad-43bc-9e10-fbf842232ab7" (UID: "52e1f8c9-d3ad-43bc-9e10-fbf842232ab7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 18:41:33 crc kubenswrapper[4700]: I0227 18:41:33.829600 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v82tm\" (UniqueName: \"kubernetes.io/projected/52e1f8c9-d3ad-43bc-9e10-fbf842232ab7-kube-api-access-v82tm\") on node \"crc\" DevicePath \"\"" Feb 27 18:41:33 crc kubenswrapper[4700]: I0227 18:41:33.829640 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52e1f8c9-d3ad-43bc-9e10-fbf842232ab7-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 18:41:34 crc kubenswrapper[4700]: I0227 18:41:34.033096 4700 generic.go:334] "Generic (PLEG): container finished" podID="52e1f8c9-d3ad-43bc-9e10-fbf842232ab7" containerID="b2b9051b2901974872ecda32a2f1a37b222b36d5fd108cc7d1ce622200a3bd3b" exitCode=0 Feb 27 18:41:34 crc kubenswrapper[4700]: I0227 18:41:34.033136 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-znqck" event={"ID":"52e1f8c9-d3ad-43bc-9e10-fbf842232ab7","Type":"ContainerDied","Data":"b2b9051b2901974872ecda32a2f1a37b222b36d5fd108cc7d1ce622200a3bd3b"} Feb 27 18:41:34 crc kubenswrapper[4700]: I0227 18:41:34.033160 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-znqck" event={"ID":"52e1f8c9-d3ad-43bc-9e10-fbf842232ab7","Type":"ContainerDied","Data":"d018cc3f0b771b523b315f462baa16df3b0e823a816862622b6da6de640d89a7"} Feb 27 18:41:34 crc kubenswrapper[4700]: I0227 18:41:34.033176 4700 scope.go:117] "RemoveContainer" containerID="b2b9051b2901974872ecda32a2f1a37b222b36d5fd108cc7d1ce622200a3bd3b" Feb 27 18:41:34 crc kubenswrapper[4700]: I0227 18:41:34.033277 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-znqck" Feb 27 18:41:34 crc kubenswrapper[4700]: I0227 18:41:34.060419 4700 scope.go:117] "RemoveContainer" containerID="9b9823ecfe1665c71565f4b1dbcadb4efb85881a733a4a36a8c2051c3efadc86" Feb 27 18:41:34 crc kubenswrapper[4700]: I0227 18:41:34.081553 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-znqck"] Feb 27 18:41:34 crc kubenswrapper[4700]: I0227 18:41:34.095799 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-znqck"] Feb 27 18:41:34 crc kubenswrapper[4700]: I0227 18:41:34.139015 4700 scope.go:117] "RemoveContainer" containerID="9f628253e56fad7a67c1983f093f20872cfe88e5386d2319bd48f063baf7906b" Feb 27 18:41:34 crc kubenswrapper[4700]: I0227 18:41:34.165883 4700 scope.go:117] "RemoveContainer" containerID="b2b9051b2901974872ecda32a2f1a37b222b36d5fd108cc7d1ce622200a3bd3b" Feb 27 18:41:34 crc kubenswrapper[4700]: E0227 18:41:34.166346 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2b9051b2901974872ecda32a2f1a37b222b36d5fd108cc7d1ce622200a3bd3b\": container with ID starting with b2b9051b2901974872ecda32a2f1a37b222b36d5fd108cc7d1ce622200a3bd3b not found: ID does not exist" containerID="b2b9051b2901974872ecda32a2f1a37b222b36d5fd108cc7d1ce622200a3bd3b" Feb 27 18:41:34 crc kubenswrapper[4700]: I0227 18:41:34.166474 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2b9051b2901974872ecda32a2f1a37b222b36d5fd108cc7d1ce622200a3bd3b"} err="failed to get container status \"b2b9051b2901974872ecda32a2f1a37b222b36d5fd108cc7d1ce622200a3bd3b\": rpc error: code = NotFound desc = could not find container \"b2b9051b2901974872ecda32a2f1a37b222b36d5fd108cc7d1ce622200a3bd3b\": container with ID starting with b2b9051b2901974872ecda32a2f1a37b222b36d5fd108cc7d1ce622200a3bd3b not found: ID does not exist" Feb 27 18:41:34 crc kubenswrapper[4700]: I0227 18:41:34.166549 4700 scope.go:117] "RemoveContainer" containerID="9b9823ecfe1665c71565f4b1dbcadb4efb85881a733a4a36a8c2051c3efadc86" Feb 27 18:41:34 crc kubenswrapper[4700]: E0227 18:41:34.170544 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b9823ecfe1665c71565f4b1dbcadb4efb85881a733a4a36a8c2051c3efadc86\": container with ID starting with 9b9823ecfe1665c71565f4b1dbcadb4efb85881a733a4a36a8c2051c3efadc86 not found: ID does not exist" containerID="9b9823ecfe1665c71565f4b1dbcadb4efb85881a733a4a36a8c2051c3efadc86" Feb 27 18:41:34 crc kubenswrapper[4700]: I0227 18:41:34.170678 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b9823ecfe1665c71565f4b1dbcadb4efb85881a733a4a36a8c2051c3efadc86"} err="failed to get container status \"9b9823ecfe1665c71565f4b1dbcadb4efb85881a733a4a36a8c2051c3efadc86\": rpc error: code = NotFound desc = could not find container \"9b9823ecfe1665c71565f4b1dbcadb4efb85881a733a4a36a8c2051c3efadc86\": container with ID starting with 9b9823ecfe1665c71565f4b1dbcadb4efb85881a733a4a36a8c2051c3efadc86 not found: ID does not exist" Feb 27 18:41:34 crc kubenswrapper[4700]: I0227 18:41:34.170749 4700 scope.go:117] "RemoveContainer" containerID="9f628253e56fad7a67c1983f093f20872cfe88e5386d2319bd48f063baf7906b" Feb 27 18:41:34 crc kubenswrapper[4700]: E0227 18:41:34.171121 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f628253e56fad7a67c1983f093f20872cfe88e5386d2319bd48f063baf7906b\": container with ID starting with 9f628253e56fad7a67c1983f093f20872cfe88e5386d2319bd48f063baf7906b not found: ID does not exist" containerID="9f628253e56fad7a67c1983f093f20872cfe88e5386d2319bd48f063baf7906b" Feb 27 18:41:34 crc kubenswrapper[4700]: I0227 18:41:34.171205 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f628253e56fad7a67c1983f093f20872cfe88e5386d2319bd48f063baf7906b"} err="failed to get container status \"9f628253e56fad7a67c1983f093f20872cfe88e5386d2319bd48f063baf7906b\": rpc error: code = NotFound desc = could not find container \"9f628253e56fad7a67c1983f093f20872cfe88e5386d2319bd48f063baf7906b\": container with ID starting with 9f628253e56fad7a67c1983f093f20872cfe88e5386d2319bd48f063baf7906b not found: ID does not exist" Feb 27 18:41:34 crc kubenswrapper[4700]: I0227 18:41:34.632101 4700 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5pqm8" Feb 27 18:41:34 crc kubenswrapper[4700]: I0227 18:41:34.723543 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5pqm8" Feb 27 18:41:35 crc kubenswrapper[4700]: I0227 18:41:35.000604 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52e1f8c9-d3ad-43bc-9e10-fbf842232ab7" path="/var/lib/kubelet/pods/52e1f8c9-d3ad-43bc-9e10-fbf842232ab7/volumes" Feb 27 18:41:36 crc kubenswrapper[4700]: I0227 18:41:36.072136 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5pqm8"] Feb 27 18:41:36 crc kubenswrapper[4700]: I0227 18:41:36.072814 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5pqm8" podUID="f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670" containerName="registry-server" containerID="cri-o://80dc0101c5ab739011c108f75c02c94192cd970c976c4eb4d52864e08ad11270" gracePeriod=2 Feb 27 18:41:36 crc kubenswrapper[4700]: I0227 18:41:36.586045 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5pqm8" Feb 27 18:41:36 crc kubenswrapper[4700]: I0227 18:41:36.702189 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670-utilities\") pod \"f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670\" (UID: \"f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670\") " Feb 27 18:41:36 crc kubenswrapper[4700]: I0227 18:41:36.702256 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckcx7\" (UniqueName: \"kubernetes.io/projected/f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670-kube-api-access-ckcx7\") pod \"f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670\" (UID: \"f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670\") " Feb 27 18:41:36 crc kubenswrapper[4700]: I0227 18:41:36.702396 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670-catalog-content\") pod \"f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670\" (UID: \"f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670\") " Feb 27 18:41:36 crc kubenswrapper[4700]: I0227 18:41:36.702971 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670-utilities" (OuterVolumeSpecName: "utilities") pod "f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670" (UID: "f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 18:41:36 crc kubenswrapper[4700]: I0227 18:41:36.721685 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670-kube-api-access-ckcx7" (OuterVolumeSpecName: "kube-api-access-ckcx7") pod "f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670" (UID: "f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670"). InnerVolumeSpecName "kube-api-access-ckcx7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:41:36 crc kubenswrapper[4700]: I0227 18:41:36.804588 4700 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 18:41:36 crc kubenswrapper[4700]: I0227 18:41:36.804889 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckcx7\" (UniqueName: \"kubernetes.io/projected/f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670-kube-api-access-ckcx7\") on node \"crc\" DevicePath \"\"" Feb 27 18:41:36 crc kubenswrapper[4700]: I0227 18:41:36.834834 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670" (UID: "f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 18:41:36 crc kubenswrapper[4700]: I0227 18:41:36.907446 4700 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 18:41:37 crc kubenswrapper[4700]: I0227 18:41:37.069004 4700 generic.go:334] "Generic (PLEG): container finished" podID="f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670" containerID="80dc0101c5ab739011c108f75c02c94192cd970c976c4eb4d52864e08ad11270" exitCode=0 Feb 27 18:41:37 crc kubenswrapper[4700]: I0227 18:41:37.069067 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5pqm8" event={"ID":"f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670","Type":"ContainerDied","Data":"80dc0101c5ab739011c108f75c02c94192cd970c976c4eb4d52864e08ad11270"} Feb 27 18:41:37 crc kubenswrapper[4700]: I0227 18:41:37.069117 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5pqm8" event={"ID":"f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670","Type":"ContainerDied","Data":"a8912e2733f2ca58161c8fa264b6417554ad67333069c79b48d1b48fd4860201"} Feb 27 18:41:37 crc kubenswrapper[4700]: I0227 18:41:37.069147 4700 scope.go:117] "RemoveContainer" containerID="80dc0101c5ab739011c108f75c02c94192cd970c976c4eb4d52864e08ad11270" Feb 27 18:41:37 crc kubenswrapper[4700]: I0227 18:41:37.069158 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5pqm8" Feb 27 18:41:37 crc kubenswrapper[4700]: I0227 18:41:37.104836 4700 scope.go:117] "RemoveContainer" containerID="9fc9801802539ab927f1ddb5a3637cf411a7dcefb8563790daf7e5b0847790fd" Feb 27 18:41:37 crc kubenswrapper[4700]: I0227 18:41:37.127185 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5pqm8"] Feb 27 18:41:37 crc kubenswrapper[4700]: I0227 18:41:37.139801 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5pqm8"] Feb 27 18:41:37 crc kubenswrapper[4700]: I0227 18:41:37.154893 4700 scope.go:117] "RemoveContainer" containerID="f5a62cf8d3a0b543c40f58460495274c3f16d8a3a47dc3c887ea09d40d6ea36c" Feb 27 18:41:37 crc kubenswrapper[4700]: I0227 18:41:37.189513 4700 scope.go:117] "RemoveContainer" containerID="80dc0101c5ab739011c108f75c02c94192cd970c976c4eb4d52864e08ad11270" Feb 27 18:41:37 crc kubenswrapper[4700]: E0227 18:41:37.190179 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80dc0101c5ab739011c108f75c02c94192cd970c976c4eb4d52864e08ad11270\": container with ID starting with 80dc0101c5ab739011c108f75c02c94192cd970c976c4eb4d52864e08ad11270 not found: ID does not exist" containerID="80dc0101c5ab739011c108f75c02c94192cd970c976c4eb4d52864e08ad11270" Feb 27 18:41:37 crc kubenswrapper[4700]: I0227 18:41:37.190284 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80dc0101c5ab739011c108f75c02c94192cd970c976c4eb4d52864e08ad11270"} err="failed to get container status \"80dc0101c5ab739011c108f75c02c94192cd970c976c4eb4d52864e08ad11270\": rpc error: code = NotFound desc = could not find container \"80dc0101c5ab739011c108f75c02c94192cd970c976c4eb4d52864e08ad11270\": container with ID starting with 80dc0101c5ab739011c108f75c02c94192cd970c976c4eb4d52864e08ad11270 not found: ID does not exist" Feb 27 18:41:37 crc kubenswrapper[4700]: I0227 18:41:37.190377 4700 scope.go:117] "RemoveContainer" containerID="9fc9801802539ab927f1ddb5a3637cf411a7dcefb8563790daf7e5b0847790fd" Feb 27 18:41:37 crc kubenswrapper[4700]: E0227 18:41:37.190784 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fc9801802539ab927f1ddb5a3637cf411a7dcefb8563790daf7e5b0847790fd\": container with ID starting with 9fc9801802539ab927f1ddb5a3637cf411a7dcefb8563790daf7e5b0847790fd not found: ID does not exist" containerID="9fc9801802539ab927f1ddb5a3637cf411a7dcefb8563790daf7e5b0847790fd" Feb 27 18:41:37 crc kubenswrapper[4700]: I0227 18:41:37.190807 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fc9801802539ab927f1ddb5a3637cf411a7dcefb8563790daf7e5b0847790fd"} err="failed to get container status \"9fc9801802539ab927f1ddb5a3637cf411a7dcefb8563790daf7e5b0847790fd\": rpc error: code = NotFound desc = could not find container \"9fc9801802539ab927f1ddb5a3637cf411a7dcefb8563790daf7e5b0847790fd\": container with ID starting with 9fc9801802539ab927f1ddb5a3637cf411a7dcefb8563790daf7e5b0847790fd not found: ID does not exist" Feb 27 18:41:37 crc kubenswrapper[4700]: I0227 18:41:37.190822 4700 scope.go:117] "RemoveContainer" containerID="f5a62cf8d3a0b543c40f58460495274c3f16d8a3a47dc3c887ea09d40d6ea36c" Feb 27 18:41:37 crc kubenswrapper[4700]: E0227 18:41:37.191172 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5a62cf8d3a0b543c40f58460495274c3f16d8a3a47dc3c887ea09d40d6ea36c\": container with ID starting with f5a62cf8d3a0b543c40f58460495274c3f16d8a3a47dc3c887ea09d40d6ea36c not found: ID does not exist" containerID="f5a62cf8d3a0b543c40f58460495274c3f16d8a3a47dc3c887ea09d40d6ea36c" Feb 27 18:41:37 crc kubenswrapper[4700]: I0227 18:41:37.191256 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5a62cf8d3a0b543c40f58460495274c3f16d8a3a47dc3c887ea09d40d6ea36c"} err="failed to get container status \"f5a62cf8d3a0b543c40f58460495274c3f16d8a3a47dc3c887ea09d40d6ea36c\": rpc error: code = NotFound desc = could not find container \"f5a62cf8d3a0b543c40f58460495274c3f16d8a3a47dc3c887ea09d40d6ea36c\": container with ID starting with f5a62cf8d3a0b543c40f58460495274c3f16d8a3a47dc3c887ea09d40d6ea36c not found: ID does not exist" Feb 27 18:41:38 crc kubenswrapper[4700]: I0227 18:41:38.991130 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670" path="/var/lib/kubelet/pods/f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670/volumes" Feb 27 18:41:39 crc kubenswrapper[4700]: I0227 18:41:39.101809 4700 generic.go:334] "Generic (PLEG): container finished" podID="489734cd-8c02-4e0b-b300-62a2ce5b6c94" containerID="70b8519010dd2402205e46f8823f4e36e179cf07706d5a2a05faa1a6cbaa6cc7" exitCode=0 Feb 27 18:41:39 crc kubenswrapper[4700]: I0227 18:41:39.101851 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wqxph/must-gather-6pk65" event={"ID":"489734cd-8c02-4e0b-b300-62a2ce5b6c94","Type":"ContainerDied","Data":"70b8519010dd2402205e46f8823f4e36e179cf07706d5a2a05faa1a6cbaa6cc7"} Feb 27 18:41:39 crc kubenswrapper[4700]: I0227 18:41:39.102550 4700 scope.go:117] "RemoveContainer" containerID="70b8519010dd2402205e46f8823f4e36e179cf07706d5a2a05faa1a6cbaa6cc7" Feb 27 18:41:40 crc kubenswrapper[4700]: I0227 18:41:40.093396 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wqxph_must-gather-6pk65_489734cd-8c02-4e0b-b300-62a2ce5b6c94/gather/0.log" Feb 27 18:41:43 crc kubenswrapper[4700]: I0227 18:41:43.988147 4700 scope.go:117] "RemoveContainer" containerID="c7ad608c50ddee2aac4992df7a6ccaaeda7c737e2a0d2fad5282b4482eac9e50" Feb 27 18:41:43 crc kubenswrapper[4700]: E0227 18:41:43.989287 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:41:49 crc kubenswrapper[4700]: I0227 18:41:49.219826 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wqxph/must-gather-6pk65"] Feb 27 18:41:49 crc kubenswrapper[4700]: I0227 18:41:49.220669 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-wqxph/must-gather-6pk65" podUID="489734cd-8c02-4e0b-b300-62a2ce5b6c94" containerName="copy" containerID="cri-o://062b668618606ea03fd04a44e970d2a53effed5ec68456e58361756b89dd4e55" gracePeriod=2 Feb 27 18:41:49 crc kubenswrapper[4700]: I0227 18:41:49.234771 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wqxph/must-gather-6pk65"] Feb 27 18:41:49 crc kubenswrapper[4700]: I0227 18:41:49.680413 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wqxph_must-gather-6pk65_489734cd-8c02-4e0b-b300-62a2ce5b6c94/copy/0.log" Feb 27 18:41:49 crc kubenswrapper[4700]: I0227 18:41:49.681168 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wqxph/must-gather-6pk65" Feb 27 18:41:49 crc kubenswrapper[4700]: I0227 18:41:49.827150 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8w67\" (UniqueName: \"kubernetes.io/projected/489734cd-8c02-4e0b-b300-62a2ce5b6c94-kube-api-access-c8w67\") pod \"489734cd-8c02-4e0b-b300-62a2ce5b6c94\" (UID: \"489734cd-8c02-4e0b-b300-62a2ce5b6c94\") " Feb 27 18:41:49 crc kubenswrapper[4700]: I0227 18:41:49.827210 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/489734cd-8c02-4e0b-b300-62a2ce5b6c94-must-gather-output\") pod \"489734cd-8c02-4e0b-b300-62a2ce5b6c94\" (UID: \"489734cd-8c02-4e0b-b300-62a2ce5b6c94\") " Feb 27 18:41:49 crc kubenswrapper[4700]: I0227 18:41:49.835629 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/489734cd-8c02-4e0b-b300-62a2ce5b6c94-kube-api-access-c8w67" (OuterVolumeSpecName: "kube-api-access-c8w67") pod "489734cd-8c02-4e0b-b300-62a2ce5b6c94" (UID: "489734cd-8c02-4e0b-b300-62a2ce5b6c94"). InnerVolumeSpecName "kube-api-access-c8w67". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:41:49 crc kubenswrapper[4700]: I0227 18:41:49.930135 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8w67\" (UniqueName: \"kubernetes.io/projected/489734cd-8c02-4e0b-b300-62a2ce5b6c94-kube-api-access-c8w67\") on node \"crc\" DevicePath \"\"" Feb 27 18:41:50 crc kubenswrapper[4700]: I0227 18:41:50.036015 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/489734cd-8c02-4e0b-b300-62a2ce5b6c94-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "489734cd-8c02-4e0b-b300-62a2ce5b6c94" (UID: "489734cd-8c02-4e0b-b300-62a2ce5b6c94"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 18:41:50 crc kubenswrapper[4700]: I0227 18:41:50.134304 4700 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/489734cd-8c02-4e0b-b300-62a2ce5b6c94-must-gather-output\") on node \"crc\" DevicePath \"\"" Feb 27 18:41:50 crc kubenswrapper[4700]: I0227 18:41:50.249199 4700 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wqxph_must-gather-6pk65_489734cd-8c02-4e0b-b300-62a2ce5b6c94/copy/0.log" Feb 27 18:41:50 crc kubenswrapper[4700]: I0227 18:41:50.249794 4700 generic.go:334] "Generic (PLEG): container finished" podID="489734cd-8c02-4e0b-b300-62a2ce5b6c94" containerID="062b668618606ea03fd04a44e970d2a53effed5ec68456e58361756b89dd4e55" exitCode=143 Feb 27 18:41:50 crc kubenswrapper[4700]: I0227 18:41:50.249854 4700 scope.go:117] "RemoveContainer" containerID="062b668618606ea03fd04a44e970d2a53effed5ec68456e58361756b89dd4e55" Feb 27 18:41:50 crc kubenswrapper[4700]: I0227 18:41:50.249861 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wqxph/must-gather-6pk65" Feb 27 18:41:50 crc kubenswrapper[4700]: I0227 18:41:50.286519 4700 scope.go:117] "RemoveContainer" containerID="70b8519010dd2402205e46f8823f4e36e179cf07706d5a2a05faa1a6cbaa6cc7" Feb 27 18:41:50 crc kubenswrapper[4700]: I0227 18:41:50.424364 4700 scope.go:117] "RemoveContainer" containerID="062b668618606ea03fd04a44e970d2a53effed5ec68456e58361756b89dd4e55" Feb 27 18:41:50 crc kubenswrapper[4700]: E0227 18:41:50.424879 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"062b668618606ea03fd04a44e970d2a53effed5ec68456e58361756b89dd4e55\": container with ID starting with 062b668618606ea03fd04a44e970d2a53effed5ec68456e58361756b89dd4e55 not found: ID does not exist" containerID="062b668618606ea03fd04a44e970d2a53effed5ec68456e58361756b89dd4e55" Feb 27 18:41:50 crc kubenswrapper[4700]: I0227 18:41:50.424941 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"062b668618606ea03fd04a44e970d2a53effed5ec68456e58361756b89dd4e55"} err="failed to get container status \"062b668618606ea03fd04a44e970d2a53effed5ec68456e58361756b89dd4e55\": rpc error: code = NotFound desc = could not find container \"062b668618606ea03fd04a44e970d2a53effed5ec68456e58361756b89dd4e55\": container with ID starting with 062b668618606ea03fd04a44e970d2a53effed5ec68456e58361756b89dd4e55 not found: ID does not exist" Feb 27 18:41:50 crc kubenswrapper[4700]: I0227 18:41:50.424981 4700 scope.go:117] "RemoveContainer" containerID="70b8519010dd2402205e46f8823f4e36e179cf07706d5a2a05faa1a6cbaa6cc7" Feb 27 18:41:50 crc kubenswrapper[4700]: E0227 18:41:50.425348 4700 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70b8519010dd2402205e46f8823f4e36e179cf07706d5a2a05faa1a6cbaa6cc7\": container with ID starting with 70b8519010dd2402205e46f8823f4e36e179cf07706d5a2a05faa1a6cbaa6cc7 not found: ID does not exist" containerID="70b8519010dd2402205e46f8823f4e36e179cf07706d5a2a05faa1a6cbaa6cc7" Feb 27 18:41:50 crc kubenswrapper[4700]: I0227 18:41:50.425381 4700 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70b8519010dd2402205e46f8823f4e36e179cf07706d5a2a05faa1a6cbaa6cc7"} err="failed to get container status \"70b8519010dd2402205e46f8823f4e36e179cf07706d5a2a05faa1a6cbaa6cc7\": rpc error: code = NotFound desc = could not find container \"70b8519010dd2402205e46f8823f4e36e179cf07706d5a2a05faa1a6cbaa6cc7\": container with ID starting with 70b8519010dd2402205e46f8823f4e36e179cf07706d5a2a05faa1a6cbaa6cc7 not found: ID does not exist" Feb 27 18:41:50 crc kubenswrapper[4700]: I0227 18:41:50.993264 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="489734cd-8c02-4e0b-b300-62a2ce5b6c94" path="/var/lib/kubelet/pods/489734cd-8c02-4e0b-b300-62a2ce5b6c94/volumes" Feb 27 18:41:57 crc kubenswrapper[4700]: I0227 18:41:57.982159 4700 scope.go:117] "RemoveContainer" containerID="c7ad608c50ddee2aac4992df7a6ccaaeda7c737e2a0d2fad5282b4482eac9e50" Feb 27 18:41:57 crc kubenswrapper[4700]: E0227 18:41:57.982836 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:42:00 crc kubenswrapper[4700]: I0227 18:42:00.158788 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536962-mmx6b"] Feb 27 18:42:00 crc kubenswrapper[4700]: E0227 18:42:00.160047 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52e1f8c9-d3ad-43bc-9e10-fbf842232ab7" containerName="registry-server" Feb 27 18:42:00 crc kubenswrapper[4700]: I0227 18:42:00.160069 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="52e1f8c9-d3ad-43bc-9e10-fbf842232ab7" containerName="registry-server" Feb 27 18:42:00 crc kubenswrapper[4700]: E0227 18:42:00.160110 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670" containerName="extract-utilities" Feb 27 18:42:00 crc kubenswrapper[4700]: I0227 18:42:00.160123 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670" containerName="extract-utilities" Feb 27 18:42:00 crc kubenswrapper[4700]: E0227 18:42:00.160144 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52e1f8c9-d3ad-43bc-9e10-fbf842232ab7" containerName="extract-utilities" Feb 27 18:42:00 crc kubenswrapper[4700]: I0227 18:42:00.160158 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="52e1f8c9-d3ad-43bc-9e10-fbf842232ab7" containerName="extract-utilities" Feb 27 18:42:00 crc kubenswrapper[4700]: E0227 18:42:00.160173 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="489734cd-8c02-4e0b-b300-62a2ce5b6c94" containerName="gather" Feb 27 18:42:00 crc kubenswrapper[4700]: I0227 18:42:00.160184 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="489734cd-8c02-4e0b-b300-62a2ce5b6c94" containerName="gather" Feb 27 18:42:00 crc kubenswrapper[4700]: E0227 18:42:00.160250 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670" containerName="registry-server" Feb 27 18:42:00 crc kubenswrapper[4700]: I0227 18:42:00.160262 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670" containerName="registry-server" Feb 27 18:42:00 crc kubenswrapper[4700]: E0227 18:42:00.160283 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="489734cd-8c02-4e0b-b300-62a2ce5b6c94" containerName="copy" Feb 27 18:42:00 crc kubenswrapper[4700]: I0227 18:42:00.160294 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="489734cd-8c02-4e0b-b300-62a2ce5b6c94" containerName="copy" Feb 27 18:42:00 crc kubenswrapper[4700]: E0227 18:42:00.160318 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670" containerName="extract-content" Feb 27 18:42:00 crc kubenswrapper[4700]: I0227 18:42:00.160330 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670" containerName="extract-content" Feb 27 18:42:00 crc kubenswrapper[4700]: E0227 18:42:00.160359 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52e1f8c9-d3ad-43bc-9e10-fbf842232ab7" containerName="extract-content" Feb 27 18:42:00 crc kubenswrapper[4700]: I0227 18:42:00.160371 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="52e1f8c9-d3ad-43bc-9e10-fbf842232ab7" containerName="extract-content" Feb 27 18:42:00 crc kubenswrapper[4700]: I0227 18:42:00.160745 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="f02bc8f0-7e31-4a40-96b7-0bf4b4c4f670" containerName="registry-server" Feb 27 18:42:00 crc kubenswrapper[4700]: I0227 18:42:00.160790 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="489734cd-8c02-4e0b-b300-62a2ce5b6c94" containerName="copy" Feb 27 18:42:00 crc kubenswrapper[4700]: I0227 18:42:00.160816 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="489734cd-8c02-4e0b-b300-62a2ce5b6c94" containerName="gather" Feb 27 18:42:00 crc kubenswrapper[4700]: I0227 18:42:00.160834 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="52e1f8c9-d3ad-43bc-9e10-fbf842232ab7" containerName="registry-server" Feb 27 18:42:00 crc kubenswrapper[4700]: I0227 18:42:00.161992 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536962-mmx6b" Feb 27 18:42:00 crc kubenswrapper[4700]: I0227 18:42:00.164562 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d62kc" Feb 27 18:42:00 crc kubenswrapper[4700]: I0227 18:42:00.165586 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 18:42:00 crc kubenswrapper[4700]: I0227 18:42:00.166568 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 18:42:00 crc kubenswrapper[4700]: I0227 18:42:00.171333 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536962-mmx6b"] Feb 27 18:42:00 crc kubenswrapper[4700]: I0227 18:42:00.306016 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p24nw\" (UniqueName: \"kubernetes.io/projected/5ee20344-c058-4d9c-93a4-fe301ccf6b9e-kube-api-access-p24nw\") pod \"auto-csr-approver-29536962-mmx6b\" (UID: \"5ee20344-c058-4d9c-93a4-fe301ccf6b9e\") " pod="openshift-infra/auto-csr-approver-29536962-mmx6b" Feb 27 18:42:00 crc kubenswrapper[4700]: I0227 18:42:00.411395 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p24nw\" (UniqueName: \"kubernetes.io/projected/5ee20344-c058-4d9c-93a4-fe301ccf6b9e-kube-api-access-p24nw\") pod \"auto-csr-approver-29536962-mmx6b\" (UID: \"5ee20344-c058-4d9c-93a4-fe301ccf6b9e\") " pod="openshift-infra/auto-csr-approver-29536962-mmx6b" Feb 27 18:42:00 crc kubenswrapper[4700]: I0227 18:42:00.438237 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p24nw\" (UniqueName: \"kubernetes.io/projected/5ee20344-c058-4d9c-93a4-fe301ccf6b9e-kube-api-access-p24nw\") pod \"auto-csr-approver-29536962-mmx6b\" (UID: \"5ee20344-c058-4d9c-93a4-fe301ccf6b9e\") " pod="openshift-infra/auto-csr-approver-29536962-mmx6b" Feb 27 18:42:00 crc kubenswrapper[4700]: I0227 18:42:00.488763 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536962-mmx6b" Feb 27 18:42:01 crc kubenswrapper[4700]: I0227 18:42:01.031526 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536962-mmx6b"] Feb 27 18:42:01 crc kubenswrapper[4700]: I0227 18:42:01.390506 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536962-mmx6b" event={"ID":"5ee20344-c058-4d9c-93a4-fe301ccf6b9e","Type":"ContainerStarted","Data":"64c6440fc3f2e37827b0ce2612af343c59861d29607dd8b9c91eb9a338db1cbc"} Feb 27 18:42:03 crc kubenswrapper[4700]: I0227 18:42:03.422149 4700 generic.go:334] "Generic (PLEG): container finished" podID="5ee20344-c058-4d9c-93a4-fe301ccf6b9e" containerID="09a06199d218c894f92b55c374357fbf4b0776b97553fb49294b0393e910d534" exitCode=0 Feb 27 18:42:03 crc kubenswrapper[4700]: I0227 18:42:03.422248 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536962-mmx6b" event={"ID":"5ee20344-c058-4d9c-93a4-fe301ccf6b9e","Type":"ContainerDied","Data":"09a06199d218c894f92b55c374357fbf4b0776b97553fb49294b0393e910d534"} Feb 27 18:42:04 crc kubenswrapper[4700]: I0227 18:42:04.840527 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536962-mmx6b" Feb 27 18:42:04 crc kubenswrapper[4700]: I0227 18:42:04.919822 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p24nw\" (UniqueName: \"kubernetes.io/projected/5ee20344-c058-4d9c-93a4-fe301ccf6b9e-kube-api-access-p24nw\") pod \"5ee20344-c058-4d9c-93a4-fe301ccf6b9e\" (UID: \"5ee20344-c058-4d9c-93a4-fe301ccf6b9e\") " Feb 27 18:42:04 crc kubenswrapper[4700]: I0227 18:42:04.930863 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ee20344-c058-4d9c-93a4-fe301ccf6b9e-kube-api-access-p24nw" (OuterVolumeSpecName: "kube-api-access-p24nw") pod "5ee20344-c058-4d9c-93a4-fe301ccf6b9e" (UID: "5ee20344-c058-4d9c-93a4-fe301ccf6b9e"). InnerVolumeSpecName "kube-api-access-p24nw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:42:05 crc kubenswrapper[4700]: I0227 18:42:05.023619 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p24nw\" (UniqueName: \"kubernetes.io/projected/5ee20344-c058-4d9c-93a4-fe301ccf6b9e-kube-api-access-p24nw\") on node \"crc\" DevicePath \"\"" Feb 27 18:42:05 crc kubenswrapper[4700]: I0227 18:42:05.440914 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536962-mmx6b" event={"ID":"5ee20344-c058-4d9c-93a4-fe301ccf6b9e","Type":"ContainerDied","Data":"64c6440fc3f2e37827b0ce2612af343c59861d29607dd8b9c91eb9a338db1cbc"} Feb 27 18:42:05 crc kubenswrapper[4700]: I0227 18:42:05.440961 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64c6440fc3f2e37827b0ce2612af343c59861d29607dd8b9c91eb9a338db1cbc" Feb 27 18:42:05 crc kubenswrapper[4700]: I0227 18:42:05.441038 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536962-mmx6b" Feb 27 18:42:05 crc kubenswrapper[4700]: I0227 18:42:05.941089 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536956-sz7xx"] Feb 27 18:42:05 crc kubenswrapper[4700]: I0227 18:42:05.953768 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536956-sz7xx"] Feb 27 18:42:07 crc kubenswrapper[4700]: I0227 18:42:07.001326 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="074ce5d9-7bd8-4205-9c43-24c9cf22ea32" path="/var/lib/kubelet/pods/074ce5d9-7bd8-4205-9c43-24c9cf22ea32/volumes" Feb 27 18:42:12 crc kubenswrapper[4700]: I0227 18:42:12.981228 4700 scope.go:117] "RemoveContainer" containerID="c7ad608c50ddee2aac4992df7a6ccaaeda7c737e2a0d2fad5282b4482eac9e50" Feb 27 18:42:12 crc kubenswrapper[4700]: E0227 18:42:12.982180 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:42:14 crc kubenswrapper[4700]: I0227 18:42:14.706234 4700 scope.go:117] "RemoveContainer" containerID="d487919697c779b8b8845178b947df4809622ee74a2201b36d2cf94c70270ba3" Feb 27 18:42:27 crc kubenswrapper[4700]: I0227 18:42:27.981509 4700 scope.go:117] "RemoveContainer" containerID="c7ad608c50ddee2aac4992df7a6ccaaeda7c737e2a0d2fad5282b4482eac9e50" Feb 27 18:42:27 crc kubenswrapper[4700]: E0227 18:42:27.982661 4700 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-f9dq7_openshift-machine-config-operator(84b45511-b94c-479f-98db-fd2c4eceec46)\"" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" Feb 27 18:42:38 crc kubenswrapper[4700]: I0227 18:42:38.981842 4700 scope.go:117] "RemoveContainer" containerID="c7ad608c50ddee2aac4992df7a6ccaaeda7c737e2a0d2fad5282b4482eac9e50" Feb 27 18:42:39 crc kubenswrapper[4700]: I0227 18:42:39.858678 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerStarted","Data":"b47e047e5315ba4c8f50802302b0905966f67ab104d0cf961afb16551a78daa7"} Feb 27 18:42:44 crc kubenswrapper[4700]: I0227 18:42:44.160678 4700 patch_prober.go:28] interesting pod/route-controller-manager-77b96b4458-bm8hf container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 18:42:44 crc kubenswrapper[4700]: I0227 18:42:44.161342 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf" podUID="21dd241c-061a-4e48-b039-f5877cdc8aba" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 27 18:43:22 crc kubenswrapper[4700]: I0227 18:43:22.773679 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/telemetry-operator-controller-manager-5fdb694969-44djz" podUID="629a3ab5-c8ce-49cb-bd57-355d643253c5" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.99:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 27 18:43:23 crc kubenswrapper[4700]: I0227 18:43:23.751688 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="0a50deb6-9983-40f1-bf2c-f8ce7584e809" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Feb 27 18:43:23 crc kubenswrapper[4700]: I0227 18:43:23.754420 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="0a50deb6-9983-40f1-bf2c-f8ce7584e809" containerName="ceilometer-notification-agent" probeResult="failure" output="command timed out" Feb 27 18:43:28 crc kubenswrapper[4700]: I0227 18:43:28.753352 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="0a50deb6-9983-40f1-bf2c-f8ce7584e809" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Feb 27 18:43:33 crc kubenswrapper[4700]: I0227 18:43:33.753296 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="0a50deb6-9983-40f1-bf2c-f8ce7584e809" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Feb 27 18:43:33 crc kubenswrapper[4700]: I0227 18:43:33.755912 4700 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ceilometer-0" Feb 27 18:43:33 crc kubenswrapper[4700]: I0227 18:43:33.757294 4700 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="ceilometer-central-agent" containerStatusID={"Type":"cri-o","ID":"2c30263af4bf57a3c0b2c9a34c54fd66f9b9aeaca3163c644e80399eaa73104f"} pod="openstack/ceilometer-0" containerMessage="Container ceilometer-central-agent failed liveness probe, will be restarted" Feb 27 18:43:33 crc kubenswrapper[4700]: I0227 18:43:33.757685 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0a50deb6-9983-40f1-bf2c-f8ce7584e809" containerName="ceilometer-central-agent" containerID="cri-o://2c30263af4bf57a3c0b2c9a34c54fd66f9b9aeaca3163c644e80399eaa73104f" gracePeriod=30 Feb 27 18:43:34 crc kubenswrapper[4700]: I0227 18:43:34.164696 4700 patch_prober.go:28] interesting pod/route-controller-manager-77b96b4458-bm8hf container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 18:43:34 crc kubenswrapper[4700]: I0227 18:43:34.164710 4700 patch_prober.go:28] interesting pod/route-controller-manager-77b96b4458-bm8hf container/route-controller-manager namespace/openshift-route-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.68:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 18:43:34 crc kubenswrapper[4700]: I0227 18:43:34.164777 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf" podUID="21dd241c-061a-4e48-b039-f5877cdc8aba" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 27 18:43:34 crc kubenswrapper[4700]: I0227 18:43:34.164853 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf" podUID="21dd241c-061a-4e48-b039-f5877cdc8aba" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.68:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 27 18:43:42 crc kubenswrapper[4700]: I0227 18:43:42.583709 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/mariadb-operator-controller-manager-556b8b874-4bmbt" podUID="c1fdff54-9c43-4fe4-925b-f87590c75e72" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.90:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 27 18:43:44 crc kubenswrapper[4700]: I0227 18:43:44.156574 4700 patch_prober.go:28] interesting pod/route-controller-manager-77b96b4458-bm8hf container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 18:43:44 crc kubenswrapper[4700]: I0227 18:43:44.156936 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf" podUID="21dd241c-061a-4e48-b039-f5877cdc8aba" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 18:43:44 crc kubenswrapper[4700]: I0227 18:43:44.160701 4700 patch_prober.go:28] interesting pod/route-controller-manager-77b96b4458-bm8hf container/route-controller-manager namespace/openshift-route-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 18:43:44 crc kubenswrapper[4700]: I0227 18:43:44.160791 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf" podUID="21dd241c-061a-4e48-b039-f5877cdc8aba" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 27 18:43:46 crc kubenswrapper[4700]: I0227 18:43:46.259536 4700 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-btxgl container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 18:43:46 crc kubenswrapper[4700]: I0227 18:43:46.259838 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-btxgl" podUID="91eb41f7-eb34-496d-b65d-20dfc7abf7f0" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 18:43:52 crc kubenswrapper[4700]: I0227 18:43:52.583693 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/mariadb-operator-controller-manager-556b8b874-4bmbt" podUID="c1fdff54-9c43-4fe4-925b-f87590c75e72" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.90:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 27 18:43:53 crc kubenswrapper[4700]: I0227 18:43:53.754198 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="0a50deb6-9983-40f1-bf2c-f8ce7584e809" containerName="ceilometer-notification-agent" probeResult="failure" output="command timed out" Feb 27 18:43:54 crc kubenswrapper[4700]: I0227 18:43:54.156680 4700 patch_prober.go:28] interesting pod/route-controller-manager-77b96b4458-bm8hf container/route-controller-manager namespace/openshift-route-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 18:43:54 crc kubenswrapper[4700]: I0227 18:43:54.157242 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf" podUID="21dd241c-061a-4e48-b039-f5877cdc8aba" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 18:43:54 crc kubenswrapper[4700]: I0227 18:43:54.157329 4700 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf" Feb 27 18:43:54 crc kubenswrapper[4700]: I0227 18:43:54.158918 4700 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="route-controller-manager" containerStatusID={"Type":"cri-o","ID":"57516f85c29d09c8565695064903ec1f31d6e6b6c6e8e8cf2ed636ca80fa1255"} pod="openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf" containerMessage="Container route-controller-manager failed liveness probe, will be restarted" Feb 27 18:43:54 crc kubenswrapper[4700]: I0227 18:43:54.158989 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf" podUID="21dd241c-061a-4e48-b039-f5877cdc8aba" containerName="route-controller-manager" containerID="cri-o://57516f85c29d09c8565695064903ec1f31d6e6b6c6e8e8cf2ed636ca80fa1255" gracePeriod=30 Feb 27 18:43:54 crc kubenswrapper[4700]: I0227 18:43:54.161586 4700 patch_prober.go:28] interesting pod/route-controller-manager-77b96b4458-bm8hf container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 18:43:54 crc kubenswrapper[4700]: I0227 18:43:54.161686 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf" podUID="21dd241c-061a-4e48-b039-f5877cdc8aba" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 27 18:43:58 crc kubenswrapper[4700]: I0227 18:43:58.260125 4700 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-btxgl container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 18:43:58 crc kubenswrapper[4700]: I0227 18:43:58.260210 4700 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-btxgl container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 18:43:58 crc kubenswrapper[4700]: I0227 18:43:58.260706 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-btxgl" podUID="91eb41f7-eb34-496d-b65d-20dfc7abf7f0" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 18:43:58 crc kubenswrapper[4700]: I0227 18:43:58.260737 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-btxgl" podUID="91eb41f7-eb34-496d-b65d-20dfc7abf7f0" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 18:44:00 crc kubenswrapper[4700]: I0227 18:44:00.212327 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536964-828mg"] Feb 27 18:44:00 crc kubenswrapper[4700]: E0227 18:44:00.213097 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ee20344-c058-4d9c-93a4-fe301ccf6b9e" containerName="oc" Feb 27 18:44:00 crc kubenswrapper[4700]: I0227 18:44:00.213113 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ee20344-c058-4d9c-93a4-fe301ccf6b9e" containerName="oc" Feb 27 18:44:00 crc kubenswrapper[4700]: I0227 18:44:00.213405 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ee20344-c058-4d9c-93a4-fe301ccf6b9e" containerName="oc" Feb 27 18:44:00 crc kubenswrapper[4700]: I0227 18:44:00.214252 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536964-828mg" Feb 27 18:44:00 crc kubenswrapper[4700]: I0227 18:44:00.217839 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 18:44:00 crc kubenswrapper[4700]: I0227 18:44:00.217952 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d62kc" Feb 27 18:44:00 crc kubenswrapper[4700]: I0227 18:44:00.218189 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 18:44:00 crc kubenswrapper[4700]: I0227 18:44:00.224775 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536964-828mg"] Feb 27 18:44:00 crc kubenswrapper[4700]: I0227 18:44:00.325238 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snx2k\" (UniqueName: \"kubernetes.io/projected/0f0c5a56-55ed-4dee-b36f-8faacb74ddbb-kube-api-access-snx2k\") pod \"auto-csr-approver-29536964-828mg\" (UID: \"0f0c5a56-55ed-4dee-b36f-8faacb74ddbb\") " pod="openshift-infra/auto-csr-approver-29536964-828mg" Feb 27 18:44:00 crc kubenswrapper[4700]: I0227 18:44:00.427136 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snx2k\" (UniqueName: \"kubernetes.io/projected/0f0c5a56-55ed-4dee-b36f-8faacb74ddbb-kube-api-access-snx2k\") pod \"auto-csr-approver-29536964-828mg\" (UID: \"0f0c5a56-55ed-4dee-b36f-8faacb74ddbb\") " pod="openshift-infra/auto-csr-approver-29536964-828mg" Feb 27 18:44:00 crc kubenswrapper[4700]: I0227 18:44:00.449039 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snx2k\" (UniqueName: \"kubernetes.io/projected/0f0c5a56-55ed-4dee-b36f-8faacb74ddbb-kube-api-access-snx2k\") pod \"auto-csr-approver-29536964-828mg\" (UID: \"0f0c5a56-55ed-4dee-b36f-8faacb74ddbb\") " pod="openshift-infra/auto-csr-approver-29536964-828mg" Feb 27 18:44:00 crc kubenswrapper[4700]: I0227 18:44:00.528892 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536964-828mg" Feb 27 18:44:01 crc kubenswrapper[4700]: I0227 18:44:01.033891 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536964-828mg"] Feb 27 18:44:01 crc kubenswrapper[4700]: I0227 18:44:01.259356 4700 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-btxgl container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 18:44:01 crc kubenswrapper[4700]: I0227 18:44:01.259693 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-btxgl" podUID="91eb41f7-eb34-496d-b65d-20dfc7abf7f0" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 18:44:01 crc kubenswrapper[4700]: I0227 18:44:01.259392 4700 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-btxgl container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 18:44:01 crc kubenswrapper[4700]: I0227 18:44:01.259819 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-btxgl" podUID="91eb41f7-eb34-496d-b65d-20dfc7abf7f0" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 18:44:01 crc kubenswrapper[4700]: I0227 18:44:01.886117 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536964-828mg" event={"ID":"0f0c5a56-55ed-4dee-b36f-8faacb74ddbb","Type":"ContainerStarted","Data":"9d19055699c59904be86433776f50da475b64be1571b685b73c626c7a099978a"} Feb 27 18:44:04 crc kubenswrapper[4700]: I0227 18:44:04.160713 4700 patch_prober.go:28] interesting pod/route-controller-manager-77b96b4458-bm8hf container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 18:44:04 crc kubenswrapper[4700]: I0227 18:44:04.161106 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf" podUID="21dd241c-061a-4e48-b039-f5877cdc8aba" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 27 18:44:04 crc kubenswrapper[4700]: I0227 18:44:04.262675 4700 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-btxgl container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 18:44:04 crc kubenswrapper[4700]: I0227 18:44:04.262731 4700 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-btxgl container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 18:44:04 crc kubenswrapper[4700]: I0227 18:44:04.262750 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-btxgl" podUID="91eb41f7-eb34-496d-b65d-20dfc7abf7f0" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 18:44:04 crc kubenswrapper[4700]: I0227 18:44:04.262784 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-btxgl" podUID="91eb41f7-eb34-496d-b65d-20dfc7abf7f0" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 18:44:04 crc kubenswrapper[4700]: I0227 18:44:04.262862 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-btxgl" Feb 27 18:44:04 crc kubenswrapper[4700]: I0227 18:44:04.262895 4700 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-config-operator/openshift-config-operator-7777fb866f-btxgl" Feb 27 18:44:04 crc kubenswrapper[4700]: I0227 18:44:04.263844 4700 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="openshift-config-operator" containerStatusID={"Type":"cri-o","ID":"b098603a8b33cfd21624b9e449a9096954f7000132a4fa6026d0684ca90830b5"} pod="openshift-config-operator/openshift-config-operator-7777fb866f-btxgl" containerMessage="Container openshift-config-operator failed liveness probe, will be restarted" Feb 27 18:44:04 crc kubenswrapper[4700]: I0227 18:44:04.263886 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-config-operator/openshift-config-operator-7777fb866f-btxgl" podUID="91eb41f7-eb34-496d-b65d-20dfc7abf7f0" containerName="openshift-config-operator" containerID="cri-o://b098603a8b33cfd21624b9e449a9096954f7000132a4fa6026d0684ca90830b5" gracePeriod=30 Feb 27 18:44:05 crc kubenswrapper[4700]: I0227 18:44:05.263652 4700 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-btxgl container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 18:44:05 crc kubenswrapper[4700]: I0227 18:44:05.263767 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-btxgl" podUID="91eb41f7-eb34-496d-b65d-20dfc7abf7f0" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 18:44:07 crc kubenswrapper[4700]: I0227 18:44:07.260095 4700 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-btxgl container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 18:44:07 crc kubenswrapper[4700]: I0227 18:44:07.261054 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-btxgl" podUID="91eb41f7-eb34-496d-b65d-20dfc7abf7f0" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 18:44:10 crc kubenswrapper[4700]: I0227 18:44:10.259392 4700 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-btxgl container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 18:44:10 crc kubenswrapper[4700]: I0227 18:44:10.260051 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-btxgl" podUID="91eb41f7-eb34-496d-b65d-20dfc7abf7f0" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 18:44:12 crc kubenswrapper[4700]: I0227 18:44:12.797656 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-xgf2x" podUID="269a9c0b-7242-44dc-b860-386757217052" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.100:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 27 18:44:12 crc kubenswrapper[4700]: I0227 18:44:12.797679 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-xgf2x" podUID="269a9c0b-7242-44dc-b860-386757217052" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.100:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 27 18:44:13 crc kubenswrapper[4700]: I0227 18:44:13.259955 4700 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-btxgl container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 18:44:13 crc kubenswrapper[4700]: I0227 18:44:13.260037 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-btxgl" podUID="91eb41f7-eb34-496d-b65d-20dfc7abf7f0" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 18:44:14 crc kubenswrapper[4700]: I0227 18:44:14.005498 4700 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:6443/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 18:44:14 crc kubenswrapper[4700]: I0227 18:44:14.005543 4700 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:6443/livez?exclude=etcd\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 18:44:14 crc kubenswrapper[4700]: I0227 18:44:14.006121 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/livez?exclude=etcd\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 18:44:14 crc kubenswrapper[4700]: I0227 18:44:14.006220 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 27 18:44:14 crc kubenswrapper[4700]: I0227 18:44:14.157029 4700 patch_prober.go:28] interesting pod/route-controller-manager-77b96b4458-bm8hf container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 18:44:14 crc kubenswrapper[4700]: I0227 18:44:14.157098 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf" podUID="21dd241c-061a-4e48-b039-f5877cdc8aba" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 18:44:14 crc kubenswrapper[4700]: I0227 18:44:14.296127 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/kube-state-metrics-0" podUID="f100325f-1efc-423e-9e23-e964c5547135" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.0.225:8080/livez\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 27 18:44:15 crc kubenswrapper[4700]: I0227 18:44:15.383724 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-btxgl" Feb 27 18:44:16 crc kubenswrapper[4700]: I0227 18:44:16.053893 4700 generic.go:334] "Generic (PLEG): container finished" podID="f6a3d37c-2b18-40bc-a990-56dc67093abf" containerID="9b3bda1dc90f49c9ff580eb83c7a8226a2c585ebb5d1e3050e741877bb0048bd" exitCode=1 Feb 27 18:44:16 crc kubenswrapper[4700]: I0227 18:44:16.053998 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg" event={"ID":"f6a3d37c-2b18-40bc-a990-56dc67093abf","Type":"ContainerDied","Data":"9b3bda1dc90f49c9ff580eb83c7a8226a2c585ebb5d1e3050e741877bb0048bd"} Feb 27 18:44:16 crc kubenswrapper[4700]: I0227 18:44:16.056405 4700 scope.go:117] "RemoveContainer" containerID="9b3bda1dc90f49c9ff580eb83c7a8226a2c585ebb5d1e3050e741877bb0048bd" Feb 27 18:44:16 crc kubenswrapper[4700]: I0227 18:44:16.062090 4700 generic.go:334] "Generic (PLEG): container finished" podID="21dd241c-061a-4e48-b039-f5877cdc8aba" containerID="57516f85c29d09c8565695064903ec1f31d6e6b6c6e8e8cf2ed636ca80fa1255" exitCode=0 Feb 27 18:44:16 crc kubenswrapper[4700]: I0227 18:44:16.062203 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf" event={"ID":"21dd241c-061a-4e48-b039-f5877cdc8aba","Type":"ContainerDied","Data":"57516f85c29d09c8565695064903ec1f31d6e6b6c6e8e8cf2ed636ca80fa1255"} Feb 27 18:44:16 crc kubenswrapper[4700]: I0227 18:44:16.071590 4700 generic.go:334] "Generic (PLEG): container finished" podID="00a21bc7-24a6-4a6d-ba48-18db1d9233c6" containerID="5910fd2a75940d588d7efce67c326ac6bb5acc14456bbbf5a5b0cd5cf441df15" exitCode=1 Feb 27 18:44:16 crc kubenswrapper[4700]: I0227 18:44:16.071657 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-75684d597f-6cfj6" event={"ID":"00a21bc7-24a6-4a6d-ba48-18db1d9233c6","Type":"ContainerDied","Data":"5910fd2a75940d588d7efce67c326ac6bb5acc14456bbbf5a5b0cd5cf441df15"} Feb 27 18:44:16 crc kubenswrapper[4700]: I0227 18:44:16.072161 4700 scope.go:117] "RemoveContainer" containerID="5910fd2a75940d588d7efce67c326ac6bb5acc14456bbbf5a5b0cd5cf441df15" Feb 27 18:44:16 crc kubenswrapper[4700]: I0227 18:44:16.118162 4700 generic.go:334] "Generic (PLEG): container finished" podID="91eb41f7-eb34-496d-b65d-20dfc7abf7f0" containerID="b098603a8b33cfd21624b9e449a9096954f7000132a4fa6026d0684ca90830b5" exitCode=0 Feb 27 18:44:16 crc kubenswrapper[4700]: I0227 18:44:16.118250 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-btxgl" event={"ID":"91eb41f7-eb34-496d-b65d-20dfc7abf7f0","Type":"ContainerDied","Data":"b098603a8b33cfd21624b9e449a9096954f7000132a4fa6026d0684ca90830b5"} Feb 27 18:44:16 crc kubenswrapper[4700]: I0227 18:44:16.138476 4700 generic.go:334] "Generic (PLEG): container finished" podID="269a9c0b-7242-44dc-b860-386757217052" containerID="3160bda50adeb5e779a8a3dd3d559bd716f376bdb530dd70b547d38a9f57686b" exitCode=1 Feb 27 18:44:16 crc kubenswrapper[4700]: I0227 18:44:16.138527 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-xgf2x" event={"ID":"269a9c0b-7242-44dc-b860-386757217052","Type":"ContainerDied","Data":"3160bda50adeb5e779a8a3dd3d559bd716f376bdb530dd70b547d38a9f57686b"} Feb 27 18:44:16 crc kubenswrapper[4700]: I0227 18:44:16.139308 4700 scope.go:117] "RemoveContainer" containerID="3160bda50adeb5e779a8a3dd3d559bd716f376bdb530dd70b547d38a9f57686b" Feb 27 18:44:16 crc kubenswrapper[4700]: I0227 18:44:16.152787 4700 generic.go:334] "Generic (PLEG): container finished" podID="1d823d65-036b-4040-af8a-9b6698071d32" containerID="fce40df3dc9d2d61893154a6aaf99d4b24e2acd4d1a0f9ed79a57e8f15f12f4e" exitCode=1 Feb 27 18:44:16 crc kubenswrapper[4700]: I0227 18:44:16.152891 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-cc5d6c7cf-x6qzc" event={"ID":"1d823d65-036b-4040-af8a-9b6698071d32","Type":"ContainerDied","Data":"fce40df3dc9d2d61893154a6aaf99d4b24e2acd4d1a0f9ed79a57e8f15f12f4e"} Feb 27 18:44:16 crc kubenswrapper[4700]: I0227 18:44:16.153515 4700 scope.go:117] "RemoveContainer" containerID="fce40df3dc9d2d61893154a6aaf99d4b24e2acd4d1a0f9ed79a57e8f15f12f4e" Feb 27 18:44:16 crc kubenswrapper[4700]: I0227 18:44:16.163307 4700 generic.go:334] "Generic (PLEG): container finished" podID="e00942b5-6f82-48d4-8413-2f568daf5d7f" containerID="248645c71439bdeefedba807c37d79e4f14d17653d863f4eadc1add08df3142d" exitCode=1 Feb 27 18:44:16 crc kubenswrapper[4700]: I0227 18:44:16.163869 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-54688575f-8vhl9" event={"ID":"e00942b5-6f82-48d4-8413-2f568daf5d7f","Type":"ContainerDied","Data":"248645c71439bdeefedba807c37d79e4f14d17653d863f4eadc1add08df3142d"} Feb 27 18:44:16 crc kubenswrapper[4700]: I0227 18:44:16.176039 4700 scope.go:117] "RemoveContainer" containerID="248645c71439bdeefedba807c37d79e4f14d17653d863f4eadc1add08df3142d" Feb 27 18:44:16 crc kubenswrapper[4700]: I0227 18:44:16.184088 4700 generic.go:334] "Generic (PLEG): container finished" podID="c1fdff54-9c43-4fe4-925b-f87590c75e72" containerID="addc0e2864607524e975a6f9978b3beb1b9fb3eea2493fdb9f54af42143ddec5" exitCode=1 Feb 27 18:44:16 crc kubenswrapper[4700]: I0227 18:44:16.184172 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-556b8b874-4bmbt" event={"ID":"c1fdff54-9c43-4fe4-925b-f87590c75e72","Type":"ContainerDied","Data":"addc0e2864607524e975a6f9978b3beb1b9fb3eea2493fdb9f54af42143ddec5"} Feb 27 18:44:16 crc kubenswrapper[4700]: I0227 18:44:16.185146 4700 scope.go:117] "RemoveContainer" containerID="addc0e2864607524e975a6f9978b3beb1b9fb3eea2493fdb9f54af42143ddec5" Feb 27 18:44:16 crc kubenswrapper[4700]: I0227 18:44:16.258815 4700 generic.go:334] "Generic (PLEG): container finished" podID="9c1db815-93de-4bcc-9698-22bf6588ac88" containerID="4762e7fa2995139519f13446bac228e27846df633ed053edabd012a9604ef662" exitCode=1 Feb 27 18:44:16 crc kubenswrapper[4700]: I0227 18:44:16.259175 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-545456dc4-stc8v" event={"ID":"9c1db815-93de-4bcc-9698-22bf6588ac88","Type":"ContainerDied","Data":"4762e7fa2995139519f13446bac228e27846df633ed053edabd012a9604ef662"} Feb 27 18:44:16 crc kubenswrapper[4700]: I0227 18:44:16.268560 4700 scope.go:117] "RemoveContainer" containerID="4762e7fa2995139519f13446bac228e27846df633ed053edabd012a9604ef662" Feb 27 18:44:16 crc kubenswrapper[4700]: I0227 18:44:16.268935 4700 generic.go:334] "Generic (PLEG): container finished" podID="bd70bcca-beb1-464b-a1a3-b154ff45c1f5" containerID="4bcb3d4f5af1b7fd0b6949c341ab10de90340503d697ebf122758a1985fe3cd9" exitCode=1 Feb 27 18:44:16 crc kubenswrapper[4700]: I0227 18:44:16.268980 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-vt248" event={"ID":"bd70bcca-beb1-464b-a1a3-b154ff45c1f5","Type":"ContainerDied","Data":"4bcb3d4f5af1b7fd0b6949c341ab10de90340503d697ebf122758a1985fe3cd9"} Feb 27 18:44:16 crc kubenswrapper[4700]: I0227 18:44:16.304432 4700 scope.go:117] "RemoveContainer" containerID="4bcb3d4f5af1b7fd0b6949c341ab10de90340503d697ebf122758a1985fe3cd9" Feb 27 18:44:16 crc kubenswrapper[4700]: E0227 18:44:16.638662 4700 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d823d65_036b_4040_af8a_9b6698071d32.slice/crio-conmon-fce40df3dc9d2d61893154a6aaf99d4b24e2acd4d1a0f9ed79a57e8f15f12f4e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod00a21bc7_24a6_4a6d_ba48_18db1d9233c6.slice/crio-conmon-5910fd2a75940d588d7efce67c326ac6bb5acc14456bbbf5a5b0cd5cf441df15.scope\": RecentStats: unable to find data in memory cache]" Feb 27 18:44:17 crc kubenswrapper[4700]: I0227 18:44:17.237321 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg" Feb 27 18:44:17 crc kubenswrapper[4700]: I0227 18:44:17.286856 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-xgf2x" event={"ID":"269a9c0b-7242-44dc-b860-386757217052","Type":"ContainerStarted","Data":"0f2bf29c485fbfc9b93b989380f35db003b1bbd2546d425f2fb8f300a9b8bb2b"} Feb 27 18:44:17 crc kubenswrapper[4700]: I0227 18:44:17.287852 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-xgf2x" Feb 27 18:44:17 crc kubenswrapper[4700]: I0227 18:44:17.292639 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-54688575f-8vhl9" event={"ID":"e00942b5-6f82-48d4-8413-2f568daf5d7f","Type":"ContainerStarted","Data":"0fcc3c8ef1af86c260726120460285f3d9f5a955ed8e6c1648e073b18f389e65"} Feb 27 18:44:17 crc kubenswrapper[4700]: I0227 18:44:17.293050 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-54688575f-8vhl9" Feb 27 18:44:17 crc kubenswrapper[4700]: I0227 18:44:17.295102 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg" event={"ID":"f6a3d37c-2b18-40bc-a990-56dc67093abf","Type":"ContainerStarted","Data":"060475e5d622b6e65ede845f2844c263758a36100820e2dc950b949f92ae0810"} Feb 27 18:44:17 crc kubenswrapper[4700]: I0227 18:44:17.295672 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg" Feb 27 18:44:17 crc kubenswrapper[4700]: I0227 18:44:17.298208 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf" Feb 27 18:44:17 crc kubenswrapper[4700]: I0227 18:44:17.298513 4700 patch_prober.go:28] interesting pod/route-controller-manager-77b96b4458-bm8hf container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.68:8443/healthz\": dial tcp 10.217.0.68:8443: connect: connection refused" start-of-body= Feb 27 18:44:17 crc kubenswrapper[4700]: I0227 18:44:17.298618 4700 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf" podUID="21dd241c-061a-4e48-b039-f5877cdc8aba" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.68:8443/healthz\": dial tcp 10.217.0.68:8443: connect: connection refused" Feb 27 18:44:17 crc kubenswrapper[4700]: I0227 18:44:17.300624 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-545456dc4-stc8v" event={"ID":"9c1db815-93de-4bcc-9698-22bf6588ac88","Type":"ContainerStarted","Data":"228612e3be8774152b2a8dab752e77333efd43effde3ce64c1efdcc8506d8f2e"} Feb 27 18:44:17 crc kubenswrapper[4700]: I0227 18:44:17.301049 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-545456dc4-stc8v" Feb 27 18:44:17 crc kubenswrapper[4700]: I0227 18:44:17.303754 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-cc5d6c7cf-x6qzc" event={"ID":"1d823d65-036b-4040-af8a-9b6698071d32","Type":"ContainerStarted","Data":"09aac5b7e1bfbf9587556a8a7e753b6ef413959b919cdb91c023193f7c887210"} Feb 27 18:44:17 crc kubenswrapper[4700]: I0227 18:44:17.303970 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-cc5d6c7cf-x6qzc" Feb 27 18:44:17 crc kubenswrapper[4700]: I0227 18:44:17.306368 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-btxgl" event={"ID":"91eb41f7-eb34-496d-b65d-20dfc7abf7f0","Type":"ContainerStarted","Data":"26cd2537e31d7d8245f7864284f9ad9b21b594751543d1636f09f9bf8943bca3"} Feb 27 18:44:17 crc kubenswrapper[4700]: I0227 18:44:17.306492 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-btxgl" Feb 27 18:44:17 crc kubenswrapper[4700]: I0227 18:44:17.308518 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536964-828mg" event={"ID":"0f0c5a56-55ed-4dee-b36f-8faacb74ddbb","Type":"ContainerStarted","Data":"8f939e1f3327ca83f80ca192d29ea9cd883187cef65cd0577cb3378690a83e08"} Feb 27 18:44:17 crc kubenswrapper[4700]: I0227 18:44:17.310807 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-556b8b874-4bmbt" event={"ID":"c1fdff54-9c43-4fe4-925b-f87590c75e72","Type":"ContainerStarted","Data":"2cae50e50acd1600a5f661927f72acd7a8b406f2c62e47d32671486948d1df9d"} Feb 27 18:44:17 crc kubenswrapper[4700]: I0227 18:44:17.311096 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-556b8b874-4bmbt" Feb 27 18:44:17 crc kubenswrapper[4700]: I0227 18:44:17.312410 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-vt248" event={"ID":"bd70bcca-beb1-464b-a1a3-b154ff45c1f5","Type":"ContainerStarted","Data":"25579dc8a6d412c617b98209043267dd93166c47d9f202f713c43143e97b5604"} Feb 27 18:44:17 crc kubenswrapper[4700]: I0227 18:44:17.312990 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-vt248" Feb 27 18:44:17 crc kubenswrapper[4700]: I0227 18:44:17.314327 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-75684d597f-6cfj6" event={"ID":"00a21bc7-24a6-4a6d-ba48-18db1d9233c6","Type":"ContainerStarted","Data":"7df75565b9da7170c469da8e6a918953a1e05f231f793de50ae1b787dbdde14a"} Feb 27 18:44:17 crc kubenswrapper[4700]: I0227 18:44:17.314851 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-75684d597f-6cfj6" Feb 27 18:44:17 crc kubenswrapper[4700]: I0227 18:44:17.434805 4700 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536964-828mg" podStartSLOduration=2.443949576 podStartE2EDuration="17.43478715s" podCreationTimestamp="2026-02-27 18:44:00 +0000 UTC" firstStartedPulling="2026-02-27 18:44:01.04338055 +0000 UTC m=+6201.028693297" lastFinishedPulling="2026-02-27 18:44:16.034218124 +0000 UTC m=+6216.019530871" observedRunningTime="2026-02-27 18:44:17.428850125 +0000 UTC m=+6217.414162872" watchObservedRunningTime="2026-02-27 18:44:17.43478715 +0000 UTC m=+6217.420099887" Feb 27 18:44:18 crc kubenswrapper[4700]: I0227 18:44:18.323159 4700 generic.go:334] "Generic (PLEG): container finished" podID="0f0c5a56-55ed-4dee-b36f-8faacb74ddbb" containerID="8f939e1f3327ca83f80ca192d29ea9cd883187cef65cd0577cb3378690a83e08" exitCode=0 Feb 27 18:44:18 crc kubenswrapper[4700]: I0227 18:44:18.323258 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536964-828mg" event={"ID":"0f0c5a56-55ed-4dee-b36f-8faacb74ddbb","Type":"ContainerDied","Data":"8f939e1f3327ca83f80ca192d29ea9cd883187cef65cd0577cb3378690a83e08"} Feb 27 18:44:18 crc kubenswrapper[4700]: I0227 18:44:18.327388 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf" event={"ID":"21dd241c-061a-4e48-b039-f5877cdc8aba","Type":"ContainerStarted","Data":"e3a37b228b796931a0ef511d6f6aac1c0f7e668321a6db0866d07109ca3b6d6c"} Feb 27 18:44:18 crc kubenswrapper[4700]: I0227 18:44:18.375208 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-77b96b4458-bm8hf" Feb 27 18:44:18 crc kubenswrapper[4700]: I0227 18:44:18.810583 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="0a50deb6-9983-40f1-bf2c-f8ce7584e809" containerName="ceilometer-notification-agent" probeResult="failure" output=< Feb 27 18:44:18 crc kubenswrapper[4700]: Unkown error: Expecting value: line 1 column 1 (char 0) Feb 27 18:44:18 crc kubenswrapper[4700]: > Feb 27 18:44:18 crc kubenswrapper[4700]: I0227 18:44:18.810657 4700 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ceilometer-0" Feb 27 18:44:19 crc kubenswrapper[4700]: I0227 18:44:19.677637 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536964-828mg" Feb 27 18:44:19 crc kubenswrapper[4700]: I0227 18:44:19.827062 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snx2k\" (UniqueName: \"kubernetes.io/projected/0f0c5a56-55ed-4dee-b36f-8faacb74ddbb-kube-api-access-snx2k\") pod \"0f0c5a56-55ed-4dee-b36f-8faacb74ddbb\" (UID: \"0f0c5a56-55ed-4dee-b36f-8faacb74ddbb\") " Feb 27 18:44:19 crc kubenswrapper[4700]: I0227 18:44:19.836472 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f0c5a56-55ed-4dee-b36f-8faacb74ddbb-kube-api-access-snx2k" (OuterVolumeSpecName: "kube-api-access-snx2k") pod "0f0c5a56-55ed-4dee-b36f-8faacb74ddbb" (UID: "0f0c5a56-55ed-4dee-b36f-8faacb74ddbb"). InnerVolumeSpecName "kube-api-access-snx2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:44:19 crc kubenswrapper[4700]: I0227 18:44:19.929869 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snx2k\" (UniqueName: \"kubernetes.io/projected/0f0c5a56-55ed-4dee-b36f-8faacb74ddbb-kube-api-access-snx2k\") on node \"crc\" DevicePath \"\"" Feb 27 18:44:20 crc kubenswrapper[4700]: I0227 18:44:20.351093 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536964-828mg" event={"ID":"0f0c5a56-55ed-4dee-b36f-8faacb74ddbb","Type":"ContainerDied","Data":"9d19055699c59904be86433776f50da475b64be1571b685b73c626c7a099978a"} Feb 27 18:44:20 crc kubenswrapper[4700]: I0227 18:44:20.351160 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d19055699c59904be86433776f50da475b64be1571b685b73c626c7a099978a" Feb 27 18:44:20 crc kubenswrapper[4700]: I0227 18:44:20.351307 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536964-828mg" Feb 27 18:44:20 crc kubenswrapper[4700]: I0227 18:44:20.757922 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536958-tdbth"] Feb 27 18:44:20 crc kubenswrapper[4700]: I0227 18:44:20.768533 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536958-tdbth"] Feb 27 18:44:20 crc kubenswrapper[4700]: I0227 18:44:20.999003 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b103b0e5-f0f5-4659-87d5-40a4d4edf5b7" path="/var/lib/kubelet/pods/b103b0e5-f0f5-4659-87d5-40a4d4edf5b7/volumes" Feb 27 18:44:21 crc kubenswrapper[4700]: I0227 18:44:21.263306 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-btxgl" Feb 27 18:44:21 crc kubenswrapper[4700]: I0227 18:44:21.371507 4700 generic.go:334] "Generic (PLEG): container finished" podID="0a50deb6-9983-40f1-bf2c-f8ce7584e809" containerID="2c30263af4bf57a3c0b2c9a34c54fd66f9b9aeaca3163c644e80399eaa73104f" exitCode=137 Feb 27 18:44:21 crc kubenswrapper[4700]: I0227 18:44:21.371561 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a50deb6-9983-40f1-bf2c-f8ce7584e809","Type":"ContainerDied","Data":"2c30263af4bf57a3c0b2c9a34c54fd66f9b9aeaca3163c644e80399eaa73104f"} Feb 27 18:44:22 crc kubenswrapper[4700]: I0227 18:44:22.389784 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a50deb6-9983-40f1-bf2c-f8ce7584e809","Type":"ContainerStarted","Data":"f44662def1c3c615b83316c9007605cddfc22abc379baf98a28423ab1ba5de08"} Feb 27 18:44:22 crc kubenswrapper[4700]: I0227 18:44:22.390382 4700 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="ceilometer-notification-agent" containerStatusID={"Type":"cri-o","ID":"8a2eaeba8408e444f6fc8624878ea574e19cf802c026e7b69c93814e26028409"} pod="openstack/ceilometer-0" containerMessage="Container ceilometer-notification-agent failed liveness probe, will be restarted" Feb 27 18:44:22 crc kubenswrapper[4700]: I0227 18:44:22.390446 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0a50deb6-9983-40f1-bf2c-f8ce7584e809" containerName="ceilometer-notification-agent" containerID="cri-o://8a2eaeba8408e444f6fc8624878ea574e19cf802c026e7b69c93814e26028409" gracePeriod=30 Feb 27 18:44:26 crc kubenswrapper[4700]: I0227 18:44:26.433314 4700 generic.go:334] "Generic (PLEG): container finished" podID="0a50deb6-9983-40f1-bf2c-f8ce7584e809" containerID="8a2eaeba8408e444f6fc8624878ea574e19cf802c026e7b69c93814e26028409" exitCode=0 Feb 27 18:44:26 crc kubenswrapper[4700]: I0227 18:44:26.433961 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a50deb6-9983-40f1-bf2c-f8ce7584e809","Type":"ContainerDied","Data":"8a2eaeba8408e444f6fc8624878ea574e19cf802c026e7b69c93814e26028409"} Feb 27 18:44:26 crc kubenswrapper[4700]: I0227 18:44:26.435403 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a50deb6-9983-40f1-bf2c-f8ce7584e809","Type":"ContainerStarted","Data":"e2e3cac10b873cb27919d1c2e7e627a95fe204abcc0a97e20c6e1a80757cad10"} Feb 27 18:44:27 crc kubenswrapper[4700]: I0227 18:44:27.245992 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9ckpbqg" Feb 27 18:44:31 crc kubenswrapper[4700]: I0227 18:44:31.369965 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-545456dc4-stc8v" Feb 27 18:44:31 crc kubenswrapper[4700]: I0227 18:44:31.547307 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-556b8b874-4bmbt" Feb 27 18:44:31 crc kubenswrapper[4700]: I0227 18:44:31.557850 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-54688575f-8vhl9" Feb 27 18:44:31 crc kubenswrapper[4700]: I0227 18:44:31.606129 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-5d86c7ddb7-vt248" Feb 27 18:44:31 crc kubenswrapper[4700]: I0227 18:44:31.670498 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-75684d597f-6cfj6" Feb 27 18:44:31 crc kubenswrapper[4700]: I0227 18:44:31.757898 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-55b5ff4dbb-xgf2x" Feb 27 18:44:52 crc kubenswrapper[4700]: I0227 18:44:52.318600 4700 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-cc5d6c7cf-x6qzc" Feb 27 18:45:00 crc kubenswrapper[4700]: I0227 18:45:00.171670 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536965-qb4hb"] Feb 27 18:45:00 crc kubenswrapper[4700]: E0227 18:45:00.173012 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f0c5a56-55ed-4dee-b36f-8faacb74ddbb" containerName="oc" Feb 27 18:45:00 crc kubenswrapper[4700]: I0227 18:45:00.173031 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f0c5a56-55ed-4dee-b36f-8faacb74ddbb" containerName="oc" Feb 27 18:45:00 crc kubenswrapper[4700]: I0227 18:45:00.173590 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f0c5a56-55ed-4dee-b36f-8faacb74ddbb" containerName="oc" Feb 27 18:45:00 crc kubenswrapper[4700]: I0227 18:45:00.174771 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536965-qb4hb" Feb 27 18:45:00 crc kubenswrapper[4700]: I0227 18:45:00.177622 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 27 18:45:00 crc kubenswrapper[4700]: I0227 18:45:00.178780 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 27 18:45:00 crc kubenswrapper[4700]: I0227 18:45:00.188903 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536965-qb4hb"] Feb 27 18:45:00 crc kubenswrapper[4700]: I0227 18:45:00.257919 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/326d1814-ee75-43f4-98fb-b23fc46d35ca-secret-volume\") pod \"collect-profiles-29536965-qb4hb\" (UID: \"326d1814-ee75-43f4-98fb-b23fc46d35ca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536965-qb4hb" Feb 27 18:45:00 crc kubenswrapper[4700]: I0227 18:45:00.259007 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/326d1814-ee75-43f4-98fb-b23fc46d35ca-config-volume\") pod \"collect-profiles-29536965-qb4hb\" (UID: \"326d1814-ee75-43f4-98fb-b23fc46d35ca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536965-qb4hb" Feb 27 18:45:00 crc kubenswrapper[4700]: I0227 18:45:00.259065 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqw46\" (UniqueName: \"kubernetes.io/projected/326d1814-ee75-43f4-98fb-b23fc46d35ca-kube-api-access-xqw46\") pod \"collect-profiles-29536965-qb4hb\" (UID: \"326d1814-ee75-43f4-98fb-b23fc46d35ca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536965-qb4hb" Feb 27 18:45:00 crc kubenswrapper[4700]: I0227 18:45:00.361336 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/326d1814-ee75-43f4-98fb-b23fc46d35ca-secret-volume\") pod \"collect-profiles-29536965-qb4hb\" (UID: \"326d1814-ee75-43f4-98fb-b23fc46d35ca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536965-qb4hb" Feb 27 18:45:00 crc kubenswrapper[4700]: I0227 18:45:00.361611 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/326d1814-ee75-43f4-98fb-b23fc46d35ca-config-volume\") pod \"collect-profiles-29536965-qb4hb\" (UID: \"326d1814-ee75-43f4-98fb-b23fc46d35ca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536965-qb4hb" Feb 27 18:45:00 crc kubenswrapper[4700]: I0227 18:45:00.361693 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqw46\" (UniqueName: \"kubernetes.io/projected/326d1814-ee75-43f4-98fb-b23fc46d35ca-kube-api-access-xqw46\") pod \"collect-profiles-29536965-qb4hb\" (UID: \"326d1814-ee75-43f4-98fb-b23fc46d35ca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536965-qb4hb" Feb 27 18:45:00 crc kubenswrapper[4700]: I0227 18:45:00.362578 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/326d1814-ee75-43f4-98fb-b23fc46d35ca-config-volume\") pod \"collect-profiles-29536965-qb4hb\" (UID: \"326d1814-ee75-43f4-98fb-b23fc46d35ca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536965-qb4hb" Feb 27 18:45:00 crc kubenswrapper[4700]: I0227 18:45:00.374115 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/326d1814-ee75-43f4-98fb-b23fc46d35ca-secret-volume\") pod \"collect-profiles-29536965-qb4hb\" (UID: \"326d1814-ee75-43f4-98fb-b23fc46d35ca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536965-qb4hb" Feb 27 18:45:00 crc kubenswrapper[4700]: I0227 18:45:00.387756 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqw46\" (UniqueName: \"kubernetes.io/projected/326d1814-ee75-43f4-98fb-b23fc46d35ca-kube-api-access-xqw46\") pod \"collect-profiles-29536965-qb4hb\" (UID: \"326d1814-ee75-43f4-98fb-b23fc46d35ca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536965-qb4hb" Feb 27 18:45:00 crc kubenswrapper[4700]: I0227 18:45:00.505108 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536965-qb4hb" Feb 27 18:45:00 crc kubenswrapper[4700]: I0227 18:45:00.980356 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536965-qb4hb"] Feb 27 18:45:01 crc kubenswrapper[4700]: I0227 18:45:01.863011 4700 generic.go:334] "Generic (PLEG): container finished" podID="326d1814-ee75-43f4-98fb-b23fc46d35ca" containerID="6c3918b98e21b5144d5999c5025d51a6b5f2137a5bd0ae59b69741491d4ab3cd" exitCode=0 Feb 27 18:45:01 crc kubenswrapper[4700]: I0227 18:45:01.863064 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536965-qb4hb" event={"ID":"326d1814-ee75-43f4-98fb-b23fc46d35ca","Type":"ContainerDied","Data":"6c3918b98e21b5144d5999c5025d51a6b5f2137a5bd0ae59b69741491d4ab3cd"} Feb 27 18:45:01 crc kubenswrapper[4700]: I0227 18:45:01.863308 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536965-qb4hb" event={"ID":"326d1814-ee75-43f4-98fb-b23fc46d35ca","Type":"ContainerStarted","Data":"438465131af9d941529af5b82e984af96dd3a105335e91a2a58c538355cfe54f"} Feb 27 18:45:03 crc kubenswrapper[4700]: I0227 18:45:03.312515 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536965-qb4hb" Feb 27 18:45:03 crc kubenswrapper[4700]: I0227 18:45:03.428327 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqw46\" (UniqueName: \"kubernetes.io/projected/326d1814-ee75-43f4-98fb-b23fc46d35ca-kube-api-access-xqw46\") pod \"326d1814-ee75-43f4-98fb-b23fc46d35ca\" (UID: \"326d1814-ee75-43f4-98fb-b23fc46d35ca\") " Feb 27 18:45:03 crc kubenswrapper[4700]: I0227 18:45:03.428844 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/326d1814-ee75-43f4-98fb-b23fc46d35ca-secret-volume\") pod \"326d1814-ee75-43f4-98fb-b23fc46d35ca\" (UID: \"326d1814-ee75-43f4-98fb-b23fc46d35ca\") " Feb 27 18:45:03 crc kubenswrapper[4700]: I0227 18:45:03.428999 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/326d1814-ee75-43f4-98fb-b23fc46d35ca-config-volume\") pod \"326d1814-ee75-43f4-98fb-b23fc46d35ca\" (UID: \"326d1814-ee75-43f4-98fb-b23fc46d35ca\") " Feb 27 18:45:03 crc kubenswrapper[4700]: I0227 18:45:03.429705 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/326d1814-ee75-43f4-98fb-b23fc46d35ca-config-volume" (OuterVolumeSpecName: "config-volume") pod "326d1814-ee75-43f4-98fb-b23fc46d35ca" (UID: "326d1814-ee75-43f4-98fb-b23fc46d35ca"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 18:45:03 crc kubenswrapper[4700]: I0227 18:45:03.434990 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/326d1814-ee75-43f4-98fb-b23fc46d35ca-kube-api-access-xqw46" (OuterVolumeSpecName: "kube-api-access-xqw46") pod "326d1814-ee75-43f4-98fb-b23fc46d35ca" (UID: "326d1814-ee75-43f4-98fb-b23fc46d35ca"). InnerVolumeSpecName "kube-api-access-xqw46". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:45:03 crc kubenswrapper[4700]: I0227 18:45:03.436620 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/326d1814-ee75-43f4-98fb-b23fc46d35ca-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "326d1814-ee75-43f4-98fb-b23fc46d35ca" (UID: "326d1814-ee75-43f4-98fb-b23fc46d35ca"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 18:45:03 crc kubenswrapper[4700]: I0227 18:45:03.536239 4700 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/326d1814-ee75-43f4-98fb-b23fc46d35ca-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 27 18:45:03 crc kubenswrapper[4700]: I0227 18:45:03.536270 4700 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/326d1814-ee75-43f4-98fb-b23fc46d35ca-config-volume\") on node \"crc\" DevicePath \"\"" Feb 27 18:45:03 crc kubenswrapper[4700]: I0227 18:45:03.536279 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqw46\" (UniqueName: \"kubernetes.io/projected/326d1814-ee75-43f4-98fb-b23fc46d35ca-kube-api-access-xqw46\") on node \"crc\" DevicePath \"\"" Feb 27 18:45:03 crc kubenswrapper[4700]: I0227 18:45:03.890988 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536965-qb4hb" event={"ID":"326d1814-ee75-43f4-98fb-b23fc46d35ca","Type":"ContainerDied","Data":"438465131af9d941529af5b82e984af96dd3a105335e91a2a58c538355cfe54f"} Feb 27 18:45:03 crc kubenswrapper[4700]: I0227 18:45:03.891044 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536965-qb4hb" Feb 27 18:45:03 crc kubenswrapper[4700]: I0227 18:45:03.891061 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="438465131af9d941529af5b82e984af96dd3a105335e91a2a58c538355cfe54f" Feb 27 18:45:04 crc kubenswrapper[4700]: I0227 18:45:04.436165 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536920-j946s"] Feb 27 18:45:04 crc kubenswrapper[4700]: I0227 18:45:04.449365 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536920-j946s"] Feb 27 18:45:05 crc kubenswrapper[4700]: I0227 18:45:05.002610 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e133aefd-7a5c-4ff6-8db8-f755e7a72327" path="/var/lib/kubelet/pods/e133aefd-7a5c-4ff6-8db8-f755e7a72327/volumes" Feb 27 18:45:06 crc kubenswrapper[4700]: I0227 18:45:06.410863 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 18:45:06 crc kubenswrapper[4700]: I0227 18:45:06.413365 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 18:45:14 crc kubenswrapper[4700]: I0227 18:45:14.946707 4700 scope.go:117] "RemoveContainer" containerID="b6054d92e394579612e9125748b8a31d02b92057bdd6b80d9c35254d3dd86e0a" Feb 27 18:45:14 crc kubenswrapper[4700]: I0227 18:45:14.986090 4700 scope.go:117] "RemoveContainer" containerID="0b22d4489434c448b1d3fdc0c20aebee56e79aca5fd52010bde72614dce6bcf6" Feb 27 18:45:36 crc kubenswrapper[4700]: I0227 18:45:36.410608 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 18:45:36 crc kubenswrapper[4700]: I0227 18:45:36.411334 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 18:46:00 crc kubenswrapper[4700]: I0227 18:46:00.178749 4700 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536966-728cm"] Feb 27 18:46:00 crc kubenswrapper[4700]: E0227 18:46:00.180718 4700 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="326d1814-ee75-43f4-98fb-b23fc46d35ca" containerName="collect-profiles" Feb 27 18:46:00 crc kubenswrapper[4700]: I0227 18:46:00.180746 4700 state_mem.go:107] "Deleted CPUSet assignment" podUID="326d1814-ee75-43f4-98fb-b23fc46d35ca" containerName="collect-profiles" Feb 27 18:46:00 crc kubenswrapper[4700]: I0227 18:46:00.181192 4700 memory_manager.go:354] "RemoveStaleState removing state" podUID="326d1814-ee75-43f4-98fb-b23fc46d35ca" containerName="collect-profiles" Feb 27 18:46:00 crc kubenswrapper[4700]: I0227 18:46:00.183013 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536966-728cm" Feb 27 18:46:00 crc kubenswrapper[4700]: I0227 18:46:00.186389 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 18:46:00 crc kubenswrapper[4700]: I0227 18:46:00.186926 4700 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 18:46:00 crc kubenswrapper[4700]: I0227 18:46:00.187205 4700 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-d62kc" Feb 27 18:46:00 crc kubenswrapper[4700]: I0227 18:46:00.222301 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536966-728cm"] Feb 27 18:46:00 crc kubenswrapper[4700]: I0227 18:46:00.302007 4700 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzn4v\" (UniqueName: \"kubernetes.io/projected/9bfd17e7-ee84-4500-8b85-876c2333c3d6-kube-api-access-qzn4v\") pod \"auto-csr-approver-29536966-728cm\" (UID: \"9bfd17e7-ee84-4500-8b85-876c2333c3d6\") " pod="openshift-infra/auto-csr-approver-29536966-728cm" Feb 27 18:46:00 crc kubenswrapper[4700]: I0227 18:46:00.404120 4700 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzn4v\" (UniqueName: \"kubernetes.io/projected/9bfd17e7-ee84-4500-8b85-876c2333c3d6-kube-api-access-qzn4v\") pod \"auto-csr-approver-29536966-728cm\" (UID: \"9bfd17e7-ee84-4500-8b85-876c2333c3d6\") " pod="openshift-infra/auto-csr-approver-29536966-728cm" Feb 27 18:46:00 crc kubenswrapper[4700]: I0227 18:46:00.429614 4700 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzn4v\" (UniqueName: \"kubernetes.io/projected/9bfd17e7-ee84-4500-8b85-876c2333c3d6-kube-api-access-qzn4v\") pod \"auto-csr-approver-29536966-728cm\" (UID: \"9bfd17e7-ee84-4500-8b85-876c2333c3d6\") " pod="openshift-infra/auto-csr-approver-29536966-728cm" Feb 27 18:46:00 crc kubenswrapper[4700]: I0227 18:46:00.529210 4700 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536966-728cm" Feb 27 18:46:01 crc kubenswrapper[4700]: I0227 18:46:01.101954 4700 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536966-728cm"] Feb 27 18:46:01 crc kubenswrapper[4700]: I0227 18:46:01.102606 4700 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 18:46:01 crc kubenswrapper[4700]: I0227 18:46:01.355878 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536966-728cm" event={"ID":"9bfd17e7-ee84-4500-8b85-876c2333c3d6","Type":"ContainerStarted","Data":"7aa23d2efbe36223b4d4236b19a8fcb0b5639e03aca5353a09d7421580e6ff90"} Feb 27 18:46:03 crc kubenswrapper[4700]: I0227 18:46:03.377650 4700 generic.go:334] "Generic (PLEG): container finished" podID="9bfd17e7-ee84-4500-8b85-876c2333c3d6" containerID="d2941039e90ba74cd825dab67410e736e6ddedaa50ede93e425f65515aa0d23d" exitCode=0 Feb 27 18:46:03 crc kubenswrapper[4700]: I0227 18:46:03.377745 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536966-728cm" event={"ID":"9bfd17e7-ee84-4500-8b85-876c2333c3d6","Type":"ContainerDied","Data":"d2941039e90ba74cd825dab67410e736e6ddedaa50ede93e425f65515aa0d23d"} Feb 27 18:46:04 crc kubenswrapper[4700]: I0227 18:46:04.859205 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536966-728cm" Feb 27 18:46:05 crc kubenswrapper[4700]: I0227 18:46:05.030001 4700 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qzn4v\" (UniqueName: \"kubernetes.io/projected/9bfd17e7-ee84-4500-8b85-876c2333c3d6-kube-api-access-qzn4v\") pod \"9bfd17e7-ee84-4500-8b85-876c2333c3d6\" (UID: \"9bfd17e7-ee84-4500-8b85-876c2333c3d6\") " Feb 27 18:46:05 crc kubenswrapper[4700]: I0227 18:46:05.051477 4700 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bfd17e7-ee84-4500-8b85-876c2333c3d6-kube-api-access-qzn4v" (OuterVolumeSpecName: "kube-api-access-qzn4v") pod "9bfd17e7-ee84-4500-8b85-876c2333c3d6" (UID: "9bfd17e7-ee84-4500-8b85-876c2333c3d6"). InnerVolumeSpecName "kube-api-access-qzn4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 18:46:05 crc kubenswrapper[4700]: I0227 18:46:05.133131 4700 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qzn4v\" (UniqueName: \"kubernetes.io/projected/9bfd17e7-ee84-4500-8b85-876c2333c3d6-kube-api-access-qzn4v\") on node \"crc\" DevicePath \"\"" Feb 27 18:46:05 crc kubenswrapper[4700]: I0227 18:46:05.407965 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536966-728cm" event={"ID":"9bfd17e7-ee84-4500-8b85-876c2333c3d6","Type":"ContainerDied","Data":"7aa23d2efbe36223b4d4236b19a8fcb0b5639e03aca5353a09d7421580e6ff90"} Feb 27 18:46:05 crc kubenswrapper[4700]: I0227 18:46:05.408026 4700 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7aa23d2efbe36223b4d4236b19a8fcb0b5639e03aca5353a09d7421580e6ff90" Feb 27 18:46:05 crc kubenswrapper[4700]: I0227 18:46:05.408119 4700 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536966-728cm" Feb 27 18:46:05 crc kubenswrapper[4700]: I0227 18:46:05.972561 4700 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536960-z9fjb"] Feb 27 18:46:05 crc kubenswrapper[4700]: I0227 18:46:05.987918 4700 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536960-z9fjb"] Feb 27 18:46:06 crc kubenswrapper[4700]: I0227 18:46:06.411219 4700 patch_prober.go:28] interesting pod/machine-config-daemon-f9dq7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 18:46:06 crc kubenswrapper[4700]: I0227 18:46:06.411354 4700 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 18:46:06 crc kubenswrapper[4700]: I0227 18:46:06.411513 4700 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" Feb 27 18:46:06 crc kubenswrapper[4700]: I0227 18:46:06.413077 4700 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b47e047e5315ba4c8f50802302b0905966f67ab104d0cf961afb16551a78daa7"} pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 18:46:06 crc kubenswrapper[4700]: I0227 18:46:06.413277 4700 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" podUID="84b45511-b94c-479f-98db-fd2c4eceec46" containerName="machine-config-daemon" containerID="cri-o://b47e047e5315ba4c8f50802302b0905966f67ab104d0cf961afb16551a78daa7" gracePeriod=600 Feb 27 18:46:06 crc kubenswrapper[4700]: I0227 18:46:06.996064 4700 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2960fff4-655f-432d-a3d0-8e2090558d8b" path="/var/lib/kubelet/pods/2960fff4-655f-432d-a3d0-8e2090558d8b/volumes" Feb 27 18:46:07 crc kubenswrapper[4700]: I0227 18:46:07.431915 4700 generic.go:334] "Generic (PLEG): container finished" podID="84b45511-b94c-479f-98db-fd2c4eceec46" containerID="b47e047e5315ba4c8f50802302b0905966f67ab104d0cf961afb16551a78daa7" exitCode=0 Feb 27 18:46:07 crc kubenswrapper[4700]: I0227 18:46:07.431979 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerDied","Data":"b47e047e5315ba4c8f50802302b0905966f67ab104d0cf961afb16551a78daa7"} Feb 27 18:46:07 crc kubenswrapper[4700]: I0227 18:46:07.432017 4700 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-f9dq7" event={"ID":"84b45511-b94c-479f-98db-fd2c4eceec46","Type":"ContainerStarted","Data":"ef7524305d95442178a5e1aaf0bc9a9719f8767a8c927fe06b50e994f1629226"} Feb 27 18:46:07 crc kubenswrapper[4700]: I0227 18:46:07.432045 4700 scope.go:117] "RemoveContainer" containerID="c7ad608c50ddee2aac4992df7a6ccaaeda7c737e2a0d2fad5282b4482eac9e50" Feb 27 18:46:15 crc kubenswrapper[4700]: I0227 18:46:15.126330 4700 scope.go:117] "RemoveContainer" containerID="1516063b9488f41ba916b1c4e0f7a55394b7ac28e08688f60a4f2308a9c540bf"